Issues
- 1
LightSeq on GCP
#530 opened by mmcgra21 - 4
[FeedBack][1.0.8][com.miniclip.eightballpool]-
#529 opened by ap0jwd - 1
Is llama inference available now?
#513 opened by frankxyy - 0
is Baichuan2 model inference available now?
#528 opened by cg3dland - 0
参数格式不对args format wrong
#527 opened by whiteSteelRain - 4
identifier "__hisnan" is undefined
#524 opened by jimmieliu - 1
请问lightseq可以支持segmentAnyting的推理优化吗
#516 opened by sanbuphy - 0
Exception: Installed CUDA version 12.3 does not match the version torch was compiled with 12.1, unable to compile cuda/cpp extensions without a matching cuda version.
#526 opened by skill-diver - 1
- 1
- 4
为什么连给的example也有bug?
#517 opened by Moran232 - 0
lightseq是否支持clip模型的int8量化?
#522 opened by shhn1 - 0
Can int8 in pre-training large model ???
#521 opened by zhoumengbo - 3
什么时候可以支持bloom的几个版本的模型,比如6b的
#505 opened by liuzhipengchd - 3
llama inference test
#515 opened by HandH1998 - 0
how to resolve xlm-roberta convert fail
#520 opened by 520jefferson - 0
question about environment
#519 opened by etoilestar - 0
[Question] gptj, mpt support.
#518 opened by DongqiShen - 0
Do you have plans to support token_type_ids?
#514 opened by chenchongthu - 2
Is there a plan to support the T5-v1_1 model?
#448 opened by dengcunqin - 4
请问下可以支持llama和bloom推理加速吗
#502 opened by HuiResearch - 0
请问lightseq在推理流程中有gemm调参这一步吗?
#512 opened by frankxyy - 0
LLaMA example 结果验证
#511 opened by chenzhengda - 2
- 1
RuntimeError: Ninja is required to load C++ extensions even after pip install ninja
#503 opened by zt991211 - 0
lightseq' Transformer expects an extra layer_norm on both encoder and decoder level
#509 opened by yuting-wang-1000 - 0
Do you consider supporting the chatglm model?
#507 opened by Youggls - 0
ls_torch_hf_quant_gpt2_export.py的使用问题
#508 opened by wzh232894 - 0
How to get output scores for each output tokens of LightSeq BART model when inference
#506 opened by quancq - 5
[Question]: How to compile lightseq
#445 opened by FrostML - 4
- 0
Wrong encode_output_project_bias_kv_size !
#497 opened by JunchengYao - 0
Is there a plan to support the mT5-small model?
#496 opened by qibao77 - 4
About inference speed compared to TRT16? [急急如律令]
#490 opened by xiao2mo - 0
- 2
en2fr和en2de的模型结构存在差异?
#485 opened by MeJerry215 - 0
- 0
关于wmt14 en2de数据集的问题
#483 opened by MeJerry215 - 0
GPU speedup vs CTranslate2
#469 opened by nghuyong - 0
beam search
#455 opened by hg123-we - 1
转化之后的hdf5模型太大,lightseq 加载的时候支持多卡部署吗?
#450 opened by Liuchunyangboy - 1
- 0
- 2
请问对于LSTM的量化,有用cuBLASLt库实现的方案吗?
#449 opened by piaohe20221128 - 3
How to ensemble lightseq models? & the memory usage is too big when generating
#440 opened by baoguo1995 - 2
How to use lightseq for bert's mlm training?
#446 opened by ingale726 - 2
Possible memory leak in DecSelfAttentionLayer
#443 opened by Kangmo - 4
- 2
Multi-calorie running model problem
#444 opened by ingale726 - 1
Is it compatible with swin-t
#439 opened by Chaoran-F