Facico/Chinese-Vicuna
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
CApache-2.0
Issues
- 16
关于中断训练继续训练。
#258 opened by xxyNeepu - 1
可以更新一下requirements吗?
#257 opened by estuday - 2
运行generate脚本之后,在页面提问,很久没有产生回答,后台无报错
#247 opened by mmmminyuhan - 0
如果更改數據集格式,要如何更改代碼
#255 opened by alexaax - 1
多卡训练 bash scripts/finetune.sh报错
#245 opened by hdjghjb - 2
可以使用原始文本微调吗
#212 opened by gravitywp - 0
官方colab安裝套件失效
#254 opened by alexaax - 0
可以提供一下huggingface上的Chinese-Vicuna/llama7b_4bit_128g模型的config.json和tokenizer么?
#253 opened by jasoncow007 - 2
使用finetune.sh来指令微调llama-33b,出现ZeroDivisionError: integer division or modulo by zero错误
#252 opened by BIUBIUBIU-JIAZHOU - 0
deepspeed跑模型相关问题
#250 opened by sunpenglv - 0
从belle+guanaco数据集中抽取前5000条样本训练lora,效果不好
#249 opened by huanghaifeng1234 - 0
OSError: Not enough disk space. Needed: Unknown size (download: Unknown size, generated: Unknown size, post-processed: Unknown size)
#248 opened by thugbobby - 5
训练模型没有保存token信息
#230 opened by apachemycat - 0
这几个不同路径下的模型是否有区别?
#246 opened by hdjghjb - 0
运行chat_7B.sh聊两句话out of memory
#244 opened by hdjghjb - 5
中文乱码
#241 opened by NewEricWang - 1
请问llama7b_4bit_128g的input shape是多少呢
#243 opened by KyrieZhang11 - 0
请问多个lora模型怎么合并?
#242 opened by Orangeices - 2
多卡finetune_chat时报mat1 and mat2 shapes cannot be multiplied (1024x2 and 1x11008)
#240 opened by 18065013 - 3
RuntimeError: mat1 and mat2 shapes cannot be multiplied (164x4096 and 1x8388608)
#228 opened by adaaaaaa - 0
- 1
transformers和pydantic问题
#238 opened by ww0o0 - 0
有办法改成分类任务么,用LlamaForSequenceClassification模型类加载
#237 opened by LeonhardtWang - 0
- 1
Traceback (most recent call last):RuntimeError: "addmm_impl_cpu_" not implemented for 'Half'
#231 opened by alps008 - 1
用checkpoint-11600跑部分问题(目测10-20%的问题)有奇怪的无限循环
#235 opened by Tongcheng - 1
- 0
简单的问题,finetune_other_continue.sh中step = 样本量/(MICRO_BATCH_SIZE*GRADIENT_ACCUMULATION_STEPS)。多卡的时候,是不是应该得是 step = 样本量/batch/(MICRO_BATCH_SIZE*GRADIENT_ACCUMULATION_STEPS*gpu数量)数量呢?这边一直不是很理解
#233 opened by niuhuluzhihao - 2
scripts中不同训练脚本的差异与训练效果的问题
#207 opened by FURYFOR - 5
拉去最新分支之后,通过pip install安装好了bitsandbytes==0.37.2,但是通过finetune_other_continue执行的时候,报此模块没有__version__
#219 opened by niuhuluzhihao - 0
7B 模型单卡3090后处理非常耗时
#232 opened by f18298335152h - 1
现在哪个模型支持4060笔记本显卡下的推理或者训练吗?
#209 opened by adaaaaaa - 1
在实际应用中我如何将num_beams=4,但最终输出的时候可保证输出过程和结果是一致的
#213 opened by zjwlgr - 2
- 1
我使用7B参数的上游模型 + 100万个问答数据集做微调,时间需要48天,如何能加快
#214 opened by zjwlgr - 1
官方 finetune colab 无法运行
#215 opened by williamjqk - 2
target_modules 各参数是什么意思,如何选择参数进行针对性的微调?
#216 opened by pan365wang - 4
- 3
- 1
代码中关于EOS paddding的区别问题
#221 opened by apachemycat - 2
llama-13b-hf做推理,CUDA out of memory. 问题
#224 opened by Bingohong - 3
运行generate.py推理报错ValueError: We need an `offload_dir` to dispatch this model
#225 opened by kakuibeyond - 1
Not an issue but a question for going forwards
#227 opened by thusinh1969 - 1
怎么区分user的问题是指令问题还是通用问题?
#229 opened by suckseed5 - 0
支持不同词表大小的 llama模型训练 lora
#226 opened by greatewei - 2
- 0
Generation问题
#223 opened by Jiangchenglin521 - 0
Infra问题
#222 opened by Jiangchenglin521 - 1
使用CPU运行13B的模型,有2个bin文件怎么选择呢
#211 opened by hengxingtx - 0
运行chat_7B.sh报错
#208 opened by hongshuo-wang