codefuse-ai/MFTCoder
High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.
PythonNOASSERTION
Issues
- 3
loss计算那里 RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn
#22 opened by hhy150 - 2
实验 MFTCoder 的效果总是不尽人意
#64 opened by Chaochao2020 - 0
model type
#65 opened by XiaoMaGe-hero - 5
- 1
数据集loss 下降不均衡如何处理
#55 opened by huangmenglong - 1
RuntimeError: CUDA error: invalid device ordinal
#57 opened by lwh8915 - 0
mftcoder使用humaneval评估
#59 opened by lwh8915 - 1
请问多机训练需要怎么修改?
#43 opened by jy00161yang - 2
在codellama上微调的性能没有提升
#18 opened by HPRCEST - 4
请教4int的gptq模型能不能进行lora微调
#34 opened by wengyuan722 - 1
任务的类型也是用gpt来生成的吗?
#48 opened by shatealaboxiaowang - 2
MFTCoder论文中训练数据集
#51 opened by superqing001 - 0
convergence curves
#52 opened by twelveand0 - 0
How can i do continue pretraining?
#47 opened by hwaking - 4
qlora微调合并权重时出错
#42 opened by fangzexian - 2
- 2
- 2
请问要支持chatglm3-6b-base的话需要哪些更改
#35 opened by kevindany - 1
请问下是否支持Wandb或者Tensorboard
#36 opened by pydaxing - 2
no 7B model size?
#30 opened by yiyepiaoling0715 - 1
Inquiry about weighted_loss_mode
#15 opened by tszdanger - 1
请问FSDP的训练API啥时候会开源出来
#14 opened by peiji1981 - 9
ValueError: Asking to pad but the tokenizer does not have a padding token. Please select a token to use as `pad_token` `(tokenizer.pad_token = tokenizer.eos_token e.g.)` or add a new pad token via `tokenizer.add_special_tokens({'pad_token': '[PAD]'})`.
#32 opened by sxsxsx - 1
- 3
- 29
如何构建codefuse-llamacode的提问和终止符
#23 opened by wengyuan722 - 1
Something wrong when run 'bash run_bash.sh'
#27 opened by MaoYouSi - 5
请问,对模型进行多任务微调该怎么设计jsonl数据集?
#16 opened by a793181018 - 1
麻烦我想问下一个可行性问题,对CodeFuse-CodeGeeX2-6B进行微调时是否可以使用peft的方式中chatglm2 config进行微调?万分感谢🙏
#13 opened by whyPeanutbutter - 5
模型是否支持商用
#33 opened by zhangyukun230 - 3
safetensors_rust.SafetensorError: Error while deserializing header: HeaderTooLarge
#20 opened by zzb2019053515 - 2
little bug fix meet
#21 opened by elcky - 2
单卡v1000,微调报错
#28 opened by sxsxsx - 5
- 0
数据问题ValueError: data format not supported, please use prompt/answer, or chatML or pretrain text
#26 opened by mst272 - 0
模型训练没有进度条
#19 opened by liujingqiao - 4
data.helper 无法加载?
#11 opened by liudonglei - 3
- 2
readme.txt指出,训练数据为jsonl格式,参考项目中的xxx.jsonl文件。未搜到对应的参考jsonl文件,能否麻烦给出一个示例?谢谢🙏
#12 opened by whyPeanutbutter - 1
基于chatgpt生成的高质量python练习题数据是如何获取呀
#6 opened by 18liumin - 3
使用lora + zero3微调CodeFuse-CodeLlama-34B后,合并模型失败
#8 opened by 3m123 - 2
HumanEval测试的Pass@1不高
#3 opened by wangzhao88 - 1
about focal loss mentioned in the paper
#10 opened by iDonal - 1
训练数据包含中文数据吗
#1 opened by smashfan - 1
能否写一个完整的微调例子?
#9 opened by liudonglei - 1