AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
PythonApache-2.0
Issues
- 23
Reproduce the commense results on Boolq
#64 opened by Zhenyu001225 - 0
How to see the evalution results?
#72 opened by gf457832386 - 1
Questions about the accuracy of eight commonsense reasoning datasets vs the Llama paper
#70 opened by Yonghao-Tan - 0
Baseline evaluation
#71 opened by Yonghao-Tan - 6
- 3
Question on the source of commonsense_15k
#69 opened by clarenceluo78 - 1
Question about datasets variants
#66 opened by ZeguanXiao - 13
- 2
Can not find BottleneckConfig
#68 opened by 1148514800 - 5
- 5
Errors when I run generation
#36 opened by ChaoGaoUCR - 1
about loss
#65 opened by haoyuwangwhy - 3
- 16
Full-Parameter Fine-Tuning on commonsense
#62 opened by lucasliunju - 6
Gibberish output
#63 opened by Aradhye2002 - 1
- 5
- 0
- 2
- 0
p-tuning in finetune.py?
#56 opened by smkim0220 - 7
- 1
请问为如何两次加载不同的微调后生成的lora权重?
#54 opened by jinlong7790 - 1
How to Reproduce BLOOMz-7b and GPT-j-6 results?
#50 opened by Ocean-627 - 1
weird evaluation results: 0% accuracy
#48 opened by wum67 - 6
ValueError: The version of PEFT you are using is not compatible, please use a version that is greater than 0.5.0
#47 opened by nbasyl - 7
- 0
How does chatglm support p-tuning in code?
#53 opened by lyt719 - 10
Question regarding the source of math_10k.json
#43 opened by HuangOwen - 3
Upload evaluation outputs and adapters
#46 opened by mkeoliya - 1
- 2
Details on provided peft
#45 opened by aksh555 - 1
how to download the dataset
#44 opened by ello0211 - 4
Questions about evaluate time
#40 opened by Yuan0320 - 2
Training reproduce
#41 opened by ChaoGaoUCR - 2
Peft version problem
#39 opened by marlin-codes - 4
- 4
- 2
AdapterH, AdapterP code
#35 opened by ChaoGaoUCR - 2
Eval without Tuning/Using OPT-1.3B
#34 opened by ChaoGaoUCR - 1
- 0
Could you give me an example to fintuning the Chatglm with adapter bottleneck please?
#29 opened by zhaojunGUO - 0
how to tune chatglm6b with dialogue dataset?
#28 opened by zhaojunGUO - 1
How to use llama-13B or bigger models?
#27 opened by feiyuehchen - 1
- 3
How to overwrite the Adapter
#24 opened by YuChen17Heaven - 4
Can not reproduce GSM8K zero-shot result
#16 opened by simplelifetime - 4
Questions about inconsistent results between the paper and the the README table.
#21 opened by lpyhdzx - 2
Question about the train and eval dataset for reporting the `Finetuned Result` table
#20 opened by ToheartZhang - 1
- 1
About the fp16 parameter setting
#17 opened by noob-ctrl