FlagAI-Open/Aquila2
The official repo of Aquila2 series proposed by BAAI, including pretrained & chat large language models.
Python
Issues
- 0
BAAI/AquilaChat2-7B 是否可以归纳总结文本内容?
#144 opened by Yshiboo - 1
官方可以做一版Aquila2Chat-7B到ModelScope上吗?
#143 opened by Micla-SHL - 1
Aquila_BGE_langchain应用所需配置大概需要在什么区间?
#138 opened by Micla-SHL - 0
- 0
How to review Aquila2-7B model
#141 opened by Micla-SHL - 2
Number of training tokens
#142 opened by RicardoDominguez - 1
- 0
指定2个以上gpu时./preprocess.sh: line 3: 2269 Bus error (core dumped) ,1个就不会,log对比在内文
#131 opened by tfal-yan - 1
How can I take the model trained using the script and apply it in Aquila_BGE_langchain?
#130 opened by Micla-SHL - 0
- 0
[Question] V100 FInetuning
#136 opened by mru4913 - 0
from predict import predict
#135 opened by adol001 - 2
- 1
- 27
bash finetune/34B/finetune_qlora_single_node.sh异常
#129 opened by tfal-yan - 1
bge有没有量化版本
#123 opened by yanchaoguo - 2
- 2
Aquila2/examples/Aquila_BGE_langchain/BGE# CUDA_VISIBLE_DEVICES=4,3,2,1,0 ./preprocess.sh 行 3: 2866484 段错误
#126 opened by tfal-yan - 3
- 2
- 9
7b模型qlora sft出错,transformers Version: 4.35.0 ,错误信息见内
#125 opened by tfal-yan - 2
- 2
是否支持function功能
#116 opened by hunter-xue - 8
16K的上下文长度是真实的吗?
#99 opened by DirtyKnightForVi - 5
RAG的问答能力很差。
#107 opened by HouYueJie - 1
使用NTK的情况下,推理输入超过2048 token模型无法正常回答
#105 opened by 0-1CxH - 13
- 1
请问34B模型全参数微调需要几张卡(A100、80G)?
#110 opened by zdaaaaa - 14
效果没有想象当中的好
#101 opened by lucasjinreal - 2
34B模型微调数据
#106 opened by Little-rookie-ee - 2
AutoGPTQ support
#100 opened by lucasjinreal - 3
aquilachat2-34b-16k这个的v1.2版本可以在哪里下载呀
#104 opened by Edisonwei54 - 6
flagai安装报错
#77 opened by zxzxde - 2
AttributeError: 'AquilaForCausalLM' object has no attribute 'save_checkpoint'
#84 opened by sincerity-being - 3
AquilaSQL-7B 的支持最大token数是多少呢?
#103 opened by nuaabuaa07 - 2
base-34B模型HF链接失效
#96 opened by HalcyonLiang - 3
finetune+chat34b 处理数据时发现问题
#94 opened by Darrenzeng - 2
- 1
ReadME中关于finetune数据的instruction字段
#79 opened by yuiant - 1
llama.cpp量化后的模型问答过程陷入自说自话
#95 opened by gaord - 4
请问4bit量化后的模型如何保存
#90 opened by sunyclj - 2
hostfile的作用及修改
#92 opened by eric971120 - 3
Token indices sequence length is longer than the specified maximum sequence length for this model (5669 > 2048). Running this sequence through the model will result in indexing errors
#93 opened by Edisonwei54 - 1
镜像启动容器报错
#78 opened by Darrenzeng - 1
- 2
关于Flash attention相关
#91 opened by onlyfish79 - 4
载入tokenizer时出现ValueError,请问怎么解决?
#75 opened by chk4991 - 2
finetune parameters
#85 opened by onlyfish79 - 2
4张4090能跑起来吗
#87 opened by bird-9 - 2
请问transformer代码示例中的predict文件在哪?
#86 opened by Bonjour0531