ymcui/Chinese-BERT-wwm
Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
PythonApache-2.0
Issues
- 0
您好,我想请教一个代码上的细节
#243 opened by 4vicii - 3
有点晕:『开源版本不包含MLM任务的权重』,这句话是什么意思?
#226 opened by liangcaihua - 4
RoBERTa-wwm-ext-large应用到全新领域不收敛
#227 opened by JerryYao80 - 1
从BERT换成此模型后跑不通,应该怎么解决,显示缺少参数
#242 opened by ottt08 - 1
请问每个模型推理时需要多大的显存去加载?
#240 opened by JohnHerry - 1
- 2
Is onnx model available
#239 opened by panxin801 - 2
Is there any sharing about phoneme-BERT pretrained?
#238 opened by JohnHerry - 3
我用bert官方的分类代码去跑,用bert-wwm作为预训练模型,报错TypeError: __init__() takes 1 positional argument but 3 were given
#209 opened by dolphin-Jia - 2
NER问题
#237 opened by wanghao-007 - 2
请问可否使用wwm系列的模型做词级别的fill-mask预测?
#236 opened by LiShaoyu5 - 2
预训练数据
#235 opened by rattlesnakey - 2
如何抽取特定layer的词向量?
#234 opened by Black-Rhen - 2
- 8
关于不能用RobertaModel引用
#206 opened by Jackthebighead - 2
RoBERTa-wwm-ext-large ft的时候loss飞了
#232 opened by fword - 2
How WWM stratege works in code
#230 opened by DemonDamon - 2
链接失效求助
#231 opened by Alternate-D - 1
CJRC有预训练的模型么?
#229 opened by ucas010 - 1
How to download chinese-roberta-wwm-ext.pt ?
#228 opened by 2000yang - 4
预训练分词是会用到中文vocab后面带“##”的token吗?如果是,整词掩码预训练出来的语言模型,用在下游任务中,下游任务可以不分词吗?
#224 opened by dr-GitHub-account - 2
请问一下训练用了几块tpu呀
#223 opened by guohan950106 - 2
预训练阶段的loss问题
#222 opened by rtygbwwwerr - 4
- 2
wwm mask 细节
#219 opened by zhengjiawei001 - 3
关于词表没有中文双引号的问题
#221 opened by Dicer-Zz - 1
EXT数据集的量有多少
#218 opened by xueyuan1990 - 3
“注意:开源版本不包含MLM任务的权重;如需做MLM任务,请进行二次预训练。”
#216 opened by lxxXuan - 1
词汇级词向量
#215 opened by wangcccchen - 5
pad_token_id错误
#214 opened by CaoYiwei - 2
请问BERT-wwm-ext有使用nsp任务吗?
#213 opened by lionsterbenn - 2
- 2
请问RoBERTa-wwm-ext有没有随机初始化MLM部分的权重
#212 opened by dolphin-Jia - 2
- 4
讯飞的下载链接提示外链已经过期了
#211 opened by natureLanguageQing - 4
The way to get Chinese Word Embedding.
#208 opened by qhd1996 - 2
有没有Bert-wwm-ext-large模型?
#207 opened by zhaowei-wang-nlp - 2
rbt3将tf模型转torch报错
#204 opened by renjunxiang - 2
- 2
求问不同模型的schedule细节
#202 opened by NinedayWang - 2
求教 roberta既然在训练的时候未做NSP,那pooler_output 还有意义吗
#201 opened by rmbone - 2
wwm模型加载的时候tokenizer出来都是一个个的字,这样对吗?
#200 opened by rmbone - 3
求助:谁有pytorch预训练模型的国内下载地址吗?我从Google云下不下来,谢谢!
#199 opened by DaihuaWei - 3
roberta中cls token的计算方法
#198 opened by wikithink - 3
求教:为什么词语要对应多个mask?
#197 opened by fangwc - 2
> @beamind 请问复现结果了吗
#195 opened by houjin0803 - 2
序列标注任务警告信息
#194 opened by WillionLei - 2
这个模型可以不做fine-tuning直接用吗?
#193 opened by xiaohengheng - 2
- 2
实验结果复现问题
#196 opened by houjin0803