pre-trained-language-models
There are 71 repositories under pre-trained-language-models topic.
ymcui/Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
RUCAIBox/LLMSurvey
The official GitHub page for the survey paper "A Survey of Large Language Models".
thunlp/OpenPrompt
An Open-Source Framework for Prompt-Learning.
thunlp/PromptPapers
Must-read papers on prompt-based tuning for pre-trained language models.
ddangelov/Top2Vec
Top2Vec learns jointly embedded topic, document and word vectors.
brightmart/roberta_zh
RoBERTa中文预训练模型: RoBERTa for Chinese
zjunlp/KnowLM
An Open-sourced Knowledgable Large Language Model Framework.
cedrickchee/awesome-transformer-nlp
A curated list of NLP resources focused on Transformer networks, attention mechanism, GPT, BERT, ChatGPT, LLMs, and transfer learning.
zjunlp/KnowledgeEditingPapers
Must-read Papers on Knowledge Editing for Large Language Models.
THUDM/P-tuning
A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.
txsun1997/LMaaS-Papers
Awesome papers on Language-Model-as-a-Service (LMaaS)
sunyilgdx/SIFRank_zh
Keyphrase or Keyword Extraction 基于预训练模型的中文关键词抽取方法(论文SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model 的中文版代码)
airaria/TextPruner
A PyTorch-based model pruning toolkit for pre-trained language models
wjn1996/HugNLP
HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 HugNLP will released to @HugAILab
zjunlp/MolGen
[ICLR 2024] Domain-Agnostic Molecular Generation with Chemical Feedback
zjunlp/DART
[ICLR 2022] Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners
sunyilgdx/SIFRank
The code of our paper "SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model"
zjunlp/MKG_Analogy
[ICLR 2023] Multimodal Analogical Reasoning over Knowledge Graphs
TobiasLee/Awesome-Efficient-PLM
Must-read papers on improving efficiency for pre-trained language models.
lyy1994/awesome-data-contamination
The Paper List on Data Contamination for Large Language Models Evaluation.
Hanlard/Electra_CRF_NER
We start a company-name recognition task with a small scale and low quality training data, then using skills to enhanced model training speed and predicting performance with least artificial participation. The methods we use involve lite pre-training models such as Albert-small or Electra-small with financial corpus, knowledge of distillation and multi-stage learning. The result is that we improve the recall rate of company names recognition task from 0.73 to 0.92 and get 4 times as fast as BERT-Bilstm-CRF model.
Victorwz/VaLM
VaLM: Visually-augmented Language Modeling. ICLR 2023.
pat-jj/TagReal
[ACL'23] Open KG Completion with PLM (Bridging Text Mining and Prompt Engineering)
WangRongsheng/Chinese-LLaMA-Alpaca-Usage
📔 对Chinese-LLaMA-Alpaca进行使用说明和核心代码注解
zjunlp/ChatCell
ChatCell: Facilitating Single-Cell Analysis with Natural Language
lancopku/DynamicKD
Code for EMNLP 2021 main conference paper "Dynamic Knowledge Distillation for Pre-trained Language Models"
anas-zafar/LLM-Survey
The official GitHub page for the survey paper "Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects"
lanwuwei/GigaBERT
Zero-shot Transfer Learning from English to Arabic
ai2-ner-project/pytorch-ko-ner
PLM 기반 한국어 개체명 인식 (NER)
xuanyuan14/ARES
SIGIR'22 paper: Axiomatically Regularized Pre-training for Ad hoc Search
nkcs-iclab/linglong
LingLong (玲珑): a small-scale Chinese pretrained language model
XingLuxi/Cal-FLOPs-for-PLM
Calculating FLOPs of Pre-trained Models in NLP
cliang1453/super-structured-lottery-tickets
Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization (ACL 2021)
zjunlp/knowledge-rumination
[EMNLP 2023] Knowledge Rumination for Pre-trained Language Models
yuzhimanhua/SeeTopic
Seed-Guided Topic Discovery with Out-of-Vocabulary Seeds (NAACL'22)
Navy10021/SLS
SLS : Neural Information Retrieval(IR)-based Semantic Search model