pre-trained-language-models

There are 71 repositories under pre-trained-language-models topic.

  • Chinese-LLaMA-Alpaca

    ymcui/Chinese-LLaMA-Alpaca

    中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)

    Language:Python18.6k1847321.9k
  • RUCAIBox/LLMSurvey

    The official GitHub page for the survey paper "A Survey of Large Language Models".

    Language:Python10.8k15965837
  • thunlp/OpenPrompt

    An Open-Source Framework for Prompt-Learning.

    Language:Python4.4k44261457
  • thunlp/PromptPapers

    Must-read papers on prompt-based tuning for pre-trained language models.

  • ddangelov/Top2Vec

    Top2Vec learns jointly embedded topic, document and word vectors.

    Language:Python3k37332374
  • brightmart/roberta_zh

    RoBERTa中文预训练模型: RoBERTa for Chinese

    Language:Python2.7k5295411
  • zjunlp/KnowLM

    An Open-sourced Knowledgable Large Language Model Framework.

    Language:Python1.3k11136128
  • cedrickchee/awesome-transformer-nlp

    A curated list of NLP resources focused on Transformer networks, attention mechanism, GPT, BERT, ChatGPT, LLMs, and transfer learning.

  • zjunlp/KnowledgeEditingPapers

    Must-read Papers on Knowledge Editing for Large Language Models.

  • THUDM/P-tuning

    A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.

    Language:Python9272350111
  • txsun1997/LMaaS-Papers

    Awesome papers on Language-Model-as-a-Service (LMaaS)

  • sunyilgdx/SIFRank_zh

    Keyphrase or Keyword Extraction 基于预训练模型的中文关键词抽取方法(论文SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model 的中文版代码)

    Language:Python42283078
  • airaria/TextPruner

    A PyTorch-based model pruning toolkit for pre-trained language models

    Language:Python38251734
  • wjn1996/HugNLP

    HugNLP is a unified and comprehensive NLP library based on HuggingFace Transformer. Please hugging for NLP now!😊 HugNLP will released to @HugAILab

    Language:Python2478013
  • zjunlp/MolGen

    [ICLR 2024] Domain-Agnostic Molecular Generation with Chemical Feedback

    Language:Python14472212
  • zjunlp/DART

    [ICLR 2022] Differentiable Prompt Makes Pre-trained Language Models Better Few-shot Learners

    Language:Python1306916
  • sunyilgdx/SIFRank

    The code of our paper "SIFRank: A New Baseline for Unsupervised Keyphrase Extraction Based on Pre-trained Language Model"

    Language:Python1214920
  • zjunlp/MKG_Analogy

    [ICLR 2023] Multimodal Analogical Reasoning over Knowledge Graphs

    Language:Python10563011
  • TobiasLee/Awesome-Efficient-PLM

    Must-read papers on improving efficiency for pre-trained language models.

  • lyy1994/awesome-data-contamination

    The Paper List on Data Contamination for Large Language Models Evaluation.

  • Hanlard/Electra_CRF_NER

    We start a company-name recognition task with a small scale and low quality training data, then using skills to enhanced model training speed and predicting performance with least artificial participation. The methods we use involve lite pre-training models such as Albert-small or Electra-small with financial corpus, knowledge of distillation and multi-stage learning. The result is that we improve the recall rate of company names recognition task from 0.73 to 0.92 and get 4 times as fast as BERT-Bilstm-CRF model.

    Language:Python805314
  • Victorwz/VaLM

    VaLM: Visually-augmented Language Modeling. ICLR 2023.

    Language:Python56253
  • pat-jj/TagReal

    [ACL'23] Open KG Completion with PLM (Bridging Text Mining and Prompt Engineering)

    Language:Python51241
  • WangRongsheng/Chinese-LLaMA-Alpaca-Usage

    📔 对Chinese-LLaMA-Alpaca进行使用说明和核心代码注解

    Language:Jupyter Notebook48216
  • zjunlp/ChatCell

    ChatCell: Facilitating Single-Cell Analysis with Natural Language

    Language:Python47809
  • lancopku/DynamicKD

    Code for EMNLP 2021 main conference paper "Dynamic Knowledge Distillation for Pre-trained Language Models"

    Language:Python40616
  • anas-zafar/LLM-Survey

    The official GitHub page for the survey paper "Large language models: a comprehensive survey of its applications, challenges, limitations, and future prospects"

  • lanwuwei/GigaBERT

    Zero-shot Transfer Learning from English to Arabic

    Language:Python29625
  • ai2-ner-project/pytorch-ko-ner

    PLM 기반 한국어 개체명 인식 (NER)

    Language:Python28003
  • xuanyuan14/ARES

    SIGIR'22 paper: Axiomatically Regularized Pre-training for Ad hoc Search

    Language:Python24102
  • nkcs-iclab/linglong

    LingLong (玲珑): a small-scale Chinese pretrained language model

    Language:Python18001
  • XingLuxi/Cal-FLOPs-for-PLM

    Calculating FLOPs of Pre-trained Models in NLP

    Language:Python18214
  • cliang1453/super-structured-lottery-tickets

    Super Tickets in Pre-Trained Language Models: From Model Compression to Improving Generalization (ACL 2021)

    Language:Python17211
  • zjunlp/knowledge-rumination

    [EMNLP 2023] Knowledge Rumination for Pre-trained Language Models

    Language:Python17401
  • yuzhimanhua/SeeTopic

    Seed-Guided Topic Discovery with Out-of-Vocabulary Seeds (NAACL'22)

    Language:C16322
  • Navy10021/SLS

    SLS : Neural Information Retrieval(IR)-based Semantic Search model

    Language:Jupyter Notebook13114