gpt2
There are 363 repositories under gpt2 topic.
jessevig/bertviz
BertViz: Visualize Attention in NLP Models (BERT, GPT2, BART, etc.)
yuanzhoulvpi2017/zero_nlp
中文nlp解决方案(大模型、数据、模型、训练、推理)
yangjianxin1/GPT2-chitchat
GPT2 for Chinese chitchat/用于中文闲聊的GPT2模型(实现了DialoGPT的MMI**)
brightmart/roberta_zh
RoBERTa中文预训练模型: RoBERTa for Chinese
Tencent/TurboTransformers
a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.
920232796/bert_seq2seq
pytorch实现 Bert 做seq2seq任务,使用unilm方案,现在也可以做自动摘要,文本分类,情感分析,NER,词性标注等任务,支持t5模型,支持GPT2进行文章续写。
Hello-SimpleAI/chatgpt-comparison-detection
Human ChatGPT Comparison Corpus (HC3), Detectors, and more! 🔥
liucongg/GPT2-NewsTitle
Chinese NewsTitle Generation Project by GPT2.带有超级详细注释的中文GPT2新闻标题生成项目。
graykode/gpt-2-Pytorch
Simple Text-Generator with OpenAI gpt-2 Pytorch Implementation
shibing624/textgen
TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet and so on. 文本生成模型,实现了包括LLaMA,ChatGLM,BLOOM,GPT2,Seq2Seq,BART,T5,UDA等模型的训练和预测,开箱即用。
sail-sg/Adan
Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models
aub-mind/arabert
Pre-trained Transformers for Arabic Language Understanding and Generation (Arabic BERT, Arabic GPT2, Arabic ELECTRA)
rinnakk/japanese-pretrained-models
Code for producing Japanese pretrained models provided by rinna Co., Ltd.
benedekrozemberczki/GraphWaveletNeuralNetwork
A PyTorch implementation of "Graph Wavelet Neural Network" (ICLR 2019)
yangjianxin1/CPM
Easy-to-use CPM for Chinese text generation(基于CPM的中文文本生成)
jasoncao11/nlp-notebook
NLP 领域常见任务的实现,包括新词发现、以及基于pytorch的词向量、中文文本分类、实体识别、摘要文本生成、句子相似度判断、三元组抽取、预训练模型等。
Xirider/finetune-gpt2xl
Guide: Finetune GPT2-XL (1.5 Billion Parameters) and finetune GPT-NEO (2.7 B) on a single GPU with Huggingface Transformers using DeepSpeed
mindspore-courses/step_into_llm
MindSpore online courses: Step into LLM
kyzhouhzau/NLPGNN
1. Use BERT, ALBERT and GPT2 as tensorflow2.0's layer. 2. Implement GCN, GAN, GIN and GraphSAGE based on message passing.
affjljoo3581/GPT2
PyTorch Implementation of OpenAI GPT-2
NLP-kr/tensorflow-ml-nlp-tf2
텐서플로2와 머신러닝으로 시작하는 자연어처리 (로지스틱회귀부터 BERT와 GPT3까지) 실습자료
benedekrozemberczki/SGCN
A PyTorch implementation of "Signed Graph Convolutional Network" (ICDM 2018).
NetEase-FuXi/EET
Easy and Efficient Transformer : Scalable Inference Solution For Large NLP model
akanyaani/gpt-2-tensorflow2.0
OpenAI GPT2 pre-training and sequence prediction implementation in Tensorflow 2.0
teddykoker/image-gpt
PyTorch Implementation of OpenAI's Image GPT
dpressel/mint
MinT: Minimal Transformer Library and Tutorials
matthias-wright/flaxmodels
Pretrained deep learning models for Jax/Flax: StyleGAN2, GPT2, VGG, ResNet, etc.
gyunggyung/KoGPT2-FineTuning
🔥 Korean GPT-2, KoGPT2 FineTuning cased. 한국어 가사 데이터 학습 🔥
matlab-deep-learning/transformer-models
Deep Learning Transformer models in MATLAB
Sea-Snell/JAXSeq
Train very large language models in Jax.
lipiji/Guyu
Chinese GPT2: pre-training and fine-tuning framework for text generation
Laz4rz/GPT-2
Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish
rusiaaman/XLnet-gen
XLNet for generating language.
arian-askari/ChatGPT-RetrievalQA-CIKM2023
A dataset for training/evaluating Question Answering Retrieval models on ChatGPT responses with the possibility to training/evaluating on real human responses.
bme-chatbots/dialogue-generation
Generating responses with pretrained XLNet and GPT-2 in PyTorch.
gmftbyGMFTBY/OpenDialog
An Open-Source Package for Chinese Open-domain Conversational Chatbot (中文闲聊对话系统,一键部署微信闲聊机器人)