madehong
I'am a Ph.D student of EECS in Peking University, and I focus on NLP and deep learning.
Peking UniversityBeijing
Pinned Repositories
AdaBound
An optimizer that trains as fast as Adam and as good as SGD.
ALBERT
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
attention-is-all-you-need-pytorch
A PyTorch implementation of the Transformer model in "Attention is All You Need".
bert-finetune
Codes for fine-tuning Bert for kinds of tasks.
BERT-pytorch
Google AI 2018 BERT pytorch implementation
madehong.github.io
Codes for my homepage.
Seq2Seq4ATE
Codes for paper Exploring Sequence-to-Sequence Learning for Aspect Term Extraction.
spread
WuDaoCorpus
迄今为止全球最大的中文语料库
madehong's Repositories
madehong/Seq2Seq4ATE
Codes for paper Exploring Sequence-to-Sequence Learning for Aspect Term Extraction.
madehong/WuDaoCorpus
迄今为止全球最大的中文语料库
madehong/bert-finetune
Codes for fine-tuning Bert for kinds of tasks.
madehong/ALBERT
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
madehong/madehong.github.io
Codes for my homepage.
madehong/alpaca-chinese-dataset
alpaca中文指令微调数据集
madehong/Alpaca-CoT
We extend CoT data to Alpaca to boost its reasoning ability. We are constantly expanding our collection of instruction-tuning data. The instruction collection can be found at https://huggingface.co/datasets/QingyiSi/Alpaca-CoT/tree/main (我们将CoT数据扩展到Alpaca以提高其推理能力,同时我们将不断收集更多的instruction-tuning数据集。)
madehong/BELLE-prompt
BELLE: Bloom-Enhanced Large Language model Engine(开源中文对话大模型-70亿参数)
madehong/CDial-GPT
A Large-scale Chinese Short-Text Conversation Dataset and Chinese pre-training dialog models
madehong/Chinese-alpaca-lora
骆驼:A Chinese finetuned instruction LLaMA. Developed by 陈启源 @ 华中师范大学 & 李鲁鲁 @ 商汤科技 & 冷子昂 @ 商汤科技
madehong/cleanlab
The standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.
madehong/CLUECorpus2020
Large-scale Pre-training Corpus for Chinese 100G 中文预训练语料
madehong/ERNIE2Pytorch
ERNIE Pytorch Version
madehong/fast-bert
Super easy library for BERT based NLP models
madehong/FLAN
madehong/GPT2-Chinese
Chinese version of GPT2 training code, using BERT tokenizer.
madehong/LSH_Attention
Calculate Softmax layer of Attention in O(LlogL)(L=sequence length) instead of O(L^2) using polytope Locality-Sensitive Hashing(https://arxiv.org/abs/1802.05751 ).
madehong/NeuroNLP2
Deep neural models for core NLP tasks (Pytorch version)
madehong/nlp_chinese_corpus
大规模中文自然语言处理语料 Large Scale Chinese Corpus for NLP
madehong/pkuthss
LaTeX template for dissertations in Peking University
madehong/pyGAT
Pytorch implementation of the Graph Attention Network model by Veličković et. al (2017, https://arxiv.org/abs/1710.10903)
madehong/RLHF
Implementation of Chinese ChatGPT
madehong/SemBERT
Semantics-aware BERT for Language Understanding (AAAI 2020)
madehong/SG-Net
AAAI2020: SG-Net: Syntax-guided machine reading comprehension
madehong/Statistical-Learning-Methods
Implement Statistical Leanring Methods, Li Hang the hard way. 李航《统计学习方法》一书的硬核 Python 实现
madehong/summarize-from-feedback-RL
Code for "Learning to summarize from human feedback"
madehong/Tencent2020_Rank1st
The code for 2020 Tencent College Algorithm Contest, and the online result ranks 1st.
madehong/transformer-xl-chinese
transformer xl在中文文本生成上的尝试(可写小说、古诗)(transformer xl for text generation of chinese)
madehong/TwinBert
pytorch implementation of the TwinBert paper
madehong/vimrc
The ultimate Vim configuration (vimrc)