HaoBytes
PhD Student @ University of Manchester Research Intern @ Microsoft Research
University of ManchesterManchester
HaoBytes's Stars
openai/chatgpt-retrieval-plugin
The ChatGPT Retrieval Plugin lets you easily find personal or work documents by asking questions in natural language.
zhpmatrix/PaperReading
每天阅读过的论文的简要笔记
truera/trulens
Evaluation and Tracking for LLM Experiments
songyingxin/NLPer-Interview
该仓库主要记录 NLP 算法工程师相关的面试题
facebookresearch/atlas
Code repository for supporting the paper "Atlas Few-shot Learning with Retrieval Augmented Language Models",(https//arxiv.org/abs/2208.03299)
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
google-research/FLAN
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
SinclairCoder/Instruction-Tuning-Papers
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
Hellisotherpeople/DebateSum
Corresponding code repo for the paper at COLING 2020 - ARGMIN 2020: "DebateSum: A large-scale argument mining and summarization dataset"
michael-wzhu/PromptCBLUE
PromptCBLUE: a large-scale instruction-tuning dataset for multi-task and few-shot learning in the medical domain in Chinese
CogStack/OpenGPT
A framework for creating grounded instruction based datasets and training conversational domain expert Large Language Models (LLMs).
explosion/spacy-llm
🦙 Integrating LLMs into structured NLP pipelines
yandex/YaLM-100B
Pretrained language model with 100B parameters
THUDM/GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
firebase/firebase-admin-python
Firebase Admin Python SDK
declare-lab/tango
A family of diffusion models for text-to-audio generation.
schlevik/flan-alpaca
This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as Flan-T5.
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
OpenBMB/BMTools
Tool Learning for Big Models, Open-Source Solutions of ChatGPT-Plugins
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
reoneo97/Whats_Cooking
Predicting Cuisines from Recipe
tloen/alpaca-lora
Instruct-tune LLaMA on consumer hardware
ccclyu/awesome-deeplogic
A collection of papers of neural-symbolic AI (mainly focus on NLP applications)
yuweihao/reclor
Code for "ReClor: A Reading Comprehension Dataset Requiring Logical Reasoning" (ICLR 2020)
yizhongw/self-instruct
Aligning pretrained language models with instruction data generated by themselves.
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
lucidrains/toolformer-pytorch
Implementation of Toolformer, Language Models That Can Use Tools, by MetaAI
pwin/owlready2
togethercomputer/OpenChatKit