Pinned Repositories
cogcomp-nlp
CogComp's Natural Language Processing Libraries and Demos: Modules include lemmatizer, ner, pos, prep-srl, quantifier, question type, relation-extraction, similarity, temporal normalizer, tokenizer, transliteration, verb-sense, and more.
PyMarlin
Lightweight Deep Learning Model Training library based on PyTorch
adapter-transformers
Huggingface Transformers + Adapters = ❤️
AgentGym
Code and implementations for the paper "AgentGym: Evolving Large Language Model-based Agents across Diverse Environments" by Zhiheng Xi et al.
alpaca-lora
Instruct-tune LLaMA on consumer hardware
Annotated-WikiExtractor
Simple Wikipedia plain text extractor with article link annotations and Hadoop support.
appworld
🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL2024
DialoGPT
Large-scale pretraining for dialogue
Docker-Containers
Joint-NER-RelEx-Coref
Code for Joint Modeling of NER, Relation Extraction and Coreference Resolution using Constrained Conditional Models
shatu's Repositories
shatu/adapter-transformers
Huggingface Transformers + Adapters = ❤️
shatu/AgentGym
Code and implementations for the paper "AgentGym: Evolving Large Language Model-based Agents across Diverse Environments" by Zhiheng Xi et al.
shatu/alpaca-lora
Instruct-tune LLaMA on consumer hardware
shatu/appworld
🌍 Repository for "AppWorld: A Controllable World of Apps and People for Benchmarking Interactive Coding Agent", ACL2024
shatu/DialoGPT
Large-scale pretraining for dialogue
shatu/Docker-Containers
shatu/awesome-system-design-resources
This repository contains System Design resources which are useful while preparing for interviews and learning Distributed Systems
shatu/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.
shatu/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.
shatu/Generating_Text_Summary_With_GPT2
A simple approach to use GPT2-medium (345M) for generating high quality text summaries with minimal training.
shatu/gorilla
Gorilla: An API store for LLMs
shatu/langchain
⚡ Building applications with LLMs through composability ⚡
shatu/NeuralDialog-CVAE
Tensorflow Implementation of Knowledge-Guided CVAE for dialog generation ACL 2017. It is released by Tiancheng Zhao (Tony) from Dialog Research Center, LTI, CMU
shatu/OpenHands
🙌 OpenHands: Code Less, Make More
shatu/ParlAI
A framework for training and evaluating AI models on a variety of openly available dialogue datasets.
shatu/prm800k
800,000 step-level correctness labels on LLM solutions to MATH problems
shatu/PyMarlin
Lightweight Deep Learning Model Training library based on PyTorch
shatu/pytorch-pretrained-BERT
📖The Big-&-Extending-Repository-of-Transformers: Pretrained PyTorch models for Google's BERT, OpenAI GPT & GPT-2, Google/CMU Transformer-XL.
shatu/reasoning-on-cots
shatu/SelfEval-Guided-Decoding
shatu/shatu.github.io
Code for the personal website
shatu/SimCSE
EMNLP'2021: SimCSE: Simple Contrastive Learning of Sentence Embeddings https://arxiv.org/abs/2104.08821
shatu/SpaceFusion
An implementation for the SpaceFusion model, https://arxiv.org/abs/1902.11205
shatu/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
shatu/t5x
shatu/TheoremQA
The dataset and code for paper: TheoremQA: A Theorem-driven Question Answering dataset
shatu/ThoughtSource
A central, open resource for data and tools related to chain-of-thought reasoning in large language models. Developed @ Samwald research group: https://samwald.info/
shatu/tree-of-thought-llm
Tree of Thoughts: Deliberate Problem Solving with Large Language Models
shatu/unify-parameter-efficient-tuning
Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)
shatu/xLAM