CZWin32768
PhD student at Beijing Institute of Technology. Interned at Microsoft Research and ByteDance AI Lab.
Canada
CZWin32768's Stars
geekan/MetaGPT
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
meta-llama/llama3
The official Meta Llama 3 GitHub site
amusi/AI-Job-Notes
AI算法岗求职攻略(涵盖准备攻略、刷题指南、内推和AI公司清单等资料)
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
sovrasov/flops-counter.pytorch
Flops counter for convolutional networks in pytorch framework
paperswithcode/galai
Model API for GALACTICA
adapter-hub/adapters
A Unified Library for Parameter-Efficient and Modular Transfer Learning
zjunlp/LLMAgentPapers
Must-read Papers on LLM Agents.
EgoAlpha/prompt-in-context-learning
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.
deepseek-ai/DeepSeek-LLM
DeepSeek LLM: Let there be answers
databricks/megablocks
JShollaj/awesome-llm-interpretability
A curated list of Large Language Model (LLM) Interpretability resources.
deepseek-ai/DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
myshell-ai/JetMoE
Reaching LLaMA2 Performance with 0.1M Dollars
aqlaboratory/proteinnet
Standardized data set for machine learning of protein structure
alexa/massive
Tools and Modeling Code for the MASSIVE dataset
ziplab/SN-Net
[CVPR 2023 Highlight] This is the official implementation of "Stitchable Neural Networks".
IBM/ModuleFormer
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
neelnanda-io/1L-Sparse-Autoencoder
nlp-uoregon/mlmm-evaluation
Multilingual Large Language Models Evaluation Benchmark
nlp-uoregon/Okapi
Okapi: Instruction-tuned Large Language Models in Multiple Languages with Reinforcement Learning from Human Feedback
muhaochen/seq_ppi
This is the repository for PIPR. This repository contains the source code and links to some datasets used in the ISMB/ECCB-2019 paper "Multifaceted Protein-Protein Interaction Prediction Based on Siamese Residual RCNN".
DeepGraphLearning/ProtST
[ICML-23 ORAL] ProtST: Multi-Modality Learning of Protein Sequences and Biomedical Texts
FreedomIntelligence/MultilingualSIFT
MultilingualSIFT: Multilingual Supervised Instruction Fine-tuning
UMass-Foundation-Model/Mod-Squad
zqgao22/HIGH-PPI
mila-iqia/milatools
Tools to connect to and interact with the Mila cluster
ZurichNLP/multilingual-instruction-tuning
Code and data for the paper "Turning English-centric LLMs Into Polyglots: How Much Multilinguality Is Needed?"
ChillingDream/DAP
ACL 2023 Dual-Alignment Pre-training for Cross-lingual Sentence Embedding