pbw-Berwin's Stars
Significant-Gravitas/AutoGPT
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
facebookresearch/segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
tloen/alpaca-lora
Instruct-tune LLaMA on consumer hardware
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
mistralai/mistral-inference
Official inference library for Mistral models
facebookresearch/dinov2
PyTorch code and models for the DINOv2 self-supervised learning method.
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
OpenBMB/MiniCPM
MiniCPM3-4B: An edge-side LLM that surpasses GPT-3.5-Turbo.
pytorch-labs/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
aimhubio/aim
Aim 💫 — An easy-to-use & supercharged open-source experiment tracker.
OpenGVLab/InternGPT
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)
databricks/megablocks
horseee/Awesome-Efficient-LLM
A curated list for Efficient Large Language Models
ibm-granite/granite-code-models
Granite Code Models: A Family of Open Foundation Models for Code Intelligence
microsoft/tutel
Tutel MoE: An Optimized Mixture-of-Experts Implementation
sail-sg/lorahub
[COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition
princeton-nlp/LLM-Shearing
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
declare-lab/instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
pointnetwork/point-alpaca
mathllm/MathCoder
Family of LLMs for mathematical reasoning.
IBM/ModuleFormer
ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward experts. We released a collection of ModuleFormer-based Language Models (MoLM) ranging in scale from 4 billion to 8 billion parameters.
HanGuo97/flute
Fast Matrix Multiplications for Lookup Table-Quantized LLMs
shawntan/scattermoe
Triton-based implementation of Sparse Mixture of Experts.
stanford-futuredata/stk
X-PLUG/mPLUG
mPLUG: Effective and Efficient Vision-Language Learning by Cross-modal Skip-connections. (EMNLP 2022)
YicongHong/Fine-Grained-R2R
Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation
astra-vision/FAMix
[CVPR 2024] Official repository of "A Simple Recipe for Language-guided Domain Generalized Segmentation"
Lyken17/FlashATM