user074's Stars
meta-llama/llama3
The official Meta Llama 3 GitHub site
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
openai/swarm
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
princeton-nlp/SWE-agent
[NeurIPS 2024] SWE-agent takes a GitHub issue and tries to automatically fix it, using GPT-4, or your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges.
axolotl-ai-cloud/axolotl
Go ahead and axolotl questions
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
modelscope/ms-swift
Use PEFT or Full-parameter to finetune 400+ LLMs (Qwen2.5, Llama3.2, GLM4, Internlm2.5, Yi1.5, Mistral, Baichuan2, DeepSeek, ...) or 100+ MLLMs (Qwen2-VL, Qwen2-Audio, Llama3.2-Vision, Llava, InternVL2.5, MiniCPM-V-2.6, GLM4v, Xcomposer2.5, Yi-VL, DeepSeek-VL2, Phi3.5-Vision, GOT-OCR2, ...).
pytorch/torchtune
PyTorch native post-training library
meta-llama/llama-stack-apps
Agentic components of the Llama Stack APIs
pytorch/torchchat
Run PyTorch LLMs locally on servers, desktop and mobile
xjdr-alt/entropix
Entropy Based Sampling and Parallel CoT Decoding
prs-eth/Marigold
[CVPR 2024 - Oral, Best Paper Award Candidate] Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation
facebookresearch/schedule_free
Schedule-Free Optimization in PyTorch
cambrian-mllm/cambrian
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
openai/prm800k
800,000 step-level correctness labels on LLM solutions to MATH problems
landing-ai/vision-agent
Vision agent
b4rtaz/distributed-llama
Tensor parallelism is all you need. Run LLMs on an AI cluster at home using any device. Distribute the workload, divide RAM usage, and increase inference speed.
srush/awesome-o1
A bibliography and survey of the papers surrounding o1
chinasatokolo/csGraduateFellowships
A curated list of fellowships for graduate students in Computer Science and related fields.
uclaml/SPPO
The official implementation of Self-Play Preference Optimization (SPPO)
MARIO-Math-Reasoning/Super_MARIO
riiswa/kanrl
Kolmogorov-Arnold Network for Reinforcement Leaning, initial experiments
ikostrikov/implicit_q_learning
SergioMEV/slurm-for-dummies
A dummy's guide to setting up (and using) HPC clusters on Ubuntu 22.04LTS using Slurm and Munge. Created by the Quant Club @ UIowa.
Sea-Snell/Implicit-Language-Q-Learning
Official code from the paper "Offline RL for Natural Language Generation with Implicit Language Q Learning"
rejunity/tiny-asic-1_58bit-matrix-mul
Tiny ASIC implementation for "The Era of 1-bit LLMs All Large Language Models are in 1.58 Bits" matrix multiplication unit
fargusplumdoodle/dump_dir
Copy a bunch of files into your clipboard to provide context for LLMs
kyegomez/Lets-Verify-Step-by-Step
"Improving Mathematical Reasoning with Process Supervision" by OPENAI
elisakreiss/concadia