WMWMW1's Stars
happyapplehorse/agere
The tool is used for building and driving workflows specifically tailored for AI initiatives. It can be used to construct AI agents.
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
WMWMW1/LM-from-scratch
LM from scratch
huggingface/accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
pinecone-io/examples
Jupyter Notebooks to help you get hands-on with Pinecone vector databases
snap-stanford/MLAgentBench
DLLXW/baby-llama2-chinese
用于从头预训练+SFT一个小参数量的中文LLaMa2的仓库;24G单卡即可运行得到一个具备简单中文问答能力的chat-llama2.
flowersteam/Grounding_LLMs_with_online_RL
We perform functional grounding of LLMs' knowledge in BabyAI-Text
assafelovic/gpt-researcher
LLM based autonomous agent that does online comprehensive research on any given topic
sweetice/Deep-reinforcement-learning-with-pytorch
PyTorch implementation of DQN, AC, ACER, A2C, A3C, PG, DDPG, TRPO, PPO, SAC, TD3 and ....
nikhilbarhate99/PPO-PyTorch
Minimal implementation of clipped objective Proximal Policy Optimization (PPO) in PyTorch
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama3 for WhatsApp & Messenger.
yangjianxin1/Firefly-LLaMA2-Chinese
Firefly中文LLaMA-2大模型,支持增量预训练Baichuan2、Llama2、Llama、Falcon、Qwen、Baichuan、InternLM、Bloom等大模型
Farama-Foundation/chatarena
ChatArena (or Chat Arena) is a Multi-Agent Language Game Environments for LLMs. The goal is to develop communication and collaboration capabilities of AIs.
facebookresearch/ImageBind
ImageBind One Embedding Space to Bind Them All
binance/binance-spot-api-docs
Official Documentation for the Binance Spot APIs and Streams
bgavran/Category_Theory_Machine_Learning
List of papers studying machine learning through the lens of category theory
mlii/mfrl
Mean Field Multi-Agent Reinforcement Learning
BorealisAI/mtmfrl
Multi Type Mean Field Reinforcement Learning
jasonyip184/Index-Tracking-Portfolio-Optimization
ggerganov/llama.cpp
LLM inference in C/C++
qmfin/index_data
Dataset for index tracking
OrigamiDream/gato
Unofficial Gato: A Generalist Agent
mingkai-zheng/GENIUS
Can GPT-4 Perform Neural Architecture Search?