Pinned Repositories
clojure_findutils
create utilities such as find
gluon-tutorials-zh
通过MXNet/Gluon来动手学习深度学习
kaggle_cifar-10_mxnet
machine-learning
nanoDPO
A nimble and innovative implementation of the Direct Preference Optimization (DPO) algorithm with Causal Transformer and LSTM model, inspired by the paper of DPO in fine-tuning unsupervised Language Models
nanoPPO
An efficient implementation of the Proximal Policy Optimization (PPO) algorithm with linear and attention policy for reinforcement learning.
nanoTransformer
A PyTorch-based featuring an efficiently implemented Transformer model. The core of our attention mechanisms is powered by torch.einsum, ensuring clean, readable, and highly optimized tensor operations.
nChain
a flexible and efficient implementation to create LLM bots over extensible dataset.
xgboost
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Flink and DataFlow
autogen
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
jamesliu's Repositories
jamesliu/nanoDPO
A nimble and innovative implementation of the Direct Preference Optimization (DPO) algorithm with Causal Transformer and LSTM model, inspired by the paper of DPO in fine-tuning unsupervised Language Models
jamesliu/nChain
a flexible and efficient implementation to create LLM bots over extensible dataset.
jamesliu/amago
a simple and scalable agent for training adaptive policies with sequence-based RL
jamesliu/anthropic-cookbook
A collection of notebooks/recipes showcasing some fun and effective ways of using Claude.
jamesliu/autogen
Enable Next-Gen Large Language Model Applications. Join our Discord: https://discord.gg/pAbnFJrkgZ
jamesliu/bayjarvis-app
jamesliu/devika
Devika is an Agentic AI Software Engineer that can understand high-level human instructions, break them down into steps, research relevant information, and write code to achieve the given objective. Devika aims to be a competitive open-source alternative to Devin by Cognition AI.
jamesliu/lag-llama
jamesliu/litgpt
Hackable implementation of state-of-the-art open-source LLMs based on nanoGPT. Supports flash attention, 4-bit and 8-bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
jamesliu/LLaMA-Factory
Unify Efficient Fine-tuning of 100+ LLMs
jamesliu/llm-foundry
LLM training code for MosaicML foundation models
jamesliu/llm.c
LLM training in simple, raw C/CUDA
jamesliu/LLMLingua
To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
jamesliu/llmtime
jamesliu/long-range-arena
Long Range Arena for Benchmarking Efficient Transformers
jamesliu/magicoder
Magicoder: Source Code Is All You Need
jamesliu/network_architecture
jamesliu/OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
jamesliu/package-stats
jamesliu/Pearl
A Production-ready Reinforcement Learning AI Agent Library brought by the Applied Reinforcement Learning team at Meta.
jamesliu/pypi-bayjarvis-packages
jamesliu/recommendation_system
jamesliu/repopack
📦 Repopack is a powerful tool that packs your entire repository into a single, AI-friendly file. Perfect for when you need to feed your codebase to Large Language Models (LLMs) or other AI tools like Claude, ChatGPT, and Gemini.
jamesliu/robo-advisor-with-python
jamesliu/routerbench
The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System
jamesliu/ScaleLLM
A high-performance inference system for large language models, designed for production environments.
jamesliu/SPIN
The official implementation of Self-Play Fine-Tuning (SPIN)
jamesliu/Time-LLM
[ICLR 2024] Official implementation of "Time-LLM: Time Series Forecasting by Reprogramming Large Language Models"
jamesliu/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
jamesliu/Transformers_Are_What_You_Dont_Need
The best repository showing why transformers don’t work in time series forecasting and showcasing the best SOTA non transformer models.