mshtelma's Stars
mlabonne/llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
guidance-ai/guidance
A guidance language for controlling large language models.
huggingface/candle
Minimalist ML framework for Rust
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
state-spaces/mamba
Mamba SSM architecture
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
LargeWorldModel/LWM
Large World Model -- Modeling Text and Video with Millions Context
deepseek-ai/DeepSeek-Coder
DeepSeek Coder: Let the Code Write Itself
NExT-GPT/NExT-GPT
Code and models for NExT-GPT: Any-to-Any Multimodal Large Language Model
amazon-science/chronos-forecasting
Chronos: Pretrained (Language) Models for Probabilistic Time Series Forecasting
NVIDIA/GenerativeAIExamples
Generative AI reference workflows optimized for accelerated infrastructure and microservice architecture.
intel/neural-compressor
SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
huggingface/datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
facebookresearch/schedule_free
Schedule-Free Optimization in PyTorch
zjunlp/LLMAgentPapers
Must-read Papers on LLM Agents.
S-LoRA/S-LoRA
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
systemdesignfightclub/SDFC
Roadmap and Resource Compilation for System Design Fight Club
SqueezeAILab/LLMCompiler
[ICML 2024] LLMCompiler: An LLM Compiler for Parallel Function Calling
Lightning-AI/lightning-thunder
Make PyTorch models up to 40% faster! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once; across one or thousands of GPUs.
Azure/GPT-RAG
Sharing the learning along the way we been gathering to enable Azure OpenAI at enterprise scale in a secure manner. GPT-RAG core is a Retrieval-Augmented Generation pattern running in Azure, using Azure Cognitive Search for retrieval and Azure OpenAI large language models to power ChatGPT-style and Q&A experiences.
ContextualAI/gritlm
Generative Representational Instruction Tuning
ToluClassics/candle-tutorial
Tutorial for Porting PyTorch Transformer Models to Candle (Rust)
normster/llm_rules
RuLES: a benchmark for evaluating rule-following in language models
lucidrains/CALM-pytorch
Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind
corl-team/rebased
Official implementation of the paper "Linear Transformers with Learnable Kernel Functions are Better In-Context Models"
AI4Finance-Foundation/Deep-Reinforcement-Learning-for-Stock-Trading-DDPG-Algorithm-NIPS-2018
Practical Deep Reinforcement Learning Approach for Stock Trading. NeurIPS 2018 AI in Finance.
austin-starks/Deep-RL-Stocks
Reinforcement Learning for Stock Market Prediction
allisonwang-db/pyspark-data-sources
Custom PySpark Data Sources
rmosleydb/agent-studio
A Databricks framework for quick Agent solutions
Data-drone/OpenMMLab-testing
Running OpenMMLab on Databricks