longlive1234's Stars
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
krahets/hello-algo
《Hello 算法》:动画图解、一键运行的数据结构与算法教程。支持 Python, Java, C++, C, C#, JS, Go, Swift, Rust, Ruby, Kotlin, TS, Dart 代码。简体版和繁体版同步更新,English version ongoing
ollama/ollama
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
mlabonne/llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
karpathy/LLM101n
LLM101n: Let's build a Storyteller
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
ddbourgin/numpy-ml
Machine learning, in numpy
graykode/nlp-tutorial
Natural Language Processing Tutorial for Deep Learning Researchers
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
microsoft/Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
state-spaces/mamba
Mamba SSM architecture
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Blealtan/efficient-kan
An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).
Nixtla/neuralforecast
Scalable and user friendly neural :brain: forecasting algorithms.
ddz16/TSFpaper
This repository contains a reading list of papers on Time Series Forecasting/Prediction (TSF) and Spatio-Temporal Forecasting/Prediction (STF). These papers are mainly categorized according to the type of model.
KimMeen/Time-LLM
[ICLR 2024] Official implementation of " 🦙 Time-LLM: Time Series Forecasting by Reprogramming Large Language Models"
NX-AI/xlstm
Official repository of the xLSTM.
kwuking/TimeMixer
[ICLR 2024] Official implementation of "TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting"
test-time-training/ttt-lm-pytorch
Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
NVlabs/MambaVision
Official PyTorch Implementation of MambaVision: A Hybrid Mamba-Transformer Vision Backbone
AntonioTepsich/Convolutional-KANs
This project extends the idea of the innovative architecture of Kolmogorov-Arnold Networks (KAN) to the Convolutional Layers, changing the classic linear transformation of the convolution to learnable non linear activations in each pixel.
GistNoesis/FourierKAN
EurekaLabsAI/micrograd
The Autograd Engine
IvanDrokin/torch-conv-kan
This project is dedicated to the implementation and research of Kolmogorov-Arnold convolutional networks. The repository includes implementations of 1D, 2D, and 3D convolutions with different kernels, ResNet-like and DenseNet-like models, training code based on accelerate/PyTorch, as well as scripts for experiments with CIFAR-10 and Tiny ImageNet.
test-time-training/ttt-lm-jax
Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
XiudingCai/Awesome-Mamba-Collection
A curated collection of papers, tutorials, videos, and other valuable resources related to Mamba.
Zyphra/tree_attention
Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters
XiudingCai/MambaTS-pytorch
Official code for MambaTS: Improved Selective State Space Models for Long-term Time Series Forecasting