thedarkzeno's Stars
KindXiaoming/pykan
Kolmogorov Arnold Networks
abetlen/llama-cpp-python
Python bindings for llama.cpp
outlines-dev/outlines
Structured Text Generation
microsoft/LLMLingua
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Blealtan/efficient-kan
An efficient pure-PyTorch implementation of Kolmogorov-Arnold Network (KAN).
ridgerchu/matmulfreellm
Implementation for MatMul-free LM.
mintisan/awesome-kan
A comprehensive collection of KAN(Kolmogorov-Arnold Network)-related resources, including libraries, projects, tutorials, papers, and more, for researchers and developers in the Kolmogorov-Arnold Network field.
atong01/conditional-flow-matching
TorchCFM: a Conditional Flow Matching library
lucidrains/RETRO-pytorch
Implementation of RETRO, Deepmind's Retrieval based Attention net, in Pytorch
GistNoesis/FourierKAN
AIGText/Glyph-ByT5
[ECCV2024] This is an official inference code of the paper "Glyph-ByT5: A Customized Text Encoder for Accurate Visual Text Rendering" and "Glyph-ByT5-v2: A Strong Aesthetic Baseline for Accurate Multilingual Visual Text Rendering""
BobMcDear/attorch
A subset of PyTorch's neural network modules, written in Python using OpenAI's Triton.
cloneofsimo/minRF
Minimal implementation of scalable rectified flow transformers, based on SD3's approach
ZiyaoLi/fast-kan
FastKAN: Very Fast Implementation of Kolmogorov-Arnold Networks (KAN)
arcee-ai/DistillKit
An Open Source Toolkit For LLM Distillation
HKUNLP/ChunkLlama
[ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"
Beomi/InfiniTransformer
Unofficial PyTorch/🤗Transformers(Gemma/Llama3) implementation of Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention
valine/NeuralFlow
Visualize the intermediate output of Mistral 7B
dingo-actual/infini-transformer
PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)
princeton-nlp/AutoCompressors
[EMNLP 2023] Adapting Language Models to Compress Long Contexts
Zyphra/Zamba2
PyTorch implementation of models from the Zamba2 series.
astramind-ai/BitMat
An efficent implementation of the method proposed in "The Era of 1-bit LLMs"
genyrosk/gym-chess
A simple chess environment for openai/gym
SelfishGene/neuron_as_deep_net
Code behind the work "Single Cortical Neurons as Deep Artificial Neural Networks", published in Neuron 2021
gle-bellier/flow-matching
Annotated Flow Matching paper
Geotrend-research/smaller-transformers
Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.
kjslag/spacebyte
A byte-level decoder architecture that matches the performance of tokenized Transformers.
Zyphra/transformers_zamba2
vasqu/mamba2-torch
FrameNetBrasil/framed-multi30k