MarshtompCS's Stars
OpenGVLab/OmniCorpus
OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text
FranxYao/Retrieval-Head-with-Flash-Attention
Efficient retrieval head analysis with triton flash attention that supports topK probability
ivnle/synth-icl
mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
jzhang38/EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
pmichel31415/are-16-heads-really-better-than-1
Code for the paper "Are Sixteen Heads Really Better than One?"
apple/corenet
CoreNet: A library for training deep neural networks
crabml/crabml
a fast cross platform AI inference engine 🤖 using Rust 🦀 and WebGPU 🎮
ankurtaly/Integrated-Gradients
Attributing predictions made by the Inception network using the Integrated Gradients method
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
WecoAI/aideml
AIDE: the Machine Learning CodeGen Agent
TIGER-AI-Lab/LongICLBench
Code and Data for "Long-context LLMs Struggle with Long In-context Learning"
Shawn-Guo-CN/Lossless_Text_Compression_with_Transformer
This repo is to demo the concept of lossless compression with Transformers as encoder and decoder.
inseq-team/inseq
Interpretability for sequence generation models 🐛 🔍
OpenInterpreter/open-interpreter
A natural language interface for computers
EvolvingLMMs-Lab/lmms-eval
Accelerating the development of large multimodal models (LMMs) with lmms-eval
zepingyu0512/awesome-llm-understanding-mechanism
awesome papers in LLM interpretability
apple/ml-sigma-reparam
MiuLab/Taiwan-LLM
Traditional Mandarin LLMs for Taiwan
HazyResearch/based
Code for exploring Based models from "Simple linear attention language models balance the recall-throughput tradeoff"
jzhang38/LongMamba
Some preliminary explorations of Mamba's context scaling.
hf-lin/ChatMusician
bartwojcik/adaptive_computation_modules
uclnlp/EMAT
Efficient Memory-Augmented Transformers
OpenCodeInterpreter/OpenCodeInterpreter
OpenCodeInterpreter is a suite of open-source code generation systems aimed at bridging the gap between large language models and sophisticated proprietary systems like the GPT-4 Code Interpreter. It significantly enhances code generation capabilities by integrating execution and iterative refinement functionalities.
yuzhaouoe/pretraining-data-packing
LargeWorldModel/LWM
nushu-script/Nyushu
𛆁𛈬𛈤𛋒 | 女书字体 | Nüshu fonts
AntreasAntoniou/kubejobs
FranxYao/Long-Context-Data-Engineering
Implementation of paper Data Engineering for Scaling Language Models to 128K Context