iteratorlee's Stars
public-apis/public-apis
A collective list of free APIs
Genymobile/scrcpy
Display and control your Android device
xai-org/grok-1
Grok open release
mlabonne/llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
lizongying/my-tv
我的电视 电视直播软件,安装即可使用
cumulo-autumn/StreamDiffusion
StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation
Mooler0410/LLMsPracticalGuide
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
mistralai/mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
SJTU-IPADS/PowerInfer
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
WooooDyy/LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
microsoft/LLMLingua
[EMNLP'23, ACL'24] To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
Alpha-VLLM/LLaMA2-Accessory
An Open-source Toolkit for LLM Development
Paitesanshi/LLM-Agent-Survey
stanford-crfm/helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image models in HEIM (https://arxiv.org/abs/2311.04287) and vision-language models in VHELM (https://arxiv.org/abs/2410.07112).
Eanya-Tonic/CCTV_Viewer
电视浏览器,一款简易电视视频收看软件,用于方便的在机顶盒上收看网页视频
Toyhom/Chinese-medical-dialogue-data
Chinese medical dialogue data 中文医疗对话数据集
deepseek-ai/DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
pjlab-sys4nlp/llama-moe
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)
AmberLJC/LLMSys-PaperList
Large Language Model (LLM) Systems Paper List
ray-project/llmperf-leaderboard
pratyushasharma/laser
The Truth Is In There: Improving Reasoning in Language Models with Layer-Selective Rank Reduction
cli99/llm-analysis
Latency and Memory Analysis of Transformer Models for Training and Inference
llm-random/llm-random
efeslab/fiddler
Fast Inference of MoE Models with CPU-GPU Orchestration
MARD1NO/CUDA-PPT
UNITES-Lab/MC-SMoE
[ICLR 2024 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"
hahnyuan/ASVD4LLM
Activation-aware Singular Value Decomposition for Compressing Large Language Models
yxli2123/LoSparse
Raphael-Hao/brainstorm
Compiler for Dynamic Neural Networks
YJHMITWEB/ExFlow
Explore Inter-layer Expert Affinity in MoE Model Inference