ZhuoruiLiu12's Stars
Devinterview-io/llms-interview-questions
🟣 LLMs interview questions and answers to help you prepare for your next machine learning and data science interview in 2024.
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
triton-lang/triton
Development repository for the Triton language and compiler
PKUFlyingPig/cs-self-learning
计算机自学指南
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
NLP-LOVE/ML-NLP
此项目是机器学习(Machine Learning)、深度学习(Deep Learning)、NLP面试中常考到的知识点和代码实现,也是作为一个算法工程师必会的理论基础知识。
wangshusen/RecommenderSystem
DefTruth/CUDA-Learn-Notes
📚Modern CUDA Learn Notes with PyTorch: Tensor/CUDA Cores, 📖150+ CUDA Kernels with PyTorch bindings, 📖HGEMM/SGEMM (95%~99% cuBLAS performance), 📖100+ LLM/CUDA Blogs.
shenweichen/AlgoNotes
【浅梦学习笔记】文章汇总:包含 排序&CXR预估,召回匹配,用户画像&特征工程,推荐搜索综合 计算广告,大数据,图算法,NLP&CV,求职面试 等内容
SakanaAI/evolutionary-model-merge
Official repository of Evolutionary Optimization of Model Merging Recipes
UbiquitousLearning/Efficient_Foundation_Model_Survey
Survey Paper List - Efficient LLM and Foundation Models
byungsoo-oh/ml-systems-papers
Curated collection of papers in machine learning systems
microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
bigscience-workshop/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
shunliz/Machine-Learning
机器学习原理
gkamradt/LLMTest_NeedleInAHaystack
Doing simple retrieval from LLM models at various context lengths to measure accuracy
arcee-ai/mergekit
Tools for merging pretrained large language models.
codecaution/Awesome-Mixture-of-Experts-Papers
A curated reading list of research in Mixture-of-Experts(MoE).
ggerganov/llama.cpp
LLM inference in C/C++
meta-llama/llama3
The official Meta Llama 3 GitHub site
zheyuye/resume-chinese
中文LaTeX简历模板
IST-DASLab/gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Strivin0311/long-llms-learning
A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks
DA-southampton/NLP_ability
总结梳理自然语言处理工程师(NLP)需要积累的各方面知识,包括面试题,各种基础知识,工程能力等等,提升核心竞争力
Tebmer/Awesome-Knowledge-Distillation-of-LLMs
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
wdndev/llm_interview_note
主要记录大语言大模型(LLMs) 算法(应用)工程师相关的知识及面试题
km1994/LLMs_interview_notes
该仓库主要记录 大模型(LLMs) 算法工程师相关的面试题
amusi/AI-Job-Notes
AI算法岗求职攻略(涵盖准备攻略、刷题指南、内推和AI公司清单等资料)
microsoft/TransformerCompression
For releasing code related to compression methods for transformers, accompanying our publications