rolsheng's Stars
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
meta-llama/llama
Inference code for Llama models
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
meta-llama/llama3
The official Meta Llama 3 GitHub site
KaTeX/KaTeX
Fast math typesetting for the web.
xiaolincoder/CS-Base
图解计算机网络、操作系统、计算机组成、数据库,共 1000 张图 + 50 万字,破除晦涩难懂的计算机基础知识,让天下没有难懂的八股文!🚀 在线阅读:https://xiaolincoding.com
naklecha/llama3-from-scratch
llama3 implementation one matrix multiplication at a time
BradyFU/Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Advances on Multimodal Large Language Models
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
google/sentencepiece
Unsupervised text tokenizer for Neural Network-based text generation.
NVIDIA/apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
InternLM/MindSearch
🔍 An LLM-based Multi-agent Framework of Web Search Engine (like Perplexity.ai Pro and SearchGPT)
microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
jquesnelle/yarn
YaRN: Efficient Context Window Extension of Large Language Models
zonechen1994/CV_Interview
I hope this repo can help you a lot!
alibaba/Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
Wang-ML-Lab/llm-continual-learning-survey
Continual Learning of Large Language Models: A Comprehensive Survey
arcee-ai/PruneMe
Automated Identification of Redundant Layer Blocks for Pruning in Large Language Models
astramind-ai/Mixture-of-depths
Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
SysCV/shift-dev
SHIFT Dataset DevKit - CVPR2022
kyegomez/Mixture-of-Depths
Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
BeyonderXX/TRACE
TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models
rolsheng/MM-VUFM4DS
A systematic survey of multi-modal and multi-task visual understanding foundation models for driving scenarios
sramshetty/mixture-of-depths
An unofficial implementation of "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
Mixture-AI/Mixture-of-Depths
Google DeepMind: Mixture of Depths Unofficial Implementation.
rolsheng/OpenDet-D
OpenDet-D: Open World Object Detection via Cooperative Foundation Models for Driving Scenes