Pinned Repositories
2023AIWIN_Competition
dawoshi
meta-comprehensive-rag-benchmark-kdd-cup-2024
meta-comprehensive-rag-benchmark-kdd-cup-2024 phase1 task1 rank3
minimind
【大模型】3小时完全从0训练一个仅有26M的小参数GPT,最低仅需2G显卡即可推理训练!
nlp-tutorial
nlp-tutorial includes a Chromium base Library and named entity recognition and text classification based on this implementation. Named entity recognition includes span ner and mrc ner, Text classification includes BERT text classification, including their Python training and c++ engineering
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
tensorrtllm_backend
The Triton TensorRT-LLM Backend
Tianchi-LLM-QA
阿里天池: 2023全球智能汽车AI挑战赛——赛道一:AI大模型检索问答 baseline 80+
main_content_extractor
A library to extract the main content from html. Developed for information on LLM and for feeding data into LangChain and LlamaIndex.
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
dawoshi's Repositories
dawoshi/Tianchi-LLM-QA
阿里天池: 2023全球智能汽车AI挑战赛——赛道一:AI大模型检索问答 baseline 80+
dawoshi/meta-comprehensive-rag-benchmark-kdd-cup-2024
meta-comprehensive-rag-benchmark-kdd-cup-2024 phase1 task1 rank3
dawoshi/nlp-bazel-tutorial
nlp-bazel-tutorial includes a Chromium base Library and named entity recognition and text classification based on this implementation. Named entity recognition includes span ner and mrc ner, Text classification includes BERT text classification, including their Python training and c++ engineering.
dawoshi/minimind
【大模型】3小时完全从0训练一个仅有26M的小参数GPT,最低仅需2G显卡即可推理训练!
dawoshi/2023AIWIN_Competition
dawoshi/dawoshi
dawoshi/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
dawoshi/tensorrtllm_backend
The Triton TensorRT-LLM Backend