Pinned Repositories
continue
⏩ Continue is the leading open-source AI code assistant. You can connect any models and any context to build custom autocomplete and chat experiences inside VS Code and JetBrains
eidy-web
llama-hub
A library of data loaders for LLMs made by the community -- to be used with GPT Index and/or LangChain
seacrowd-eval
Seacrowd eval base-code for ThaiLLM-Leaderboard
TensorRT-Incubator
Experimental projects related to TensorRT
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
deploy-ai-model
Float16-example
matichon-vultureprime's Repositories
matichon-vultureprime/eidy-web
matichon-vultureprime/continue
⏩ Continue is the leading open-source AI code assistant. You can connect any models and any context to build custom autocomplete and chat experiences inside VS Code and JetBrains
matichon-vultureprime/llama-hub
A library of data loaders for LLMs made by the community -- to be used with GPT Index and/or LangChain
matichon-vultureprime/seacrowd-eval
Seacrowd eval base-code for ThaiLLM-Leaderboard
matichon-vultureprime/TensorRT-Incubator
Experimental projects related to TensorRT
matichon-vultureprime/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.