kangjiahui's Stars
ggerganov/llama.cpp
LLM inference in C/C++
labmlai/annotated_deep_learning_paper_implementations
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
onnx/onnx
Open standard for machine learning interoperability
ai-shifu/ChatALL
Concurrently chat with ChatGPT, Bing Chat, Bard, Alpaca, Vicuna, Claude, ChatGLM, MOSS, 讯飞星火, 文心一言 and more, discover the best answers
QwenLM/Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
divamgupta/diffusionbee-stable-diffusion-ui
Diffusion Bee is the easiest way to run Stable Diffusion locally on your M1 Mac. Comes with a one-click installer. No dependencies or technical knowledge needed.
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
Reamd7/notion-zh_CN
notion 中文化
makenotion/notion-sdk-js
Official Notion JavaScript Client
Lordog/dive-into-llms
《动手学大模型Dive into LLMs》系列编程实践教程
mymusise/ChatGLM-Tuning
基于ChatGLM-6B + LoRA的Fintune方案
DefTruth/lite.ai.toolkit
🛠 A lite C++ toolkit of 100+ Awesome AI models, support ORT, MNN, NCNN, TNN and TensorRT. 🎉🎉
open-mmlab/mmdeploy
OpenMMLab Model Deployment Framework
eProsima/Fast-DDS
The most complete DDS - Proven: Plenty of success cases. Looking for commercial support? Contact info@eprosima.com
quic/aimet
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
openai/finetune-transformer-lm
Code and model for the paper "Improving Language Understanding by Generative Pre-Training"
kwuking/TimeMixer
[ICLR 2024] Official implementation of "TimeMixer: Decomposable Multiscale Mixing for Time Series Forecasting"
zhiqwang/yolort
yolort is a runtime stack for yolov5 on specialized accelerators such as tensorrt, libtorch, onnxruntime, tvm and ncnn.
PacktPublishing/Modern-CMake-for-Cpp
Modern CMake for C++, published by Packt
xmba15/onnx_runtime_cpp
small c++ library to quickly deploy models using onnxruntime
fengxinjie/Transformer-OCR
ltkong218/FastFlowNet
FastFlowNet: A Lightweight Network for Fast Optical Flow Estimation (ICRA 2021)
intel/ros2_openvino_toolkit
kalfazed/tensorrt_starter
This repository give a guidline to learn CUDA and TensorRT from the beginning.
quic/efficient-transformers
This library empowers users to seamlessly port pretrained models and checkpoints on the HuggingFace (HF) hub (developed using HF transformers library) into inference-ready formats that run efficiently on Qualcomm Cloud AI 100 accelerators.
wangxb96/Awesome-EdgeAI
Resources of our survey paper "A Comprehensive Survey on AI Integration at the Edge: Techniques, Applications, and Challenges"
scaidermern/top-processes
Displays the top processes according to current CPU or memory usage
kalfazed/multi-thread-programming
This is a repository to practice multi-thread programming in C++