FancyChen21's Stars
hwchase17/langchain
⚡ Building applications with LLMs through composability ⚡
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
AlexeyAB/darknet
YOLOv4 / Scaled-YOLOv4 / YOLO - Neural Networks for Object Detection (Windows and Linux version of Darknet )
apache/tvm
Open deep learning compiler stack for cpu, gpu and specialized accelerators
stanfordnlp/GloVe
Software in C and data files for the popular GloVe model for distributed word representations, a.k.a. word vectors or embeddings
facebookresearch/metaseq
Repo for external large-scale work
wenda-LLM/wenda
闻达:一个LLM调用平台。目标为针对特定环境的高效内容生成,同时考虑个人和中小企业的计算资源局限性,以及知识安全和私密性问题
kuangliu/pytorch-cifar
95.47% on CIFAR10 with PyTorch
Oneflow-Inc/oneflow
OneFlow is a deep learning framework designed to be user-friendly, scalable and efficient.
IntelLabs/distiller
Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distiller
hibayesian/awesome-automl-papers
A curated list of automated machine learning papers, articles, tutorials, slides and projects
FlagAI-Open/FlagAI
FlagAI (Fast LArge-scale General AI models) is a fast, easy-to-use and extensible toolkit for large-scale model.
THUDM/GLM
GLM (General Language Model)
lucasjinreal/yolov7_d2
🔥🔥🔥🔥 (Earlier YOLOv7 not official one) YOLO with Transformers and Instance Segmentation, with TensorRT acceleration! 🔥🔥🔥
alpa-projects/alpa
Training and serving large-scale neural networks with auto parallelization.
huawei-noah/Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
D-X-Y/Awesome-AutoDL
Automated Deep Learning: Neural Architecture Search Is Not the End (a curated list of AutoDL resources and an in-depth analysis)
hunglc007/tensorflow-yolov4-tflite
YOLOv4, YOLOv4-tiny, YOLOv3, YOLOv3-tiny Implemented in Tensorflow 2.0, Android. Convert YOLO v4 .weights tensorflow, tensorrt and tflite
666DZY666/micronet
micronet, a model compression and deploy lib. compression: 1、quantization: quantization-aware-training(QAT), High-Bit(>2b)(DoReFa/Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference)、Low-Bit(≤2b)/Ternary and Binary(TWN/BNN/XNOR-Net); post-training-quantization(PTQ), 8-bit(tensorrt); 2、 pruning: normal、regular and group convolutional channel pruning; 3、 group convolution structure; 4、batch-normalization fuse for quantization. deploy: tensorrt, fp32/fp16/int8(ptq-calibration)、op-adapt(upsample)、dynamic_shape
BBuf/tvm_mlir_learn
compiler learning resources collect.
PaddlePaddle/PaddleSlim
PaddleSlim is an open-source library for deep model compression and architecture search.
ymm-tech/thresh
Flutter-based mobile cross-platform dynamic UI framework
mit-han-lab/amc
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
inisis/brocolli
Everything in Torch Fx
leimao/Frozen-Graph-TensorFlow
Save, Load Frozen Graph and Run Inference From Frozen Graph in TensorFlow 1.x and 2.x
HankYe/PAGCP
PAGCP for the compression of YOLOv5
Wulingtian/yolov5_onnx2caffe
yolov5 onnx caffe
facerless/yolov5-tensorflow
tsingmicro-toolchain/ts.knight-modelzoo
tsingmicro AI model zoo
dongkuanx27/SparseBERT
(SparseBERT) Rethinking Network Pruning -- under the Pre-train and Fine-tune Paradigm (NAACL'21)