songhan
Song Han is an associate professor at MIT EECS and distinguished scientist at NVIDIA. His research interest is efficient AI computing.
MIT, NVIDIA
songhan's Stars
home-assistant/core
:house_with_garden: Open source home automation that puts local control and privacy first.
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
mit-han-lab/streaming-llm
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
gaogaotiantian/viztracer
VizTracer is a low-overhead logging/debugging/profiling tool that can trace and visualize your python code execution.
mit-han-lab/bevfusion
[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation
mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
mit-han-lab/temporal-shift-module
[ICCV 2019] TSM: Temporal Shift Module for Efficient Video Understanding
mit-han-lab/once-for-all
[ICLR 2020] Once for All: Train One Network and Specialize it for Efficient Deployment
mit-han-lab/efficientvit
EfficientViT is a new family of vision models for efficient high-resolution vision.
intel/intel-extension-for-pytorch
A Python package for extending the official PyTorch that can easily obtain performance on Intel platform
mit-han-lab/data-efficient-gans
[NeurIPS 2020] Differentiable Augmentation for Data-Efficient GAN Training
mit-han-lab/torchquantum
A PyTorch-based framework for Quantum Classical Simulation, Quantum Machine Learning, Quantum Neural Networks, Parameterized Quantum Circuits with support for easy deployments on real quantum computers.
mit-han-lab/torchsparse
[MICRO'23, MLSys'22] TorchSparse: Efficient Training and Inference Framework for Sparse Convolution on GPUs.
mit-han-lab/gan-compression
[CVPR 2020] GAN Compression: Efficient Architectures for Interactive Conditional GANs
mit-han-lab/smoothquant
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Efficient-Large-Model/VILA
VILA - a multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)
mit-han-lab/anycost-gan
[CVPR 2021] Anycost GANs for Interactive Image Synthesis and Editing
mit-han-lab/tinyengine
[NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep Learning; [NeurIPS 2022] MCUNetV3: On-Device Training Under 256KB Memory
mit-han-lab/tinyml
mit-han-lab/pvcnn
[NeurIPS 2019, Spotlight] Point-Voxel CNN for Efficient 3D Deep Learning
mit-han-lab/TinyChatEngine
TinyChatEngine: On-Device LLM Inference Library
mit-han-lab/lite-transformer
[ICLR 2020] Lite Transformer with Long-Short Range Attention
mit-han-lab/spvnas
[ECCV 2020] Searching Efficient 3D Architectures with Sparse Point-Voxel Convolution
mit-han-lab/amc
[ECCV 2018] AMC: AutoML for Model Compression and Acceleration on Mobile Devices
mit-han-lab/tiny-training
On-Device Training Under 256KB Memory [NeurIPS'22]
mit-han-lab/offsite-tuning
Offsite-Tuning: Transfer Learning without Full Model
mit-han-lab/hardware-aware-transformers
[ACL'20] HAT: Hardware-Aware Transformers for Efficient Natural Language Processing
mit-han-lab/inter-operator-scheduler
[MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration
mit-han-lab/parallel-computing-tutorial
mit-han-lab/tinychat-tutorial