KastanDay/LLM-Distributed-Quantization
Accelerating multi-node Large Language Model training with per-layer selective quantization (FP32 -> FP16) of the transformer architecture.
PythonApache-2.0
Issues
- 0
Switch from Colossal Launcher to torchrun.
#1 opened by KastanDay