Pinned issues
Issues
- 3
[Feature Request] Support of `int8_dynamic_activation_int8_weight` with asymmetrically quantized weights
#1320 opened by sanchitintel - 0
[NF4] Various bugs in how NF4 handles `.to()` to move to a different device
#1310 opened by gau-nernst - 2
- 0
- 1
- 3
`attempting to run aten.abs.default, this is not supported` with latest torchtitan + torchao
#1313 opened by lchu-ibm - 4
Very large discrepancy in the quantized model's output compared to the original model when quantizing on CPU
#1335 opened by JohnnyRacer - 0
[Tracker] autoquant v2 tracker
#1215 opened by jerryzh168 - 0
Use unittest instead of pytest everywhere
#1321 opened by jerryzh168 - 2
`int8_dynamic_activation_int8_weight` uses zero-points for weight when activation is asymmetrically quantized
#1317 opened by sanchitintel - 8
pip install torchao cannot get latest versions (only 0.1 and 2 other version in the same level)
#1300 opened by moreAImore - 7
- 3
How to trigger torchao unit tests?
#1315 opened by goldhuang - 5
Why my QAT's convert doesn't work,still float32
#1290 opened by Xxxgrey - 1
[float8] DDP GPT1.5B Torch.compile dynamo error
#1308 opened by OrenLeung - 2
torch.compile(sync_float8_amax_and_scale_history) not working with triton latest main
#1311 opened by goldhuang - 5
[float8] FP8 GPT 1.5B Delayed Scaling 2x Slower than BF16
#1297 opened by OrenLeung - 6
[Sparsity] When sparsifying using Wanda on only Linear layers, PerChannelNormObserver() being added to embedding layers, leading to RuntimeError: linalg.vector_norm: Expected a floating point or complex tensor as input. Got Long
#1133 opened by agrawal-aka - 4
CPUoffloadOptimizer issues
#1209 opened by felipemello1 - 2
Can you export a quantized model to Torchscript?
#1293 opened by will-rice - 0
SEO not helping ao
#1298 opened by msaroufim - 2
- 7
ZeroPointDomain as an arguments
#1264 opened by airMeng - 1
AMD integration tracker
#1260 opened by jcaip - 3
Int4wo quantization raises exception for FP32/FP16 i.e. works only for BF16
#1267 opened by piotr-bazan-nv - 2
Sporadic Bad Alloc Failures on CI
#1229 opened by drisspg - 3
How to skip decomposition of dequantize_affine and quantize_affine custom ops in inductor?
#1230 opened by Nullkooland - 0
Low bit shampoo
#1257 opened by msaroufim - 16
Cannot run FSDP2 with low bit optim from AO
#1189 opened by nighting0le01 - 1
aarch64 jobs failing because conda is missing
#1246 opened by msaroufim - 0
- 15
torchao.float8 not working on PyTorch 2.4.1 and how does torchao handle FP8 autocast?
#1159 opened by zigzagcai - 1
[QAT] Low-bit FSDP all-gather for QAT
#1224 opened by gau-nernst - 1
[RFC] Sparsity Future Plans
#1136 opened by jcaip - 5
[low-bit optim] Add COAT optimizer
#1190 opened by gau-nernst - 4
Low Bit Optim Instability.
#1218 opened by nighting0le01 - 4
[feat request] PagedAdamw8bit for FSDP2
#1213 opened by felipemello1 - 2
- 3
Why swapping nn.Linear with FP8Linear brings throughput decrease for 7B LLAMA2-like model?
#1199 opened by zigzagcai - 1
- 7
- 3
Unable to save checkpoints when Use low bit optimizers with FSDP1 or FSDP2
#1185 opened by nighting0le01 - 0
- 6
FP6 Speed on A100 80g
#1181 opened by shihaobai - 1
[sparse] Add GPTQ support for sparse-marlin
#1134 opened by jcaip - 0
- 1
2 sparse tests failing on main nightly job
#1158 opened by msaroufim - 3
cannot import name 'quantize_' from 'torchao.quantization' (C:\ai\Meissonic\venv\lib\site-packages\torchao\quantization\__init__.py)
#1146 opened by nitinmukesh - 0
- 3
Fail to reproduce benchmark results
#1135 opened by ThisisBillhe