bitsandbytes-foundation/bitsandbytes
Accessible large language models via k-bit quantization for PyTorch.
PythonMIT
Issues
- 3
Error when saving FSDP weights with cpu_offload=True [rank1]: AttributeError: 'Params4bit' object has no attribute 'absmax'
#1405 opened by PanagiotisFytas - 5
- 0
ROCm6.2 Error cublasLt Encountered Error!
#1428 opened by githust66 - 1
ARM Runners December 2024
#1390 opened by johnnynunez - 1
It seems that the current version of bitsandbytes is not compatible with my CUDA 12.4 library.
#1409 opened by Yukiiceeee - 8
Release v44 not available for Mac
#1378 opened by ACMCMC - 1
- 1
RuntimeError: Internal: could not parse ModelProto from /home/smusleh/miniGPT-Med/llama-2-7b-chat-hf/tokenizer.model
#1438 opened by smusleh - 0
CUDA setup failed on linux CUDA12.3 with version 0.45.0
#1440 opened by viki760 - 1
aarch64 whl in PyPi
#1437 opened by drikster80 - 1
python -m bitsandbytes ·
#1423 opened by gunnusravani - 1
Failed to import transformers.integrations.bitsandbytes because of the following error (look up to see its traceback): CUDA Setup failed despite GPU being available. Please run the following command to get more information: python -m bitsandbytes Inspect the output of the command and see if you can locate CUDA libraries. You might need to add them to your LD_LIBRARY_PATH. If you suspect a bug, please take the information from python -m bitsandbytes and open an issue at: https://github.com/TimDettmers/bitsandbytes/issues
#1387 opened by smillpine - 0
BnB load_in_8bit model violates attention mask
#1433 opened by poedator - 1
bitsandbytes-0.44.1.dev0-py3-none-macosx_13_1_arm64.whl is not a supported wheel on this platform.
#1406 opened by wuhongsheng - 4
CUDA Setup failed despite GPU being available
#1394 opened by carlxc911 - 2
CUDA Setup failed despite CUDA being available.
#1404 opened by N1ana7mi - 0
Rewrite `assert`s as exceptions
#1408 opened by akx - 0
Support for quantization of convolutional layers
#1414 opened by JohnnyRacer - 0
8-bit C-Optim optimizers
#1430 opened by odusseys - 1
- 1
'NoneType' object has no attribute 'cget_col_row_stats'
#1427 opened by githust66 - 1
Source Code for `cadam_8bit_blockwise_grad_bf16`
#1417 opened by ionutmodo - 0
- 4
- 1
Integration of BitNet into bitsandbytes
#1412 opened by ParagEkbote - 3
- 1
[Windows] [Error] Compile from source
#1398 opened by blap - 2
Support running on CPU
#1402 opened by fzyzcjy - 2
Paged optimizer resuming from checkpoint - attributeError: 'int' object has no attribute 'cpu'
#1381 opened by shivam15s - 0
- 0
Support
#1407 opened by Qarqor5555555 - 0
- 0
- 5
Questions about the details of LLM.int8
#1400 opened by bg51717 - 0
too large numeric difference with pytorch inference
#1396 opened by weixsong - 0
I ran a NF4 72B model in 2xA6000 using llamafactory
#1392 opened by charleswg - 0
CUDA Architecture 80+ Causing Incorrect Model Behavior with BitsAndBytes Quantization
#1391 opened by gunjunlee - 1
AdEMA NaN when loading from state_dict
#1382 opened by darius-lam - 0
Python 3.9 support broken in 0.44.0
#1376 opened by Benzhaomin - 0
Model architecture is modified when I use BitsAndBytesConfig with default params
#1371 opened by yunhao-tech - 2
- 2
cuda is available but import bnb error
#1355 opened by ZeroneBo - 4
Merge LoRA into 405B
#1359 opened by junzhang-zj - 1
RuntimeError: CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling cublasGemmEx
#1363 opened by LukeLIN-web - 0
Bug when using optimizer LAMB 32bits
#1350 opened by FrsECM - 1
Lion Optimizer With Triton Kernel
#1356 opened by lapp0 - 0
Model not able to quantize
#1354 opened by alielfilali01 - 0
Torch autograd support for dequantize methods
#1347 opened by yaldashbz - 1
Cannot load decoder.lm_head.weight when loading 4 bit quantized model using VisionEncoderDecoder.from_pretrained
#1343 opened by AditiJain14 - 0