Skip to content

bitsandbytes-foundation/bitsandbytes

Folders and files

NameName
Last commit message
Last commit date

Latest commit

e1f515c · Mar 19, 2025
Mar 13, 2025
Feb 4, 2024
Dec 5, 2024
Mar 7, 2025
Jan 14, 2025
Mar 19, 2025
Mar 13, 2024
Aug 22, 2024
Mar 13, 2024
Dec 5, 2024
Feb 4, 2024
Apr 11, 2024
Dec 5, 2024
Oct 23, 2024
Jan 23, 2025
Jan 22, 2025
Oct 6, 2021
May 7, 2024
Oct 6, 2021
Oct 6, 2021
Feb 20, 2025
Oct 23, 2024
Mar 13, 2024
Feb 5, 2024
Feb 5, 2024
Feb 4, 2024
Dec 2, 2024
Feb 19, 2025
Feb 25, 2025
Feb 24, 2025

Repository files navigation

bitsandbytes

Downloads Downloads Downloads

The bitsandbytes library is a lightweight Python wrapper around CUDA custom functions, in particular 8-bit optimizers, matrix multiplication (LLM.int8()), and 8 & 4-bit quantization functions.

The library includes quantization primitives for 8-bit & 4-bit operations, through bitsandbytes.nn.Linear8bitLt and bitsandbytes.nn.Linear4bit and 8-bit optimizers through bitsandbytes.optim module.

There are ongoing efforts to support further hardware backends, i.e. Intel CPU + GPU, AMD GPU, Apple Silicon, hopefully NPU.

Please head to the official documentation page:

https://huggingface.co/docs/bitsandbytes/main

License

bitsandbytes is MIT licensed.

We thank Fabio Cannizzo for his work on FastBinarySearch which we use for CPU quantization.