Skip to content

feat: Implement 4-bit and 8-bit Quantized Convolutional Layers#1907

Open
OnePunchMonk wants to merge 1 commit intobitsandbytes-foundation:mainfrom
OnePunchMonk:feat/convQuantization
Open

feat: Implement 4-bit and 8-bit Quantized Convolutional Layers#1907
OnePunchMonk wants to merge 1 commit intobitsandbytes-foundation:mainfrom
OnePunchMonk:feat/convQuantization

Conversation

@OnePunchMonk
Copy link
Copy Markdown

This PR introduces support for 4-bit and 8-bit quantized convolutional layers as requested in #1414.

Included:

  • \Conv1d4bit, \Conv2d4bit, \Conv3d4bit\ with FP4 and NF4 types
  • \Conv1d8bitLt, \Conv2d8bitLt, \Conv3d8bitLt\
  • Complete test suite matching bitsandbytes standards
  • Resolves the missing support for vision and audio models relying natively on convolutions.

Fixes #1414

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Support for quantization of convolutional layers

1 participant