Skip to content

Quantization support for GroupedTensor: MXFP8 #2450

@ptrendx

Description

@ptrendx

Implement quantization support for the GroupedTensor type for MXFP8 format.
The needed modifications to the existing kernel:

  • ignore padding in the allocation

Metadata

Metadata

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions