Skip to content

[pull] master from Comfy-Org:master#644

Merged
pull[bot] merged 4 commits intocode:masterfrom
Comfy-Org:master
Feb 28, 2026
Merged

[pull] master from Comfy-Org:master#644
pull[bot] merged 4 commits intocode:masterfrom
Comfy-Org:master

Conversation

@pull
Copy link

@pull pull bot commented Feb 28, 2026

See Commits and Changes for more details.


Created by pull[bot] (v2.0.0-alpha.4)

Can you help keep this open source service alive? 💖 Please sponsor : )

ReinerBforartists and others added 4 commits February 27, 2026 19:03
Allow non QuantizedTensor layer to set want_requant to get the post lora
calculation stochastic cast down to the original input dtype.

This is then used by the legacy fp8 Linear implementation to set the
compute_dtype to the preferred lora dtype but then want_requant it back
down to fp8.

This fixes the issue with --fast fp8_matrix_mult is combined with
--fast dynamic_vram which doing a lora on an fp8_ non QT model.
@pull pull bot locked and limited conversation to collaborators Feb 28, 2026
@pull pull bot added the ⤵️ pull label Feb 28, 2026
@pull pull bot merged commit ac4412d into code:master Feb 28, 2026
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants