Skip to content

Conversation

[ghstack-poisoned]
HAOCHENYE added a commit that referenced this pull request Jan 29, 2026
…f16 in DecoderLayer

When FSDPModule is applied to DecoderLayer, it automatically converts position_embedding to bf16, which causes the forward pass output to differ from HuggingFace models beyond the acceptable threshold. This leads to test failures for Qwen VL. The impact on text-only components is currently uncertain.


ghstack-source-id: 7bff952
Pull-Request: #1462
HAOCHENYE added a commit to HAOCHENYE/xtuner that referenced this pull request Jan 29, 2026
…f16 in DecoderLayer

When FSDPModule is applied to DecoderLayer, it automatically converts position_embedding to bf16, which causes the forward pass output to differ from HuggingFace models beyond the acceptable threshold. This leads to test failures for Qwen VL. The impact on text-only components is currently uncertain.


ghstack-source-id: 7bff952
Pull-Request: InternLM#1462
[ghstack-poisoned]
HAOCHENYE added a commit that referenced this pull request Jan 29, 2026
…f16 in DecoderLayer

When FSDPModule is applied to DecoderLayer, it automatically converts position_embedding to bf16, which causes the forward pass output to differ from HuggingFace models beyond the acceptable threshold. This leads to test failures for Qwen VL. The impact on text-only components is currently uncertain.


ghstack-source-id: 8007bdb
Pull-Request: #1462
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants