Skip to content

Conversation

@erictang000
Copy link
Collaborator

@erictang000 erictang000 commented Jan 15, 2026

After this Megatron-Bridge PR: NVIDIA-NeMo/Megatron-Bridge#1766

We should update LoRA + Megatron to only export LoRA weights, and to enable LoRA on vllm with megatron. This should help with making weight syncing for LoRA more efficient, and help reduce trainer/inference mismatch for LoRA.

TODOs:

  • Make changes to allow vLLM latest (0.13.0) to work with SkyRL
  • Patch vLLM LoRA weight loading (LRUCacheWorkerLoRAManager - verl) to allow for in memory LoRA weight loading
  • provide instructions for installing vLLM from a nightly commit (vllm: 482914849cf9ce61d3e0dffaa35096bb34de58f5)

erictang000 added a commit that referenced this pull request Jan 17, 2026
…lm and mcore extras (#887)

Upgrading vllm to latest (minor changes).

Needed for #885 and should subsume #882

Upgrades transformer-engine from 2.9.0 -> 2.10.0 for the megatron
backend due to incompatibility with triton 3.5 (required by torch 2.9.0)

Keeps the vllm_engine.py path backwards compatible for versions of vllm
< 0.11.2 (important for flash-rl integration)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant