Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
25 commits
Select commit Hold shift + click to select a range
40979a3
update to use mixed_precision_wrapper api
yaoyu-33 Oct 14, 2025
55923a3
Update refit_verifier.py
yaoyu-33 Oct 16, 2025
6c0353c
Merge remote-tracking branch 'origin/main' into yuya/update-to-use-mb…
yaoyu-33 Oct 21, 2025
0f93ad0
Update Megatron-Bridge submodule to latest main
yaoyu-33 Oct 21, 2025
371e458
Update Megatron-LM submodule to 85ca99106990c731887e6526d7a0f498911526d6
yaoyu-33 Oct 21, 2025
464934e
feature: additional kl metrics
ZhiyuLi-Nvidia Oct 24, 2025
afbfa41
Update nemo_rl/algorithms/loss_functions.py
ZhiyuLi-Nvidia Oct 24, 2025
9d118af
update naming
ZhiyuLi-Nvidia Oct 24, 2025
fc33e3b
Merge branch 'main' into yuya/update-to-use-mbridge-tot
yaoyu-33 Oct 24, 2025
710005a
update setup.py
yaoyu-33 Oct 24, 2025
e9a3e46
Merge branch 'yuya/update-to-use-mbridge-tot' of https://github.com/N…
yaoyu-33 Oct 24, 2025
0c18360
Merge commit 'refs/pull/1358/head' of github.com:NVIDIA-NeMo/RL
ashors1 Oct 27, 2025
3352cb9
update links in doc
ZhiyuLi-Nvidia Oct 27, 2025
d63f8f2
Merge commit 'refs/pull/1420/head' of github.com:NVIDIA-NeMo/RL
ashors1 Oct 27, 2025
97e2e20
updates
ashors1 Oct 28, 2025
1c2e019
update megatron bridge and MLM
ashors1 Oct 28, 2025
a1ce5df
update mlm
ashors1 Oct 31, 2025
8e42282
update megatron bridge
ashors1 Oct 31, 2025
224ff4a
clean up MLM branch
ashors1 Oct 31, 2025
69787ca
update mbridge
ashors1 Oct 31, 2025
9c9711a
Merge branch 'main' of github.com:NVIDIA-NeMo/RL into ashors/gpt-oss-tot
ashors1 Oct 31, 2025
d9b0e1c
cherry-pick fix for dependency issues and update uv.lock
terrykong Oct 28, 2025
83c77bf
fixes
ashors1 Oct 31, 2025
aaa1c12
update MLM
ashors1 Oct 31, 2025
0f7e8e8
Merge branch 'main' of github.com:NVIDIA-NeMo/RL into ashors/gpt-oss-tot
ashors1 Nov 4, 2025
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .gitmodules
Original file line number Diff line number Diff line change
@@ -1,12 +1,12 @@
[submodule "3rdparty/Megatron-LM"]
path = 3rdparty/Megatron-LM-workspace/Megatron-LM
url = https://github.com/terrykong/Megatron-LM.git
branch = yuya/nemo-rl-use-dev
branch = ashors/dev-with-gpt-oss
shallow = true
[submodule "3rdparty/Megatron-Bridge"]
path = 3rdparty/Megatron-Bridge-workspace/Megatron-Bridge
url = https://github.com/NVIDIA-NeMo/Megatron-Bridge.git
branch = main
branch = ashors/gpt-oss-tot
shallow = true
[submodule "3rdparty/Automodel-workspace/Automodel"]
path = 3rdparty/Automodel-workspace/Automodel
Expand Down
6 changes: 6 additions & 0 deletions nemo_rl/models/generation/vllm/vllm_worker.py
Original file line number Diff line number Diff line change
Expand Up @@ -132,6 +132,12 @@ def __init__(
self.cfg = config

self.model_name = self.cfg["model_name"]
## use the bf16 version of the model rather than the quantized version
## megatron --> hf export is done in bf16 so this ensures the vllm
## model is compatible with megatron
if "openai/gpt-oss" in self.model_name:
size = self.model_name.split("-")[-1]
self.model_name = f"unsloth/gpt-oss-{size}-BF16"
self.tensor_parallel_size = self.cfg["vllm_cfg"]["tensor_parallel_size"]
self.pipeline_parallel_size = self.cfg["vllm_cfg"]["pipeline_parallel_size"]
self.expert_parallel_size = self.cfg["vllm_cfg"]["expert_parallel_size"]
Expand Down
Loading
Loading