ci: exclude vllm_inference and megatron from nightly recipe CI#1554
Merged
pstjohn merged 1 commit intoNVIDIA:mainfrom Apr 18, 2026
Merged
Conversation
Contributor
|
Important Review skippedAuto reviews are disabled on this repository. Please check the settings in the CodeRabbit UI or the ⚙️ Run configurationConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro Plus Run ID: You can disable this status message by setting the Use the checkbox below for a quick retry:
✨ Finishing Touches🧪 Generate unit tests (beta)
Comment |
pstjohn
approved these changes
Apr 18, 2026
auto-merge was automatically disabled
April 18, 2026 14:27
Head branch was pushed to by a user without write access
auto-merge was automatically disabled
April 18, 2026 14:50
Head branch was pushed to by a user without write access
02bc46d to
814266f
Compare
814266f to
402d1d8
Compare
vllm_inference: vLLM 0.15.1 requires transformers<5 but the 26.03 container ships transformers 5.x. Excluded until container catches up. megatron recipes: already run via the dedicated mbridge-recipes workflow (unit-tests-mbridge-recipes.yaml). Remove duplicate runs from the recipes workflow to save CI resources. Both were already excluded from PR changed-files detection. This also excludes them from scheduled (nightly) ALL_DIRS enumeration. Signed-off-by: svc-bionemo <267129667+svc-bionemo@users.noreply.github.com>
402d1d8 to
5c6fa63
Compare
pstjohn
approved these changes
Apr 18, 2026
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Problem
The 26.03 container update brought
transformers==5.5.4, which breaks the vllm_inference recipe CI:vLLM 0.15.1 (current) + transformers 5.x
transformers < 5install_vllm.shinstalls vLLM (which pins transformers <5), then upgrades transformers back to 5.xnvidia/esm2_*Hub models have"tokenizer_class": "TokenizersBackend"in theirtokenizer_config.json, a class that only exists in transformers 5.x. If vLLM downgrades to <5, tokenizer loading failsvLLM 0.19.1 (latest, supports transformers 5.x)
transformers >= 5.5.1✅register_opaque_type(hoist=True)invllm/utils/torch_utils.py, gated behindis_torch_equal_or_newer("2.11.0.dev")2.11.0a0+nv26.03) matches that version check, but does not have the upstreamhoistparameter onregister_opaque_typeyetTypeError: register_opaque_type() got an unexpected keyword argument 'hoist'at import timeMegatron duplicates
eden_megatron,evo2_megatron) already have a dedicated CI workflow (unit-tests-mbridge-recipes.yaml) but were also running as duplicates in this recipes workflow on nightlyFix
Exclude both
vllm_inferenceandmegatronfrom theALL_DIRSnightly enumeration inunit-tests-recipes.yml:hoistAPI (enabling vLLM 0.19.1) or vLLM releases a version compatible with both the NGC torch build and transformers 5.xunit-tests-mbridge-recipes.yaml)Both were already excluded from PR
changed-filesdetection but scheduled runs bypass that filter.Related
HFInferenceParams.is_compileablefor transformers 5.x (merged)PreTrainedTokenizerFastinstead ofTokenizersBackendFailing CI run
https://github.com/NVIDIA/bionemo-framework/actions/runs/24601629903