Open
Conversation
There was a problem hiding this comment.
Cursor Bugbot has reviewed your changes and found 1 potential issue.
Bugbot Autofix is OFF. To automatically fix reported issues with cloud agents, enable autofix in the Cursor dashboard.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.

Doing this to support quack rms norm and custom selectice AC targets. Also in general cleaner to have custom paths for models we care about and serve even if non-MoE. Especially as we later adopt fp8/fp4 kernels to use
Note
Medium Risk
Adds new custom model implementations and changes model-config selection for Qwen3.5, which can affect training/inference correctness and attention masking/position-id behavior for these models. Main risk is regressions in Qwen3/Qwen3.5 loading and attention backends (SDPA/Flash/ring attention) rather than broader system impact.
Overview
Adds custom PrimeRL dense implementations for
Qwen3and text-onlyQwen3.5, and wires them intoAutoModelForCausalLMPrimeRLsoimpl=custom/auto-selection can instantiate these models.Updates model loading to force Qwen3.5 text-only config when not doing VLM training (switching from composite config to
text_configwhile preserving_attn_implementationand_name_or_path). Extendssubstitute_ring_attnto patch ring-attention_compute_attentionfor the new Qwen3/Qwen3.5 FlashAttention classes.Written by Cursor Bugbot for commit 7d59e25. This will update automatically on new commits. Configure here.