From b4ed1fbf294f7922872dd31cebb9b37f3159af62 Mon Sep 17 00:00:00 2001 From: "Chang Liu (Enterprise Products)" <9713593+chang-l@users.noreply.github.com> Date: Mon, 17 Nov 2025 11:30:16 -0800 Subject: [PATCH] Indexer rope fix Signed-off-by: Chang Liu (Enterprise Products) <9713593+chang-l@users.noreply.github.com> --- tensorrt_llm/_torch/attention_backend/sparse/dsa.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/tensorrt_llm/_torch/attention_backend/sparse/dsa.py b/tensorrt_llm/_torch/attention_backend/sparse/dsa.py index 1d447a8557d..04e0e46c7a1 100644 --- a/tensorrt_llm/_torch/attention_backend/sparse/dsa.py +++ b/tensorrt_llm/_torch/attention_backend/sparse/dsa.py @@ -723,7 +723,8 @@ def __init__(self, self.rotary_emb = RotaryEmbedding( pos_embd_params.rope, head_dim=self.rope_dim, - is_neox=pos_embd_params.is_neox, + # RoPE in indexer is not interleaved + is_neox=True, ) self.softmax_scale = self.head_dim**-0.5