Skip to content

Commit a43cf9e

Browse files
chang-llfr-0531
authored andcommitted
Address comment
Signed-off-by: Chang Liu (Enterprise Products) <9713593+chang-l@users.noreply.github.com> Signed-off-by: Fanrong Li <23290157+lfr-0531@users.noreply.github.com>
1 parent f68ff81 commit a43cf9e

File tree

2 files changed

+1
-4
lines changed

2 files changed

+1
-4
lines changed

tensorrt_llm/_torch/attention_backend/sparse/dsa.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1114,7 +1114,7 @@ def forward(self, qr: torch.Tensor, hidden_states: torch.Tensor,
11141114
weights.to(torch.float32))
11151115

11161116

1117-
class DSATrtllmAttention(TrtllmAttention, nn.Module):
1117+
class DSATrtllmAttention(TrtllmAttention):
11181118
Metadata = DSAtrtllmAttentionMetadata
11191119

11201120
def __init__(
@@ -1151,7 +1151,6 @@ def __init__(
11511151
skip_create_weights_in_init=skip_create_weights_in_init,
11521152
attention_chunk_size=attention_chunk_size,
11531153
**kwargs)
1154-
nn.Module.__init__(self)
11551154

11561155
self.indexer = Indexer(quant_config, pos_embd_params, mla_params,
11571156
skip_create_weights_in_init,

tensorrt_llm/_torch/models/modeling_deepseekv3.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -383,8 +383,6 @@ def split_kv_b_proj(kv_b_proj: torch.Tensor,
383383
continue
384384
elif names[-1] == "next_layer_layernorm":
385385
continue
386-
elif names[-1] == "indexer" or 'mqa.indexer' in name:
387-
continue
388386
else:
389387
module_weights = filter_weights(name, weights)
390388
if hasattr(module, 'load_weights'):

0 commit comments

Comments
 (0)