Skip to content

Commit 591a913

Browse files
committed
[None][fix] fix post-merge issue raised by NVIDIA#5488
Signed-off-by: nv-guomingz <137257613+nv-guomingz@users.noreply.github.com>
1 parent d49374b commit 591a913

File tree

2 files changed

+2
-2
lines changed

2 files changed

+2
-2
lines changed

tests/unittest/llmapi/test_executor.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -399,7 +399,7 @@ def test_ZeroMqQueue_serialization_complicated_dataclass():
399399
TokenRangeRetentionConfig = tllm.KvCacheRetentionConfig.TokenRangeRetentionConfig
400400
kvcache_config = tllm.KvCacheRetentionConfig(
401401
[TokenRangeRetentionConfig(0, 2, 30, datetime.timedelta(seconds=30))],
402-
80)
402+
80, None, tllm.KvCacheTransferMode.DRAM, "test_dir")
403403

404404
sampling_params = SamplingParams(max_tokens=4,
405405
embedding_bias=torch.randn(2, 2))

tests/unittest/llmapi/test_llm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -357,7 +357,7 @@ def test_llm_with_kv_cache_retention_config():
357357
kv_cache_retention_config = KvCacheRetentionConfig([
358358
KvCacheRetentionConfig.TokenRangeRetentionConfig(
359359
0, 2, 30, datetime.timedelta(seconds=30))
360-
], 80)
360+
], 80, None, tllm.KvCacheTransferMode.DRAM, "test_dir")
361361

362362
llm = LLM(model=llama_model_path,
363363
kv_cache_config=global_kvcache_config,

0 commit comments

Comments
 (0)