Skip to content

Commit 03db7c4

Browse files
authored
fix: fix parameter name in run-vllm (#93)
1 parent 95d7e20 commit 03db7c4

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

run-vllm.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -33,8 +33,8 @@ def __call__(
3333
m = params["_max_tokens"]
3434
kwargs["max_num_batched_tokens"] = m
3535
kwargs["max_model_len"] = min(m, model_max_tokens or m, model_seq_length or m)
36-
if kwargs["tensor_parallel_size"] > 0:
37-
tensor_parallel_size = kwargs["tensor_parallel_size"]
36+
if params["tensor_parallel_size"] > 0:
37+
tensor_parallel_size = params["tensor_parallel_size"]
3838
else:
3939
tensor_parallel_size = math.gcd(
4040
torch.cuda.device_count(),

0 commit comments

Comments
 (0)