Skip to content

Commit 1eca8fb

Browse files
authored
Fix rank handling in multi-node PP setup (#129)
1 parent bead072 commit 1eca8fb

File tree

1 file changed

+4
-1
lines changed

1 file changed

+4
-1
lines changed

unifiedcache/integration/vllm/uc_connector.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,6 +35,7 @@
3535
KVConnectorMetadata,
3636
KVConnectorRole,
3737
)
38+
from vllm.distributed.parallel_state import get_world_group
3839
from vllm.v1.core.kv_cache_utils import hash_request_tokens
3940
from vllm.v1.core.sched.output import SchedulerOutput
4041

@@ -121,7 +122,9 @@ def __init__(self, vllm_config: "VllmConfig", role: KVConnectorRole):
121122
self.use_layerwise = True
122123
self.kv_caches: dict[str, torch.Tensor] = {}
123124
self.total_tp_size = vllm_config.parallel_config.tensor_parallel_size
124-
self.rank = vllm_config.parallel_config.rank
125+
self.rank = (
126+
-1 if role == KVConnectorRole.SCHEDULER else get_world_group().local_rank
127+
)
125128
self.load_paras: dict[str, LoadPara] = {}
126129
self.save_paras: dict[str, SavePara] = {}
127130
# dump tasks record request -> block -> list[task]

0 commit comments

Comments
 (0)