Skip to content

Commit e891bae

Browse files
authored
add Xeon GNR linux.24xl.gnr into runners list (#97)
Signed-off-by: Tsai, Louie <louie.tsai@intel.com>
1 parent ffdcb72 commit e891bae

File tree

2 files changed

+14
-1
lines changed

2 files changed

+14
-1
lines changed

.github/scripts/generate_vllm_benchmark_matrix.py

Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,13 +18,15 @@
1818
"linux.aws.h100",
1919
"linux.rocm.gpu.gfx942.1",
2020
"linux.24xl.spr-metal",
21+
"linux.24xl.gnr",
2122
"linux.dgx.b200",
2223
],
2324
# NB: There is no 2xH100 runner at the momement, so let's use the next one
2425
# in the list here which is 4xH100
2526
2: [
2627
"linux.aws.h100.4",
2728
"linux.rocm.gpu.gfx942.2",
29+
"linux.24xl.gnr",
2830
],
2931
4: [
3032
"linux.aws.h100.4",
@@ -50,6 +52,7 @@
5052
"linux.rocm.gpu.gfx942.4": "rocm",
5153
"linux.rocm.gpu.gfx942.8": "rocm",
5254
"linux.24xl.spr-metal": "cpu",
55+
"linux.24xl.gnr": "cpu",
5356
}
5457

5558
# All the different names vLLM uses to refer to their benchmark configs
@@ -72,6 +75,7 @@
7275
"meta-llama/Meta-Llama-3.1-70B-Instruct": [
7376
"linux.dgx.b200",
7477
"linux.rocm.gpu.gfx942", # TODO: Fail on ROCm
78+
"linux.24xl.gnr",
7579
],
7680
"mistralai/Mixtral-8x7B-Instruct-v0.1": [
7781
"linux.dgx.b200",
@@ -88,40 +92,49 @@
8892
"linux.aws.a100",
8993
"linux.aws.h100",
9094
"linux.rocm.gpu.gfx942", # TODO: Fail on ROCm
95+
"linux.24xl.gnr",
9196
],
9297
"google/gemma-3-27b-it": [
9398
"linux.aws.a100",
9499
"linux.aws.h100",
95100
"linux.rocm.gpu.gfx942", # TODO (huydhn): Fail on ROCm
101+
"linux.24xl.gnr",
96102
],
97103
"meta-llama/Llama-4-Scout-17B-16E-Instruct": [
98104
"linux.aws.a100",
99105
"linux.aws.h100",
100106
"linux.rocm.gpu.gfx942", # TODO: Fail on ROCm
107+
"linux.24xl.gnr",
101108
],
102109
"meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": [
103110
"linux.aws.a100",
104111
"linux.aws.h100",
105112
"linux.rocm.gpu.gfx942", # TODO (huydhn): Hang on ROCm
113+
"linux.24xl.gnr",
106114
],
107115
# Run gpt-oss on both H100 and B200
108116
"openai/gpt-oss-20b": [
109117
"linux.aws.a100",
118+
"linux.24xl.gnr",
110119
],
111120
"openai/gpt-oss-120b": [
112121
"linux.aws.a100",
122+
"linux.24xl.gnr",
113123
],
114124
# Deepseek can only run on B200
115125
"deepseek-ai/DeepSeek-V3.1": [
116126
"linux.aws.a100",
117127
"linux.aws.h100",
128+
"linux.24xl.gnr",
118129
],
119130
"deepseek-ai/DeepSeek-V3.2-Exp": [
120131
"linux.aws.a100",
121132
"linux.aws.h100",
133+
"linux.24xl.gnr",
122134
],
123135
"deepseek-ai/DeepSeek-R1": [
124136
"linux.aws.a100",
137+
"linux.24xl.gnr",
125138
"linux.aws.h100",
126139
],
127140
}

.github/workflows/vllm-benchmark.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ on:
2525
A comma-separated list of runners from .github/scripts/generate_vllm_benchmark_matrix.py to run the benchmark (optional, default to run everything)
2626
required: true
2727
type: string
28-
default: h100,rocm,spr,b200
28+
default: h100,rocm,spr,gnr,b200
2929
pull_request:
3030
paths:
3131
- .github/workflows/vllm-benchmark.yml

0 commit comments

Comments
 (0)