Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 9 additions & 7 deletions .github/configs/nvidia-master.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -3197,7 +3197,7 @@ gptoss-fp4-b200-vllm:
- { tp: 8, conc-start: 4, conc-end: 4 }

minimaxm2.5-fp8-b200-vllm:
image: vllm/vllm-openai:v0.17.0-cu130
image: vllm/vllm-openai:v0.18.0-cu130
model: MiniMaxAI/MiniMax-M2.5
model-prefix: minimaxm2.5
runner: b200
Expand All @@ -3208,18 +3208,20 @@ minimaxm2.5-fp8-b200-vllm:
- isl: 1024
osl: 1024
search-space:
- { tp: 2, conc-start: 4, conc-end: 64 }
- { tp: 4, conc-start: 4, conc-end: 64 }
- { tp: 2, conc-start: 4, conc-end: 256 }
- { tp: 4, conc-start: 4, conc-end: 256 }
- { tp: 4, ep: 4, conc-start: 64, conc-end: 64 }
- { tp: 8, conc-start: 4, conc-end: 8 }
- isl: 1024
osl: 8192
search-space:
- { tp: 2, conc-start: 4, conc-end: 64 }
- { tp: 4, conc-start: 4, conc-end: 64 }
- { tp: 2, conc-start: 4, conc-end: 256 }
- { tp: 4, conc-start: 4, conc-end: 256 }
- isl: 8192
osl: 1024
search-space:
- { tp: 2, conc-start: 4, conc-end: 64 }
- { tp: 4, conc-start: 4, conc-end: 64 }
- { tp: 2, conc-start: 4, conc-end: 256 }
- { tp: 4, conc-start: 4, conc-end: 256 }

gptoss-fp4-h100-vllm:
image: vllm/vllm-openai:v0.18.0
Expand Down
6 changes: 4 additions & 2 deletions benchmarks/single_node/minimaxm2.5_fp8_b200.sh
Original file line number Diff line number Diff line change
Expand Up @@ -24,8 +24,8 @@ hf download "$MODEL"
SERVER_LOG=/workspace/server.log
PORT=${PORT:-8888}

export VLLM_USE_FLASHINFER_MOE_FP8=0
export VLLM_MOE_USE_DEEP_GEMM=0
export VLLM_USE_DEEP_GEMM=0
export VLLM_FLASHINFER_ALLREDUCE_BACKEND=mnnvl

if [ "$EP_SIZE" -ge 1 ]; then
EP=" --enable-expert-parallel"
Expand All @@ -43,6 +43,8 @@ $EP \
--gpu-memory-utilization 0.95 \
--max-model-len $MAX_MODEL_LEN \
--block-size=32 \
--kv-cache-dtype fp8 \
--stream-interval 20 --no-enable-prefix-caching \
--trust-remote-code > $SERVER_LOG 2>&1 &

SERVER_PID=$!
Expand Down
8 changes: 8 additions & 0 deletions perf-changelog.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -1108,3 +1108,11 @@
description:
- "Update vLLM image from v0.15.1 to v0.18.0 for gptoss H100 and H200 configs"
pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/960


- config-keys:
- minimaxm2.5-fp8-b200-vllm
description:
- "Update vLLM image from v0.17.0 to v0.18.0 for MiniMax-M2.5 FP8 B200"
- "Add tp4 ep4 search-space entries (conc 32-256) for all seq-len configs"
pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/947
Loading