diff --git a/benchmarks/single_node/dsr1_fp8_mi355x.sh b/benchmarks/single_node/dsr1_fp8_mi355x.sh index d629437cf..1ce51ec87 100644 --- a/benchmarks/single_node/dsr1_fp8_mi355x.sh +++ b/benchmarks/single_node/dsr1_fp8_mi355x.sh @@ -44,7 +44,7 @@ python3 -m sglang.launch_server \ --trust-remote-code \ --chunked-prefill-size 196608 \ --mem-fraction-static 0.8 --disable-radix-cache \ - --num-continuous-decode-steps 4 \ + --num-continuous-decode-steps 8 \ --max-prefill-tokens 196608 \ --kv-cache-dtype fp8_e4m3 \ --cuda-graph-max-bs "$CONC" $EVAL_CONTEXT_ARGS > $SERVER_LOG 2>&1 & diff --git a/perf-changelog.yaml b/perf-changelog.yaml index 2bd2f025c..fef4fd11f 100644 --- a/perf-changelog.yaml +++ b/perf-changelog.yaml @@ -1,3 +1,9 @@ +- config-keys: + - dsr1-fp8-mi355x-sglang + description: + - "Tune --num-continuous-decode-steps 4 → 8 (+4.7% avg output throughput gain)" + pr-link: https://github.com/SemiAnalysisAI/InferenceX/pull/1109 + - config-keys: - 70b-fp8-*-vllm description: