From 7c602007c440a0f90a1b8311639ffd89fba7732a Mon Sep 17 00:00:00 2001 From: Pedro Cuenca Date: Thu, 16 Oct 2025 19:16:56 +0200 Subject: [PATCH] Fix latex Related: #3129 --- gpt-oss-on-intel-xeon.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/gpt-oss-on-intel-xeon.md b/gpt-oss-on-intel-xeon.md index b5dc7e3093..9b51ac7a4d 100644 --- a/gpt-oss-on-intel-xeon.md +++ b/gpt-oss-on-intel-xeon.md @@ -104,7 +104,7 @@ We are in container now, do following steps. For each batch size we 1. Build a fixed-length 1024‑token left‑padded batch. 2. Run a single warm‑up round. -3. set `max_new_tokens=1024` and measure total latency, then get $throughput = (OUTPUT\\_TOKENS * batch\\_size) / total\\_latency$. +3. set `max_new_tokens=1024` and measure total latency, then get \\(throughput = (OUTPUT\_TOKENS * batch\_size) / total\_latency\\). Run `numactl -l python benchmark.py` for the following codes.