diff --git a/examples/demo-apps/android/LlamaDemo/app/src/main/java/com/example/executorchllamademo/LlmBenchmarkRunner.java b/examples/demo-apps/android/LlamaDemo/app/src/main/java/com/example/executorchllamademo/LlmBenchmarkRunner.java index efb9ab21eae..7236fe317b0 100644 --- a/examples/demo-apps/android/LlamaDemo/app/src/main/java/com/example/executorchllamademo/LlmBenchmarkRunner.java +++ b/examples/demo-apps/android/LlamaDemo/app/src/main/java/com/example/executorchllamademo/LlmBenchmarkRunner.java @@ -105,15 +105,15 @@ public void onGenerationStopped() { results.add( new BenchmarkMetric( benchmarkModel, - "model_load_time(ns)", - mStatsDump.loadEnd - mStatsDump.loadStart, + "model_load_time(ms)", + (mStatsDump.loadEnd - mStatsDump.loadStart) * 1e-6, 0.0f)); // LLM generate time results.add( new BenchmarkMetric( benchmarkModel, - "generate_time(ns)", - mStatsDump.generateEnd - mStatsDump.generateStart, + "generate_time(ms)", + (mStatsDump.generateEnd - mStatsDump.generateStart) * 1e-6, 0.0f)); // Token per second results.add( diff --git a/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/BenchmarkActivity.java b/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/BenchmarkActivity.java index 4360bdcef85..2397bcfb851 100644 --- a/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/BenchmarkActivity.java +++ b/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/BenchmarkActivity.java @@ -56,7 +56,7 @@ protected void onCreate(Bundle savedInstanceState) { for (int i = 0; i < numIter; i++) { long start = System.nanoTime(); module.forward(); - long forwardMs = System.nanoTime() - start; + double forwardMs = (System.nanoTime() - start) * 1e-6; stats.latency.add(forwardMs); } @@ -68,13 +68,13 @@ protected void onCreate(Bundle savedInstanceState) { results.add( new BenchmarkMetric( benchmarkModel, - "avg_inference_latency(ns)", + "avg_inference_latency(ms)", stats.latency.stream().mapToDouble(l -> l).average().orElse(0.0f), 0.0f)); // Model load time results.add( new BenchmarkMetric( - benchmarkModel, "model_load_time(ns)", stats.loadEnd - stats.loadStart, 0.0f)); + benchmarkModel, "model_load_time(ms)", (stats.loadEnd - stats.loadStart) * 1e-6, 0.0f)); // Load status results.add(new BenchmarkMetric(benchmarkModel, "load_status", stats.errorCode, 0)); @@ -90,7 +90,7 @@ protected void onCreate(Bundle savedInstanceState) { class Stats { long loadStart; long loadEnd; - List latency = new ArrayList<>(); + List latency = new ArrayList<>(); int errorCode = 0; @Override diff --git a/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/LlmBenchmarkActivity.java b/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/LlmBenchmarkActivity.java index 04702562ba4..3bc38aad403 100644 --- a/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/LlmBenchmarkActivity.java +++ b/extension/android/benchmark/app/src/main/java/org/pytorch/minibench/LlmBenchmarkActivity.java @@ -97,15 +97,15 @@ public void onGenerationStopped() { results.add( new BenchmarkMetric( benchmarkModel, - "model_load_time(ns)", - mStatsInfo.loadEnd - mStatsInfo.loadStart, + "model_load_time(ms)", + (mStatsInfo.loadEnd - mStatsInfo.loadStart) * 1e-6, 0.0f)); // LLM generate time results.add( new BenchmarkMetric( benchmarkModel, - "generate_time(ns)", - mStatsInfo.generateEnd - mStatsInfo.generateStart, + "generate_time(ms)", + (mStatsInfo.generateEnd - mStatsInfo.generateStart) * 1e-6, 0.0f)); // Token per second results.add(