diff --git a/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md b/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md index 7f2e79d500..1c9a1872e9 100644 --- a/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md +++ b/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md @@ -3,13 +3,13 @@ title: Run a Phi-4-mini chatbot powered by ONNX Runtime weight: 2 overview: | - This Learning Path shows you how to use a 32-core Azure Dpls_v6 instance powered by an Arm Neoverse-N2 CPU to build a simple chatbot server that you can then use to provide a chatbot to serve a small number of concurrent users. + This Learning Path shows you how to use a 32-core Azure Dpls_v6 instance powered by an Arm Neoverse N2 CPU to build a simple chatbot that you can use to serve a small number of concurrent users. - This architecture is suitable for businesses looking to deploy the latest Generative AI technologies with RAG capabilities using their existing CPU compute capacity and deployment pipelines. + This architecture is suitable for deploying the latest Generative AI technologies with RAG capabilities using their existing CPU compute capacity and deployment pipelines. - The demo uses the ONNX runtime, which Arm has enhanced with its own Kleidi technologies. Further optimizations are achieved by using the smaller Phi-4-mini model, which has been optimized at INT4 quantization to minimize memory usage. + The demo uses the ONNX runtime, which Arm has integrated with KleidiAI. Further optimizations are achieved by using the smaller Phi-4-mini model, which has been optimized at INT4 quantization to minimize memory usage. - Chat with the chatbot LLM below to see the performance for yourself, and then follow the Learning Path to build your own Generative AI service on Arm Neoverse. + Chat with the LLM below to see the performance for yourself, and then follow the Learning Path to build your own Generative AI service on Arm Neoverse. demo_steps: diff --git a/content/learning-paths/servers-and-cloud-computing/onnx/_index.md b/content/learning-paths/servers-and-cloud-computing/onnx/_index.md index 37bdde2483..27dd6e766e 100644 --- a/content/learning-paths/servers-and-cloud-computing/onnx/_index.md +++ b/content/learning-paths/servers-and-cloud-computing/onnx/_index.md @@ -7,7 +7,7 @@ who_is_this_for: This is an advanced topic for developers, ML engineers, and clo learning_objectives: - Quantize and run the Phi-4-mini model with ONNX Runtime on Azure. - - Analyze performance on Arm Neoverse-N2 based Azure Cobalt 100 VMs. + - Analyze performance on Arm Neoverse N2 based Azure Cobalt 100 VMs. prerequisites: - An [Arm-based instance](/learning-paths/servers-and-cloud-computing/csp/) from an appropriate cloud service provider. This Learning Path has been tested on an Azure Cobalt 100 virtual machine.