diff --git a/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md b/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md index 4e88f92d4c..7f2e79d500 100644 --- a/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md +++ b/content/learning-paths/servers-and-cloud-computing/onnx/_demo.md @@ -24,10 +24,6 @@ terms_and_conditions: demo-terms-and-conditions.txt prismjs: true # enable prismjs rendering of code snippets -example_user_prompts: - - Prompt 1? - - Prompt 2? - rag_data_cutoff_date: 2025/01/17 diff --git a/content/learning-paths/servers-and-cloud-computing/onnx/setup.md b/content/learning-paths/servers-and-cloud-computing/onnx/setup.md index d28f75f19f..fe72af9ead 100644 --- a/content/learning-paths/servers-and-cloud-computing/onnx/setup.md +++ b/content/learning-paths/servers-and-cloud-computing/onnx/setup.md @@ -93,6 +93,7 @@ Clone and build the `onnxruntime-genai` repository: cd ~ cp ./onnxruntime/build/Linux/Release/install/include/onnxruntime/onnxruntime_float16.h ./onnxruntime/build/Linux/Release/install/include/onnxruntime_float16.h cp ./onnxruntime/build/Linux/Release/install/include/onnxruntime/onnxruntime_c_api.h ./onnxruntime/build/Linux/Release/install/include/onnxruntime_c_api.h +cp ./onnxruntime/build/Linux/Release/install/include/onnxruntime/onnxruntime_ep_c_api.h ./onnxruntime/build/Linux/Release/install/include/onnxruntime_ep_c_api.h git clone https://github.com/microsoft/onnxruntime-genai.git cd onnxruntime-genai python3 build.py --config Release --update --ort_home ../onnxruntime/build/Linux/Release/install diff --git a/themes/arm-design-system-hugo-theme/layouts/partials/demo-components/llm-chatbot/javascript--llm-chatbot.html b/themes/arm-design-system-hugo-theme/layouts/partials/demo-components/llm-chatbot/javascript--llm-chatbot.html index ea2f7343ee..6dda79922c 100644 --- a/themes/arm-design-system-hugo-theme/layouts/partials/demo-components/llm-chatbot/javascript--llm-chatbot.html +++ b/themes/arm-design-system-hugo-theme/layouts/partials/demo-components/llm-chatbot/javascript--llm-chatbot.html @@ -678,17 +678,17 @@ // Update ping & popup status if (data.cluster_utilization == 'normal') { - showPopupPostConnection('Connected to Arm Neoverse V2 based LLM! Start chatting now.',"success"); + showPopupPostConnection('Connected to Arm Neoverse based LLM! Start chatting now.',"success"); //ping_txt.textContent = `Ping: ${ping}` traffic_txt.textContent = 'Server traffic: Low' } else if (data.cluster_utilization == 'high') { - showPopupPostConnection('Connected to Arm Neoverse V2 based LLM! Traffic is high, delays may occur. Start chatting now.',"success"); + showPopupPostConnection('Connected to Arm Neoverse based LLM! Traffic is high, delays may occur. Start chatting now.',"success"); //ping_txt.textContent = `Ping: ${ping}` traffic_txt.textContent = `Server traffic: High` } else if (data.cluster_utilization == 'at-limit') { - showPopupPostConnection('Connected to Arm Neoverse V2 based LLM! Traffic is high, delays may occur. Start chatting now.',"warning"); + showPopupPostConnection('Connected to Arm Neoverse based LLM! Traffic is high, delays may occur. Start chatting now.',"warning"); //ping_txt.textContent = `Ping: ${ping}` traffic_txt.textContent = `Server traffic: High` }