Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 4 additions & 1 deletion .wordlist.txt
Original file line number Diff line number Diff line change
Expand Up @@ -4559,7 +4559,7 @@ qdisc
ras
rcu
regmap
rgerganovs
rgerganov's
rotocol
rpcgss
rpmh
Expand Down Expand Up @@ -4588,3 +4588,6 @@ vmscan
workqueue
xdp
xhci
JFR
conv
servlet
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ tools_software_languages_filter:
- Coding: 26
- Containerd: 1
- DetectNet: 1
- Docker: 9
- Docker: 10
- DSTREAM: 2
- Edge AI: 1
- Edge Impulse: 1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ TinyML is machine learning optimized to run on low-power, resource-constrained d

This Learning Path focuses on using TinyML models with virtualized Arm hardware to simulate real-world AI workloads on microcontrollers and NPUs.

If you're looking to build and train your own TinyML models, follow the [Introduction to TinyML on Arm using PyTorch and ExecuTorch](/embedded-and-microcontrollers/introduction-to-tinyml-on-arm/).
If you're looking to build and train your own TinyML models, follow the [Introduction to TinyML on Arm using PyTorch and ExecuTorch](/learning-paths/embedded-and-microcontrollers/introduction-to-tinyml-on-arm/).

## What is ExecuTorch?

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,7 +32,7 @@ operatingsystems:

tools_software_languages:
- Arm Virtual Hardware
- Fixed Virtual Platform (FVP)
- Fixed Virtual Platform
- Python
- PyTorch
- ExecuTorch
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -47,7 +47,7 @@ tools_software_languages_filter:
- ASP.NET Core: 2
- Assembly: 4
- assembly: 1
- Async-profiler: 1
- async-profiler: 1
- AWS: 1
- AWS CDK: 2
- AWS CodeBuild: 1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ layout: learningpathall
The instructions in this Learning Path are for any Arm server running Ubuntu 24.04.2 LTS. You will need at least three Arm server instances with at least 64 cores and 128GB of RAM to run this example. The instructions have been tested on an AWS Graviton4 c8g.16xlarge instance

## Overview
llama.cpp is a C++ library that enables efficient inference of LLaMA and similar large language models on CPUs, optimized for local and embedded environments. Just over a year ago from its publication date, rgerganovs RPC code was merged into llama.cpp, enabling distributed inference of large LLMs across multiple CPU-based machines—even when the models don’t fit into the memory of a single machine. In this learning path, we’ll explore how to run a 405B parameter model on Arm-based CPUs.
llama.cpp is a C++ library that enables efficient inference of LLaMA and similar large language models on CPUs, optimized for local and embedded environments. Just over a year ago from its publication date, rgerganov's RPC code was merged into llama.cpp, enabling distributed inference of large LLMs across multiple CPU-based machines—even when the models don’t fit into the memory of a single machine. In this learning path, we’ll explore how to run a 405B parameter model on Arm-based CPUs.

For the purposes of this demonstration, the following experimental setup will be used:
- Total number of instances: 3
Expand Down