Filter

Ethos-U85 NPU Applications with TOSA Model Explorer: Exploring Next-Gen Edge AI Inference

TrendingRecently Added

Description

Why is this important?

The Arm Ethos-U85 NPU represents a major leap in bringing heavy inference to constrained embedded systems. With its full transformer operator support, expanded MAC throughput, and native TOSA compatibility, the Ethos-U85 enables developers to deploy models and workloads that were previously too intensive for MCU-class devices.

This project challenges you to explore the boundaries of what’s possible on Ethos-U85. The goal is to demonstrate inference performance and model complexity that is now achievable due to the architectural improvements and transformer acceleration capabilities of the Ethos-U85.

Ethos-U85 Launch

Project Summary

Using hardware such as the Alif Ensemble E4/E6/E8 DevKits (all include Ethos-U85), your task is to design and benchmark an advanced edge inference application that exploits the Ethos-U85’s compute and transformer capabilities.

You can utilise the Arm Fixed Virtual Platform Corstone-320 to prototype and test your application functionally, without access to Alif hardware. You can use this to prove functional correctness - and can then later test performance on actual silicon. We are interested to see projects both in simulation, and on final hardware.

Your project should include:

Model Deployment and Optimization
Select a computationally intensive model — ideally transformer-based or multi-branch convolutional — and deploy it on the Ethos-U85 using:

  • Model Explorer to inspect models and identify problem layers that reduce optimal delegation to the Ethos-U backend
  • The Vela compiler for optimization.

These tools can be used to:

  • Convert and visualize model graphs in TOSA format.
  • Identify unsupported operators.

Application Demonstration Implement a working example that highlights the Ethos-U85’s strengths in real-world inference. Possible categories include:

  • Transformers on Edge: lightweight BERT, ViT, or audio transformers (e.g. speech or sound event classification).
  • High-resolution Vision: semantic segmentation, object detection on large input sizes, or multi-head perception networks.
  • Multi-modal Fusion: combining audio, image, or sensor streams for contextual understanding.

Analysis and Benchmarking Report quantitative results on:

  • Inference latency, throughput (FPS or tokens/s), and memory footprint.
  • Power efficiency under load (optional).
  • Comparative performance versus Ethos-U55/U65 (use available benchmarks for reference or utilise the other Ethos-U NPUs provided in the Alif DevKits).

What kind of projects should you target?

To clearly demonstrate the leap from Ethos-U55/U65 to U85, choose projects that meet at least one of the following criteria:

  • Transformer-heavy architectures: e.g. attention blocks, transformer encoders, ViTs, or hybrid CNN+transformer models.
  • High-resolution or multi-branch networks: models with high input dimensionality or multiple processing paths that saturate NPU throughput.
  • Dense post-processing or large fully connected layers: cases where U55/U65 memory limits or MAC bandwidth previously restricted performance.
  • Multi-modal pipelines: combining multiple sensor inputs (e.g. image + IMU + audio) where the NPU must maintain concurrency or shared intermediate representations.

The Ethos-U85 is ideal for projects where model performance is constrained by attention layers, large activations, or operator types that previously required fallback to the CPU. Use the Ethos-U85 to eliminate those fallbacks and achieve full-NPU execution of advanced topologies.

What will you use?

You should be familiar with, or willing to learn about:

  • Programming: Python, C/C++
  • ExecuTorch or LiteRT
  • Techniques for optimising AI models for the edge (quantization, pruning, etc.)
  • Optimization Tools:
    • Model Explorer with TOSA adapter (and PTE adapter for ExecuTorch)
    • Vela compiler for Ethos-U
  • Bare-metal or RTOS (e.g., Zephyr)

Resources from Arm and our partners

Support Level

This project is designed to be self-serve but comes with opportunity of some community support from Arm Ambassadors, who are part of the Arm Developer program. If you are not already part of our program, click here to join.

Benefits

Standout project contributions to the community will earn digital badges. These badges can support CV or resumé building and demonstrate earned recognition.

To receive the benefits, you must show us your project through our online form. Please do not include any confidential information in your contribution. Additionally if you are affiliated with an academic institution, please ensure you have the right to share your material.