/

/

Meet AirTen: the fastest audio real-time runtime period

/

/

Meet AirTen: the fastest audio real-time runtime period

Meet AirTen: the fastest audio real-time runtime period

Meet AirTen: the fastest audio real-time runtime period

/

Case Study

We’re thrilled to officially launch AirTen - ai-coustics’ purpose-built neural network runtime. Designed especially for real-time audio AI, AirTen delivers unmatched speed, safety, and portability.

And the best part? It’s packed into a runtime smaller than the average photo stored on your phone and exclusively powers the models in our SDK.

What is AirTen?

AirTen (short for AirTensors) is our custom runtime for neural network inference - the critical phase where models move from training to action, executing in real-world applications and devices. In audio AI every millisecond counts and inference speed isn't just a ‘nice-to-have’ - it’s essential.

So we built a new kind of engine. One that has:

  • A pure no_std Rust runtime

  • Zero dependencies, ensuring unmatched portability

  • A build tiny enough for microcontrollers, yet powerful enough for desktop and web

  • Integration into ai-coustics’ model delivery pipeline and available with our SDK

Why not use existing runtimes?

There are plenty of general purpose inference engines out there, but none optimized for real-time, resource-constrained audio environments. Here’s what we ran into:

  • They couldn’t guarantee consistent timing - which can lead to pops, clicks, and audio glitches

  • They used too much memory and processing power, making them hard to run on smaller devices

  • They were difficult to set up across different platforms

  • They were missing key features our models rely on

  • They acted like a black box - hard to understand and even harder to customize

AirTen changes the game by giving you complete control and making no compromises on size, speed, or stability.

AirTen: Key benefits

Image of the table listing key benefits of AirTen
  • Real-time safe: guaranteed sub-8ms inference at 48kHz audio rates, with no clicks/pops.

  • Lightning fast: up to 2x faster worst-case execution (WCET) vs ONNX Runtime.

  • Ultra lightweight: <1MB static runtime - over 100X smaller than typical engines.

  • Cross-platform: deploys to embedded, desktop, and browser with one simple command.

  • Memory-efficient: uses 80% less RAM than ONNX Runtime (6 MB vs 30 MB).

  • Rust safety: pure Rust avoids crashes, overflows, and threading issues.

  • Fully portable: runs anywhere Rust is supported - no custom toolchains or fragile build steps.

Real-world performance: How AirTen stacks up

We tested AirTen and here’s how it compares to ONNX, one of the most popular inference engines out there:

In summary:
AirTen is smaller, uses less memory, and runs twice as fast - perfect for devices where every millisecond and megabyte counts.

Keen to try it out?

AirTen is now available with ai-coustics SDK - perfect to pair with our real-time model families Quail and Sparrow. Get in touch to learn more.

Final logo

Bring real-time audio intelligence into your voice AI stack

Bring real-time audio intelligence into your voice AI stack