loader image

Our new Developer Platform and API prices are live!

Introducing Finch 2: the updated ai|coustics audio enhancement model

Introducing Lark 2.

We’re delighted to announce the launch of Finch 2 – the next generation of our signature voice isolation model.

An upgrade and improvement on our previous Finch model, Finch 2 is now available across our API and web app. With state-of-the-art improvements in clarity, robustness, and realism, Finch 2 sets a new standard for subtractive speech enhancement.

Key takeaways

  • Finch 2 replaces our legacy Finch model with improved de-noising and de-reverb, and other exciting new upgrades
  • Along with technical boosts, Finch 2 is also more energy-efficient
  • Finch 2 pairs well with our existing Lark model for a full suite of audio enhancement tools.

What’s new in Finch 2?

Finch 2 replaces our legacy Finch model with substantial upgrades across the board. Here’s what you can expect:

  • Improved de-noising and de-reverberation
  • More robust in adverse conditions, including loud noise, strong reverb, speaker to mic distance and more
  • A more natural sound with better preservation of speaker identity
  • Significantly less AI processing artifacts
  • Reduced size, with faster inference and less electricity consumption

Like the original Finch model, Finch 2 is a subtractive model. That means that it removes information from an audio signal. For example, if there’s a lot of background noise, a subtractive machine learning model can help remove the irrelevant information from the signal. So you’ll get the speech of the person you’re interviewing, but not someone coughing or a car going past behind you.

We also have a reconstructive model called Lark, which is still available as usual in the API and web app. A reconstructive model is based on generative AI and doesn’t just subtract unwanted audio; it also repairs wanted audio to improve it. When a signal gets too mixed (i.e. a car horn over your interviewee’s voice), it can use its machine learning database to reconstruct what your interviewee was saying exactly, in their own voice

How does it compare to other voice-isolation tools?

We benchmarked our Finch 2 model against other leading tools using the widely recognized DAPS dataset. The results speak for themselves:

Finch 2 benchmark.

The DAPS (Device and Produced Speech) dataset is an open-source benchmark designed to test how well speech enhancement models can transform real-world, noisy and reverberant recordings into clean speech.

We evaluate each model using two industry-standard metrics:

  1. PESQ (Perceptual Evaluation of Speech Quality)
    Measures how closely the enhanced audio matches the original clean reference.
  2. SigMOS (Signal Mean Opinion Score)
    A no-reference metric that simulates human subjective listening tests – it measures perceived quality, not just similarity to the original and Finch 2 comes in first place

Should I use Finch 2 or Lark?

Finch 2 and Lark are powerful audio enhancement models ideal for different contexts.

Choose Finch 2 if:

  • Your voice recording is already good, and you just want to isolate it, for example separating a voice track from a movie or song
  • You’ve recorded in tough conditions with very strong background noise, lots of reverb, distant speakers and more
  • Your first priority is preserving the sound of the original voice

Choose Lark if:

  • You want to restore an old recording to make it sound like it was recorded with modern studio equipment
  • You want to add back high frequencies in bandwidth limited recordings
  • You want to fix strong compression, for example in Zoom or phone call recordings
  • You want to create a full and professional sound for your voice

Still can’t decide? The good news is you can test out both models in our web app and API, right now. 

Latest updates

Fixing the audio input for voice agents

Voice agents are revolutionising the way we interact with technology – but they can only perform as well as the audio they receive. These systems are built on a complex stack: voice capture, speech recognition (ASR), reasoning (LLMs) and text-to-speech (TTS). While each layer has improved dramatically, one foundational element remains critically underserved and has the potential to break the

Read More
"Introducing our new model: Lark 2", a graphic of an origami bird and ai-coustics logo

Announcing Lark 2: the next generation of reconstructive speech enhancement

Fans of Lark, rejoice: Lark 2 is here. Bolder, better, and stronger than ever, Lark 2 is our most advanced reconstructive speech enhancement model yet. Lark 2, like its predecessor, is built with our speciality reconstructive AI technology which goes beyond just isolating speech to repair existing speech and restore lost information – all while preserving the authentic human voice

Read More

Meet AirTen: the fastest audio real-time runtime period

We’re thrilled to officially launch AirTen – ai|coustics’ purpose-built neural network runtime. Designed especially for real-time audio AI, AirTen delivers unmatched speed, safety, and portability. And the best part? It’s packed into a runtime smaller than the average photo stored on your phone and exclusively powers the models in our SDK. What is AirTen? AirTen (short for AirTensors) is our

Read More

Ready to embrace the power of Voice AI?

Authentic human voices. Studio-quality sound. Real-time capacity. Automated workflows. It starts here.