AudioShake Launches First-Ever SDK for Real-Time Sound Separation
AudioShake, the industry leader in AI-powered audio separation, today launched its first-ever Software Development Kit (SDK) for iOS/MacOS, Android, Windows, and Linux platforms. The SDK brings AudioShake’s state-of-the-art AI models directly to developers, enabling real-time speech cleanup, background noise removal, and instrument separation in streaming workflows and on edge devices.
Built for low-latency performance, the AudioShake’s SDK opens up new real-time applications across voice agents, transcription pipelines, music remixing tools, and mobile education apps—delivering clean, editable audio–separated by voice or individual instruments–instantly from any mix.
“Real-time sound isolation opens up vast new opportunities for developers working with real-time applications,” said Jessica Powell, CEO of AudioShake, whose technology is already used by major labels, broadcasters, and AI platforms. “With these SDKs, developers can deliver unmatched separation quality—whether powering live voice tools or unlocking interactive music features. We’re excited to see how these tools transform the way we experience sound.”
Powering Creativity in Music and Education
With real-time separation, AudioShake’s SDK gives music app developers and producers new creative superpowers:
- Music Mixing: Instantly isolate vocals and instruments from any song to enable remixing, karaoke, or fan engagement. The SDK powers music platforms including Tuned Global and MashApp—one of the first consumer music apps licensed by all major labels for stem-based remixing.
- Music Education: Enable more immersive practice by isolating individual instruments in any song. Backtrackit, a popular learning app, uses AudioShake to help musicians loop, slow down, and study complex parts.
- Songwriting & Mobile Recording: Real-time vocal isolation ensures studio-quality capture—even in noisy environments. Apps like VoCap (from the team behind AutoTune) and Voloco use AudioShake’s SDK to give songwriters clean vocals on the go.
Boosting Accuracy in Voice and Speech Workflows
In speech workflows, AudioShake’s SDK empowers developers to create cleaner voice inputs, improve transcription accuracy, and elevate audio quality across diverse use cases, including:
- Clean Voice Extraction: Seamlessly integrate into consumer apps and professional audio workflows for clearer and more precise voice inputs.
- Enhanced ASR Performance: Deliver cleaner speech inputs to improve automatic speech recognition, boosting transcription and captioning accuracy.
- Easy Localization Workflows: Feed cleaner dialogue tracks and transcriptions into translation pipelines to create localized content automatically for greater audience accessibility and reach.
In speech workflows, AudioShake empowers developers to feed clean audio into downstream pipelines—dramatically improving transcription, translation, and content localization:
- Voice Isolation for Transcription & Dubbing: Isolate speakers from crowd or background noise in real time—critical for captioning, ASR, and localization.
- Enhanced ASR Accuracy: Customers using AudioShake’s SDK have reported up to a 25% improvement in automatic speech recognition accuracy—translating into faster, more reliable pipelines.
- Live Captioning at Scale: AI-Media, which powers live captioning for sports and corporate events, uses the SDK to isolate speech from crowd and background noise—improving both speed and accuracy in multilingual broadcast environments.
Get Started
The AudioShake SDK is available now for iOS/MacOS, Android, Windows, and Linux. Developers can explore documentation, sample apps, and integration guides at developer.audioshake.ai.