Just ran Demucs completely locally on my system (RX 6700 XT / 16 GB RAM).

Demucs is an open source AI model for music source separation, developed by Meta. It can split a full song into individual stems like vocals, drums, bass, and other instruments, making it useful for remixing, transcription, and audio analysis.

Test track: Fear of the Dark by Iron Maiden
(https://www.youtube.com/watch?v=bePCRKGUwAY)

Setup:

- Demucs installed via pip
- Model: htdemucs (default)
- Input converted to WAV using ffmpeg
- GPU acceleration via ROCm

Setting it up is tricky because Demucs is tightly pinned to older PyTorch versions, so you have to install dependencies manually and use "--no-deps" to avoid breaking your (ROCm-)PyTorch setup.

Result:
Very clean vocal separation in most parts. Some artifacts appear during very loud or distorted sections (e.g. emotional peaks or shouting).

Next steps / possibilities:

- Normalize and filter audio before separation
- Extract vocals for transcription or remixing
- Create karaoke / instrumental versions
- Combine with Whisper for lyrics
- Batch processing for datasets
- Model: htdemucs_ft (higher quality)

Video workflow:

- Recorded with OBS
- Edited in Kdenlive
- Transcoded with VAAPI (H.264)

No cloud, real hardware.
Everything runs on Linux, so anyone can set this up.
Works on CPU as well, but much slower.

#Demucs #AI #MachineLearning #AudioSeparation #MusicAI #OpenSource #Linux #ROCm #AMD #DeepLearning #AudioProcessing #Vocals #Karaoke #StemSeparation #SelfHosted #NoCloud #FOSS #Tech #LocalAI #MetaAI

I’d like to share a project I think deserves more visibility in the Fediverse:

https://musicprism.app/

MusicPrism is an audio stem separation tool developed by a worker-owned tech cooperative from Argentina 🇦🇷
https://wannacode.ar/

In a landscape dominated by VC-backed AI startups, it’s refreshing to see audio tech built by a worker-managed cooperative.

If you’re into music production, remixing, sampling, or audio experimentation, I’d love for you to try it and share feedback.

A boost could help this reach the right person in the Fediverse

#Fediverse #Mastodon #IndieDev #AudioTech #MusicProduction #AI #Cooperatives #StemSeparation #MadeInArgentina #Argentina #Cordoba

Something we always wanted! Meta’s new SAM Audio brings it. Now that the tech is available, it can be added to DAWs and NLEs as native features or as plugins quick. Just a matter of time. Possibly within 2026.

https://prashantmishra.xyz/blog/article/sam-audio-meta-s-new-unified-multimodal-audio-separation-model?auid=xyz-0110

#musictech #stemseparation #meta #ml #multimodal

SAM Audio: Meta's new Multimodal Audio Separation Model | Blog - Prashant Mishra

The future of audio separation is multimodal. DAWs and NLEs will soon use a similar feature as Meta's new SAM Audio has introduced.

Ableton Live 12.3 Beta Introduces Built-In Stem Separation and Workflow Enhancements

Ableton has released a public beta for Live 12.3, introducing key updates designed to streamline music production. The headline feature is built-in stem separation powered by Music.AI, allowing users to isolate vocals, drums, bass, and other elements directly within Live, eliminating the need for ex... [More info]

Steinberg Media Technologies just previewed SpectralLayers 11 and it blew my mind 🤯 The unmix for Sax & Brass section and even Lead vocals vs Backing vocals! And it all sounds so clean.

https://www.youtube.com/live/2BoEgBGiafM?feature=shared

#spectralayers #steinberg #audioplugin #machinelearning #stemseparation #mir #musicinformationretrieval #musictech #vst

SpectraLayers 11 World Premiere

Join Dom Sigalas at this streamed preview event to discover the exceptional new features in SpectraLayers 11. From its spectacular integration of AI and prec...

YouTube
Hit'n'Mix Updated RipX DAW to 7.1

Hit'n'Mix integrates DeepRemix, DeepCreate, and DeepAudio modules into RipX DAW

Review Space