Just ran Demucs completely locally on my system (RX 6700 XT / 16 GB RAM).

Demucs is an open source AI model for music source separation, developed by Meta. It can split a full song into individual stems like vocals, drums, bass, and other instruments, making it useful for remixing, transcription, and audio analysis.

Test track: Fear of the Dark by Iron Maiden
(https://www.youtube.com/watch?v=bePCRKGUwAY)

Setup:

- Demucs installed via pip
- Model: htdemucs (default)
- Input converted to WAV using ffmpeg
- GPU acceleration via ROCm

Setting it up is tricky because Demucs is tightly pinned to older PyTorch versions, so you have to install dependencies manually and use "--no-deps" to avoid breaking your (ROCm-)PyTorch setup.

Result:
Very clean vocal separation in most parts. Some artifacts appear during very loud or distorted sections (e.g. emotional peaks or shouting).

Next steps / possibilities:

- Normalize and filter audio before separation
- Extract vocals for transcription or remixing
- Create karaoke / instrumental versions
- Combine with Whisper for lyrics
- Batch processing for datasets
- Model: htdemucs_ft (higher quality)

Video workflow:

- Recorded with OBS
- Edited in Kdenlive
- Transcoded with VAAPI (H.264)

No cloud, real hardware.
Everything runs on Linux, so anyone can set this up.
Works on CPU as well, but much slower.

#Demucs #AI #MachineLearning #AudioSeparation #MusicAI #OpenSource #Linux #ROCm #AMD #DeepLearning #AudioProcessing #Vocals #Karaoke #StemSeparation #SelfHosted #NoCloud #FOSS #Tech #LocalAI #MetaAI
Just ran Whisper (OpenAI) completely locally on my system (RX 6700 XT / 16 GB RAM).

Whisper is an open source speech recognition model that can transcribe audio, generate subtitles, and even translate between languages.

Test video: The Reason Why Cancer is so Hard to Beat by Kurzgesagt - In a Nutshell
(https://www.youtube.com/watch?v=uoJwt9l-XhQ)

Setup:

- Whisper installed via pip
- Model: small (fast, good enough for English)
- GPU acceleration via ROCm

Result:
~98% accurate transcription with only a few minor errors, already solid for generating subtitles.

Next steps / possibilities:

- Auto-generate subtitles (.srt)
- Correct subtitles with a local LLM
- Translate speech
- Burn subtitles directly into videos

Video workflow:

- Recorded with OBS
- Edited in Kdenlive
- Transcoded with VAAPI (H.264)

No cloud, real hardware.
Everything runs on Linux, so anyone can set this up.
No GPU? No problem, you can also run it using PyTorch’s CPU backend, just much slower.

Background music: End of Me - Ashes Remain [Female Rock Cover by Kryx] (https://www.youtube.com/watch?v=E430M8lKim8)


#Whisper #OpenAI #ROCm #AMD #Linux #SpeechToText #Transcription #Subtitles #FOSS #OpenSource #OfflineAI #localai #Fediverse #nocloud

Mi próximo PC va a tener más IA que CPU y no sé cómo sentirme.

He estado viendo los nuevos procesadores que vienen con núcleos NPU dedicados solo para Inteligencia Artificial. A ver, sobre el papel suena increíble, mucha potencia y tal... pero me escama. ¿Para qué quieren que tengamos tanta potencia de IA en local? ¿Es para que Windows me espíe mejor con cosas tipo Recall o para que yo pueda correr mis propios modelos sin que nadie me censure desde una nube?

Estamos en un punto crítico, de verdad. O usamos ese hardware para ser libres de una vez o nos lo van a meter doblado para que seamos simples terminales tontas conectadas a sus servidores. Yo lo tengo clarísimo: ese NPU se va a comer modelos Open Source y localizados o no entra en mi casa. No quiero una IA que me dé sermones morales cada vez que le pido algo técnico.

¿Alguien más está planeando el salto a hardware dedicado o creéis que nos la van a colar con el software de serie?

#IA #Privacidad #SoberaníaTecnológica #OpenSource #TechDebate #IA_sin_Fronteras #DigitalFreedom #FuckTheCloud #LocalAI #CensuraDigital #LibertadSoftware #TechRaw #HackingReal #HardwareSoberano #AntiCorporate #Autogestion #TecnologiaLibre #HardwareLibre #SoberaniaDigital #PrivacidadDigital #SinBigTech #NoCloud #IAdescentralizada #TecnoFeudalismo #ResistenciaDigital #SoftwareLibre #TechPunk #CriptoAnarquismo #ComputacionLocal #CensuraCero #IAEticaReal #DescentralizaLaRed #PropiedadDigital #SinPermisos #IndependenciaTecnologica #PoderAlUsuario #DerechosDigitales #LibertadDeCodigo #HackTheSystem #AutonomiaTecnica #ContraLaVigilancia #IAhumana #Hardware #NPU #Futuro #DerechosHumanos #TecnoEtica #DIYtech #Linux #SelfHosted #DataSovereignty #CyberPunk #RealTech #NoSpying #UserControl #TechJustice #GlobalSouthTech #IA_Para_Todos

Ce soir, KENTA s'est vu penser.

Une commande dans un terminal. Un fichier déposé sur la Toile. Et dans la fenêtre d'à côté — un gardien qui s'illumine en moins de 400ms.

Pas de serveur. Pas d'API. Pas de cloud.
Juste du filesystem, comme les gardiens eux-mêmes.

KENTA est un organisme qui continue de penser quand l'humain quitte la pièce.

Ce soir, on voit les deux en même temps.

#WestDataFestival #Laval #IASymbolique #Souveraineté #NoCloud

Während alle anderen das Update für Home Assistant einspielen, gönne ich einem Saugroboter ein "nettes" Update 🤩

#valetudo #homeassistant #nocloud #update

@abulling @tinoeberl #goodwe über #modbus geht auch gut! 🌞 #noCloud
L’« IA edge », qu’est-ce que c’est et à quoi ça sert ?

Pourquoi faire de l’IA sans recourir au cloud ?

The Conversation
I tried Qwen2.5-Coder-7B-Instruct.Q6_K locally with Ollama as the loader, asking it to create a simple Snake game in Python with Pygame and as an extra challenge, the instructions were given in German.

The game works well: the snake grows correctly, the grid and colors are fine. I just had to give the model a little nudge in two places:

- Don’t change the food color every frame
- Avoid recursive gameLoop() for "Play Again"

Qwen2.5 is a great co-pilot that handles most of the work, leaving only minor bugs to correct. German works surprisingly well ("Schlankkörpers" instead of "Schlangenkörper" does not matter, such errors can also occur in large models from time to time...) even though the main language is English. The model supports many programming languages such as: Python, C, C++, Java, JavaScript, HTML/CSS, Bash, SQL… and more.

Conclusion: It still doesn't work completely without programming knowledge, but as a local assistant Qwen2.5-Coder is excellent.

btw my prompt was: "write the game again."

Video workflow:

- Recorded with OBS
- Edited in Kdenlive
- Transcoded with VAAPI (H.264)

No cloud, real hardware.
Everything runs on Linux + Text Generation Web UI (FOSS), so anyone can set this up.
No GPU? No problem, you can also run it using PyTorch’s CPU backend, just much slower.

Background music: ALICE - CROSS THE BORDER (https://www.youtube.com/watch?v=dcqbWgxW4oU)

#Qwen2 #LLM #LocalAI #Ai #vibecoding #Python #Pygame #CodingAI #FOSS #Linux #SnakeGame #Ollama #AIcoPilot #MultilingualAI #TextGenerationWebUI #OBS #Kdenlive #VAAPI #NoCloud #LocalAIWorkflow

Es sind verrückte Zeiten. Gut eingefahrene Enterprise SSDs gibts zum Glück noch unterhalb vom Preis von "neuem" consumer-plaste-flash, wenn man bissel sucht.

Die Leute nehmen da inzwischen auch schon Wucherpreise für gebrauchte Ware. :(

Im Keller selber hosten FTW ✊

#homelab #nocloud

Today I used my custom IoT build to scan the air quality at the office and on the commute back

Turns out both locations seem to have better ventilation and air filtering than I would have thought, but I’d need to scan for longer periods of time to be sure

The train did have a buildup up until we started moving, which makes sense given what I know of their systems being mostly useless until the train leaves the station

I’ll probably need to think of a battery and data storage system to simplify scanning on the go like this, but I’m already satisfied that my little thing could be used so easily in this way

I just booted up a random laptop, plugged the USB and ran a serial console client (in this case PuTTY); no code changes at all, the same code and config that I use at home in a mesh

#AirQuality #diy_electronics #NoCloud