Where's the hype around Local LLMs?
Local inference seems like the most reasonable way forward, in order to avoid security and privacy concerns.
I'm currently trying out Ollama and LMStudio, primarily using Qwen3:8B. It's not fast with my hardware (CPU only), but it's exciting to see it works.