🚀 Đã ra mắt SDK Python cho RamaLama AI Containers – công cụ runtime‑agnostic giúp chạy inference AI trên mọi thiết bị, kể cả phần cứng phi chuẩn. Hỗ trợ llama.cpp, vLLM, mlx, tải mô hình từ HuggingFace, Ollama, OCI, chạy offline, async. Thích hợp cho edge và môi trường air‑gapped. #AI #Python #RamaLama #Containers #EdgeComputing #TríTuệNhânTạo #Python #RamaLama #Container #ThiếtBịCạnhMạng

https://www.reddit.com/r/LocalLLaMA/comments/1qqoife/i_built_a_python_sdk_for_ramalama_ai_containers/

RT @ericcurtin17: Want to run #RamaLama AI on OpenShift DevSpaces Rohan Kumar has got you covered:

https://developers.redhat.com/articles/2025/06/13/how-run-ai-models-cloud-development-environments

RT @rhdevelopers: Discover how #RamaLama's new multimodal feature brings cutting-edge vision-language models directly to your fingertips, s…

RamaLama là công cụ CLI giúp chạy và xây dựng các mô hình LLM qua container. Bước sœur hỗ trợ MLX kèm llama.cpp, vLLM. Tương thích nhiều thiết bị: GPU Apple Silicon, Nvidia, AMD, Intel, mới hỗ trợ NPU Ascend. Mời cộng đồng thử nghiệm với cấu hình đặc biệt. GitHub: containers/ramalama. Hình thành diễn đàn dev Lukas trong weeks sau. #RamaLama #MLX #LLM #AI #Tech #GitHub #مطور #Container

https://www.reddit.com/r/LocalLLaMA/comments/1odilom/ramalama_running_llms_as_containers_adding_mlx/

Raised a bug about #ramalama today not playing well with #arm64 and and #amd gpus. However if you force the base image local inference does use #vulkan to run - and much faster than maxing out the CPU cores on my #altra.
yesterday in #fedora qa:
* made abrt obsolete fros to fix upgrades to #rawhide - https://src.fedoraproject.org/rpms/abrt/c/3e192a597d9539c4f37f1deacbeb8c99091ce299?branch=rawhide
* fixed new gnome-software not launching from overview - https://src.fedoraproject.org/rpms/gnome-software/c/7fb63f649d3ade06860bf6fce2e8a0537b482a73?branch=rawhide
* rest of the day poking around at running local llm via #ramalama and #rocm and using it with #goose ... practically useful? no! fun? kinda!
Commit - rpms/abrt - 3e192a597d9539c4f37f1deacbeb8c99091ce299 - src.fedoraproject.org

today in #fedora qa:
* wrote up some instructions for handling branching with #openqa since I'll be on vacation - https://pagure.io/infra-docs-fpo/pull-request/385
* fixed openQA tests for some changes in #anaconda webui - https://pagure.io/fedora-qa/os-autoinst-distri-fedora/c/d43b28b etc.
* some followup on https://bugzilla.redhat.com/show_bug.cgi?id=2385306
* deployed a couple of new openQA aarch64 worker hosts
* fiddled around with #ramalama a bit using that fancy graphics card. so far...well, the bugs are interesting! https://huggingface.co/unsloth/Qwen3-Coder-30B-A3B-Instruct-GGUF/discussions/1 , https://github.com/containers/ramalama/issues/1783
PR#385: sysadmin guide: Update openQA with branching instructions - infra-docs-fpo - Pagure.io

RamaLama just got multimodal! 🚀 See, understand & respond to visual info with new VLM capabilities. Shoutout to Xuan-Son Nguyen! #RamaLama #AI #llamacpp

https://developers.redhat.com/articles/2025/06/20/unleashing-multimodal-magic-ramalama

Unleashing multimodal magic with RamaLama | Red Hat Developer

RamaLama's new multimodal feature integrates vision-language models with containers. Discover how it helps developers download and serve multimodal AI models

Red Hat Developer
Add documentation for running with Ramalama local model serving in OC… · block/goose@3bec469

…I Containers (#1973) Signed-off-by: Adam Miller <[email protected]>

GitHub