OpAMP server with MCP – aka conversational Fluent Bit control

I’ve written a few times about how OpAMP (Open Agent Management Protocol) may emerge from the OpenTelemetry CNCF project, but like OTLP (OpenTelemetry Protocol), it applies to just about any observability agent, not just the OTel Collector. As a side project, giving a real-world use case work on my Python skills, as well as an excuse to work with FastMCP (and LangGraph shortly). But also to bring the evolved idea of ChatOps (see here and here).

One of the goals of ChatOps was to free us from having to actively log into specific tools to mine for information once metrics, traces, and logs reach the aggregating back ends, but being able to. If we leverage a decent LLM with Model Context Protocol tools through an app such as Claude Desktop or ChatGPT (or their mobile variants). Ideally, we have a means to free ourselves to use social collaboration tools, rather than being tied to a specific LLM toolkit.

With a UI and the ability to communicate with Fluentd and Fluent Bit without imposing changes on the agent code base (we use a supervisor model), issue commands, track what is going on, and have the option of authentication. (more improvements in this space to come).

New ChatOps – Phase 1

With the first level of the new ChatOps dynamism being through LLM desktop tooling and MCP, the following are screenshots showing how we’ve exposed part of our OpAMP server via APIs. As you can see in the screenshot within our OpAMP server, we have the concept of commands. What we have done is take some of the commands described in the OpAMP spec, call them standard commands, and then define a construct for Custom Commands (which can be dynamically added to the server and client).

The following screenshot illustrates using plain text rather than trying to come up with structured English to get the OpAMP server to shut down a Fluentd node (in this case, as we only had 1 Fluentd node, it worked out which node to stop).

Interesting considerations

What will be interesting to see is the LLM token consumption changes as the portfolio of managed agents changes, given that, to achieve the shutdown, the LLM will have had to obtain all the Fluent Bit & Fluentd instances being managed. If we provide an endpoint to find an agent instance, would the LLM reason to use that rather than trawl all the information?

Next phase

ChatGPT, Claude Desktop, and others already incorporate some level of collaboration capabilities if the users involved are on a suitable premium account (Team/Enterprise). It would be good to enable greater freedom and potentially lower costs by enabling the capability to operate through collaboration platforms such as Teams and Slack. This means the next steps need to look something along the lines of:

#AI #chatops #FluentBit #Fluentd #LangGraph #LLM #MCP #OpAMP #OpenTelemetry #OTel #OTLP

Recebi isso no e-mail. Devo me preocupar?

"[...] We’re writing to let you know that Cloud Observability has launched a new OpenTelemetry (OTel) ingestion API that supports native OpenTelemetry Protocol (OTLP) logs, trace spans, and metrics.

Starting March 23, 2026, this API will be added as a dependency for the current Cloud Logging, Cloud Trace, and Cloud Monitoring ingestion APIs. This change ensures a seamless transition as collection tools migrate to this new unified endpoint.
What you need to know

Key changes:

The existing Cloud Observability ingestion APIs (logging.googleapis.com, cloudtrace.googleapis.com, and monitoring.googleapis.com) are automatically activated when you create a Google Cloud project using the Google Cloud console or gcloud CLI. The behavior remains unchanged for projects created via API, which do not have these ingestion APIs enabled by default. Starting March 23, 2026, the new OTel ingestion endpoint telemetry.googleapis.com will automatically activate when any of these specified APIs are enabled.

[...]

Thanks for choosing Google Cloud Observability.
– The Google Cloud Team"

#OTLP #OpenTelemetry

How dy'all read #OTLP?
oddle-pee
0%
ottle-pee
0%
oh-tee-el-pee
100%
other, pweaze comment uwu
0%
Poll ended at .

🛠️ Monitor logs & events: user prompts, tool results with accept/reject decisions, #API requests with detailed cost/duration/token counts, error tracking with retry attempts & tool decision patterns

⚙️ Setup options: Launch via #VSCode or terminal with environment variables, configure centralized #OTLP exports via #gRPC or #HTTP protocols, use administrator-managed settings for organization-wide control

You asked, we listened: OTLP/gRPC support is now available in Coroot 1.14.3 onwards! 🧑‍💻 🔥 🤌 https://github.com/coroot/coroot/releases

#OpenTelemetry #OTLP #OTEL #observability #linux #opensource #softwarelibre #freesoftware #eBPF #Coroot

Wednesday Links - Edition 2025-08-13

What's new in Gradle 9.0.0 (8 min)🎉 https://gradle.org/whats-new/gradle-9/ Preparing for Spring...

DEV Community

Is anybody I know (or who reads this post 😅) using #Azure Container Apps with #OpenTelemetry collection enabled sending the telemetry to an external provider?

https://learn.microsoft.com/en-us/azure/container-apps/opentelemetry-agents#otlp-endpoint

#o11y #OTel #OTLP

Collect and read OpenTelemetry data in Azure Container Apps

Learn to record and query data collected using OpenTelemetry in Azure Container Apps (preview).

Score! Managed to hook the cluster at home up to #Grafana #cloud over #OTLP for shipping #traces 🎉 . Next up:
* Figure out how to do a better presentation in Grafana
* Optimize OTEL collection deployment, work out all the issues etc

Then the real fun begins. As I have two WIP packages to bring out:
* One for @reactphp 's filesystem
* And one for Bunny

Both of those need to be refined, but with Shawn Maddock's (https://github.com/smaddock) initial push in this direction and help

#php #reactphp

[Перевод] Эй, где моя ошибка? Как OpenTelemetry фиксирует ошибки

Языки программирования расходятся во мнении, что такое ошибки или исключения и как их обрабатывать. Возникает вопрос: что использовать, если нужны стандартизированная телеметрия и отчёты об ошибках для написанных на этих языках микросервисов? Ответом может быть OpenTelemetry. Перевели материал, из которого вы узнаете, как OpenTelemetry обрабатывает ошибки, чем отличаются ошибки в спанах и логи, а также как добавить в спаны OTel метаданные и события.

https://habr.com/ru/companies/flant/articles/892784/

#monitoring #opentelemetry #span #events #спан #события #otlp #jaeger #микросервисы #ошибки

Эй, где моя ошибка? Как OpenTelemetry фиксирует ошибки

В зависимости от языка программирования, на котором вы привыкли работать, у вас могут быть определённые представления о том, что такое ошибка, что такое исключение и как его следует обрабатывать....

Хабр
Wednesday Links - Edition 2025-03-12

The Gatherer API (32 min)🧺 https://dev.java/learn/api/streams/gatherers/ Null Safety in Spring...

DEV Community