CERN uses tiny AI models burned into silicon for real-time LHC data filtering

https://theopenreader.org/Journalism:CERN_Uses_Tiny_AI_Models_Burned_into_Silicon_for_Real-Time_LHC_Data_Filtering

CERN Uses Tiny AI Models Burned into Silicon for Real-Time LHC Data Filtering

CERN has developed ultra-small AI models embedded directly into custom chips to filter massive data streams from the Large Hadron Collider in real time, addressing the enormous data challenge of the world’s most powerful particle accelerator.

TheOpenReader
A bit of hype in the AI wording here. This could be called a chip with hardcoded logic obtained with machine learning
Is a LLM logic in weights derived from machine learning?
Well, yes. That's literally what it is.
What what is? The article has nothing to do with LLMs. It even explicitly says they don’t use LLMs.

> Is a LLM logic in weights derived from machine learning?

I was just answering this question. LLM logic in weights is fundamentally from machine learning, so yes. Wasn't really saying anything about the article.

Good one… but Is a DB query filter AI? I forgot to say though is sounds like a really cool thing to do
Strictly speaking, expert systems are AI as well, as in, an expert comes up with a bunch of if/else rules. So yes technically speaking even if they didn’t acquire the weights using ML and hand-coded them, it could still be called AI.

It is 100% valid to label an algorithm that plays tic-tac-toe as "AI"

Much of the early AI research was spent on developing various algorithms that could play board games.

Didn't even need computers, one early AI was MENACE [1], a set of 304 matchboxes which could learn how to play noughts and crosses.

[1] https://en.wikipedia.org/wiki/Matchbox_Educable_Noughts_and_...

Matchbox Educable Noughts and Crosses Engine - Wikipedia

Yup this is exactly my point, in the 80s there were plenty of “AI” companies and “fuzzy logic” was the buzzword of the day.