Richard

@richrants@toot.community
25 Followers
81 Following
169 Posts
Supply chain technology guy. I love the outdoors 🏞️, hiking 🥾 and learning how things work ⚙️; interested in natsec, infosec & foreign policy. I also like cars 🚗 & planes ✈️. Formerly of IBM and Blue Yonder; a bit of that sticks with me, I guess.
I work at TradeBeyond, yet this is a personal account with my personal opinions.
FromRhine-Main area, 🇩🇪 Germany, 🇪🇺 Europe

When an entire class of technology states on the packaging that it was made in China but intended "for overseas use only," this should really give you pause before plugging it into your network.

You will find this verbiage on a lot of Android TV streaming boxes for sale at the major retailers. There's a very good reason the country that makes this crap doesn't want it on their own networks. My advice: If you have one of these Android streaming boxes on your network or get one as a gift, toss it in the trash. I'll have a lot more about this in the New Year, but these things are responsible for building out a botnet that currently has ~2M devices and is growing rapidly. https://blog.xlab.qianxin.com/kimwolf-botnet-en/

Today is my first day at AWS.

I noticed a small bug in DynamoDB clustering implementation and I think I fixed it.

Shipped to prod already.

Going to make a coffee and will check back if everything is working.

I think this needs to be repeated, since I tend to be quite negative about all of the 'AI' hype:

I am not opposed to machine learning. I used machine learning in my PhD and it was great. I built a system for predicting the next elements you'd want to fetch from disk or a remote server that didn't require knowledge of the algorithm that you were using for traversal and would learn patterns. This performed as well as a prefetcher that did have detailed knowledge of the algorithm that defined the access path. Modern branch predictors use neural networks. Machine learning is amazing if:

  • The problem is too hard to write a rule-based system for or the requirements change sufficiently quickly that it isn't worth writing such a thing and,
  • The value of a correct answer is much higher than the cost of an incorrect answer.

The second of these is really important. Most machine-learning systems will have errors (the exceptions are those where ML is really used for compression[1]). For prefetching, branch prediction, and so on, the cost of a wrong answer is very low, you just do a small amount of wasted work, but the benefit of a correct answer is huge: you don't sit idle for a long period. These are basically perfect use cases.

Similarly, face detection in a camera is great. If you can find faces and adjust the focal depth automatically to keep them in focus, you improve photos, and if you do it wrong then the person can tap on the bit of the photo they want to be in focus to adjust it, so even if you're right only 50% of the time, you're better than the baseline of right 0% of the time.

In some cases, you can bias the results. Maybe a false positive is very bad, but a false negative is fine. Spam filters (which have used machine learning for decades) fit here. Marking a real message as spam can be problematic because the recipient may miss something important, letting the occasional spam message through wastes a few seconds. Blocking a hundred spam messages a day is a huge productivity win. You can tune the probabilities to hit this kind of threshold. And you can't easily write a rule-based algorithm for spotting spam because spammers will adapt their behaviour.

Translating a menu is probably fine, the worst that can happen is that you get to eat something unexpected. Unless you have a specific food allergy, in which case you might die from a translation error.

And that's where I start to get really annoyed by a lot of the LLM hype. It's pushing machine-learning approaches into places where there are significant harms for sometimes giving the wrong answer. And it's doing so while trying to outsource the liability to the customers who are using these machines in ways in which they are advertised as working. It's great for translation! Unless a mistranslated word could kill a business deal or start a war. It's great for summarisation! Unless missing a key point could cost you a load of money. It's great for writing code! Unless a security vulnerability would cost you lost revenue or a copyright infringement lawsuit from having accidentally put something from the training set directly in your codebase in contravention of its license would kill your business. And so on. Lots of risks that are outsourced and liabilities that are passed directly to the user.

And that's ignoring all of the societal harms.

[1] My favourite of these is actually very old. The hyphenation algorithm in TeX trains short Markov chains on a corpus of words with ground truth for correct hyphenation. The result is a Markov chain that is correct on most words in the corpus and is much smaller than the corpus. The next step uses it to predict the correct breaking points in all of the words in the corpus and records the outliers. This gives you a generic algorithm that works across a load of languages and is guaranteed to be correct for all words in the training corpus and is mostly correct for others. English and American have completely different hyphenation rules for mostly the same set of words, and both end up with around 70 outliers that need to be in the special-case list in this approach. Writing a rule-based system for American is moderately easy, but for English is very hard. American breaks on syllable boundaries, which are fairly well defined, but English breaks on root words and some of those depend on which language we stole the word from.

@derPUPE Really important to note that this is about on device learning and not about training an llm or anything like that. Details are here:

https://www.apple.com/legal/privacy/data/en/ask-siri-dictation/

This setting has been there for ten years and every couple of years it makes the rounds on social media and blogs again where people imply that this setting shares your private app content with Apple. It does not.

Legal - Siri, Dictation & Privacy- Apple

Data & Privacy

Apple Legal
Please read this speech by a French senator who tells harsh truths about the Trump regime and Russia, and challenges Europe to stand up for what is needed. Then, please ask your members of Congress to read it, too. https://thebulletin.org/2025/03/for-this-french-senator-trump-is-a-traitor-and-europe-is-now-alone/amp/
For this French senator, Trump is a traitor—and Europe is now alone

In one month, Trump has done more harm to the Atlantic alliance than in four years of his last presidency. Europeans now see him as a traitor.

Bulletin of the Atomic Scientists
Accurate.

A 19 year old has been given admin access to the core US treasury financial system by Elon Musk’s DOGE.

He goes by the name of “Mr. Big Balls” online, is on Russian forums, and has tried to purchase DDoS botnets.

Elon has said sharing this information is illegal, so please do not press the boost button.

https://www.wired.com/story/edward-coristine-tesla-sexy-path-networks-doge/

DOGE Teen Owns ‘Tesla.Sexy LLC’ and Worked at Startup That Has Hired Convicted Hackers

Experts question whether Edward Coristine, a DOGE staffer who has gone by “Big Balls” online, would pass the background check typically required for access to sensitive US government systems.

WIRED

What happens when you twist your lens' focus ring while shooting long exposures of fireworks

#photography

@tveskov @rvedotrc Ah, a Brother manual! 😄
Termite ransomware group just claimed Blue Yonder. "Our team got 680gb of data such as DB dumps Email lists for future attacks (over 16000) Documents (over 200000) Reports Insurance documents. Check for updates. Data links will be available soon." HT @AlvieriD