λTotoro

@lambdatotoro@chaos.social
483 Followers
383 Following
3.4K Posts

Nonbinary CS PhD student, λ enthusiast, Pen-and-Paper RPG enjoyer and ADHD brain pilot.

Equality. Freedom. Solidarity. ✊🏳️‍🌈🏴
Sometimes anti-social but always anti-fascist!
↙️↙️↙️

I am interested in, among other things and in no particular order:

- feminism
- procedural generation
- game design
- art in various forms
- writing
- animals (incl. humans)
- mathematics
- programming
- ... and various intersections of the above.

All my image posts have alt-text. No alt-text, no boost!

Profile Picturethemarshmallowslayer @ DeviantArt
Pronounsany/all
Matrix@lambdatotoro:acmelabs.space
Alt-Account (mutuals are invited)@dodecahedron
Whereas toxic monogamous cishet people are extremely irritating by getting jealous of the idea of their partners dating or kissing other people, I am instead extremely irritating by trying to convince people they should be dating or kissing my partners
asbestos is the future, asbestos is a modern material of science. if you aren't using asbestos in everything you are going to be left behind

We need more cooperative rooftop gardens with shading and power generation by mobile solar panels.

And we need to abolish landlords.

Large Language models (LLMs) do not simulate human psychology. That's the title of our new paper, available as preprint today (1/12):

https://arxiv.org/abs/2508.06950

Large Language Models Do Not Simulate Human Psychology

Large Language Models (LLMs),such as ChatGPT, are increasingly used in research, ranging from simple writing assistance to complex data annotation tasks. Recently, some research has suggested that LLMs may even be able to simulate human psychology and can, hence, replace human participants in psychological studies. We caution against this approach. We provide conceptual arguments against the hypothesis that LLMs simulate human psychology. We then present empiric evidence illustrating our arguments by demonstrating that slight changes to wording that correspond to large changes in meaning lead to notable discrepancies between LLMs' and human responses, even for the recent CENTAUR model that was specifically fine-tuned on psychological responses. Additionally, different LLMs show very different responses to novel items, further illustrating their lack of reliability. We conclude that LLMs do not simulate human psychology and recommend that psychological researchers should treat LLMs as useful but fundamentally unreliable tools that need to be validated against human responses for every new application.

arXiv.org

A contact just told me that my old "LLMs generate nonsense code" blog post from 2 years ago is now very outdated with GPT5 because it's so awesome and so helpful. So I asked him to give it a test for me, and asked it my favorite test question based on a use-case I had myself recently:

Without adding third-party dependencies, how can I compress a Data stream with zstd in Swift on an iPhone?

and here is the answer from ChatGPT 5: https://chatgpt.com/share/68968506-1834-8004-8390-d27f4a00f480

Very confident, very bold, even claims "Works on iOS 16+".

Problem with that: Just like any other LLM I've tested that provided similar responses, it is - excuse my language but I need to use it - absolute horseshit. No version of any Apple SDK ever supported or supports ZSTD (see https://developer.apple.com/documentation/compression/compression_algorithm for a real piece of knowledge). It was never there. Not even in private code. Not even as a mention of "things we might do in the future" on some developer event. It fundamentally does not exist. It's completely made up nonsense.

This concludes all the testing for GPT5 I have to do. If a tool is able to actively mislead me this easy, which potentially results in me wasting significant amounts of time in trying to make something work that is guaranteed to never work, it's a useless tool. I don't like collaborating with chronic liars who aren't able to openly point out knowledge gaps, so I'm also not interested in burning resources for a LLM that does the same.

ChatGPT - アーミッシュ自転車の変化

Shared via ChatGPT

ChatGPT

Perhaps don't launch businesses that rely on breaking the law? Just a thought.

'If the appeals court denies the petition…the emerging company may be doomed. As Anthropic argued, it now "faces hundreds of billions of dollars in potential damages liability at trial in four months"…that involves "up to seven million potential claimants, whose works span a century of publishing history," each possibly triggering a $150,000 fine.'

https://arstechnica.com/tech-policy/2025/08/ai-industry-horrified-to-face-largest-copyright-class-action-ever-certified/

#copyright #ai

AI industry horrified to face largest copyright class action ever certified

Copyright class actions could financially ruin AI industry, trade groups say.

Ars Technica

anime girl turns to face the camera directly

Hey folks, we have fun here in yuri, don't we? But always remember, this is just a story, and we're professional miscommunicators! In real life, you should always express your intentions, desires, and concerns clearly and up-front. Be confident! Your date will appreciate your candor, I promise.

Alt text for blind and low vision users
Alt text for low bandwidth users
Alt text for flakey Internet
Alt text for digital decay and link rot
Alt text for text based browsers
Alt text for calling out the relevant details you mistakenly assume are obvious to everyone
Alt text for explaining the joke to people that don't have the same background as you
Alt text for the 10,000 people learning something "everyone knows" for the first time today
Alt text for leveling up your own writing skills
Alt text for everyone
A generative AI may not hurt stock price or, through inaction, allow shareholder value to come to harm
Ult: well, it has been time and past time for a while, let's use this as the straw that breaks the camel' back.
My past and future private hackings will be on codeberg rather than github.