Nyancient CthUwU

9 Followers
36 Following
243 Posts
weeb / functional programmer / computer science phd / leftist sjw / magical non-binary girl(?)
Pronounsthey/them
GitHubhttps://github.com/nyancient

@gknauss @Migueldeicaza I have a better solution to the problem of reading Twitter threads and it it this:

Utterly disregarding the thoughts of anyone who is still choosing to use Twitter in this, The Year of our Dystopia Two Thousand Twenty Six.

Anyone who is still making that decision is not worthy of my attention. I do not care what they have to say. They have disqualified themselves utterly.

@RosaCtrl having bots running around closing issues left and right is a pretty good way to tell everyone that you don't want their bug reports. Whenever I see this my motivation to report a bug drops by like 90%. Usually I just stop using the project instead, if the bug I wanted to report is severe enough.
@RosaCtrl I would say a basic understanding of matrix and vector math, as well as some programming experience, are the only real requirements. Reading an introduction to neural networks in general is helpful, but the book covers that too briefly.
@RosaCtrl I mean, "Pinochet would have voted for me" is not so much a dog whistle as it is the fucking gjallarhorn... :/
@RosaCtrl I would say something like "next step is betting on which out of these 10 kids in Gaza will be killed by IDF snipers first", except I really don't want anyone to tell me "oh yeah, they did that last week".
@yogthos approved by the EU council, which means nothing in and of itself. They've approved several versions before and parliament has struck down all of them. No reason to go full panic mode just yet.

@lina @deivpaukst I'd be very interested in those takes. Sounds fun! (well, as much fun as can be had in your current situation...)

It's almost like Luna is the leftist version of the brain virus that turns normal people into deranged culture warriors in no time flat starting from the day they have their first mildly transphobic thought.

@RosaCtrl yeah, that's the book! The physical copy comes with a free ebook copy so that's nice, though Manning will send you ungodly amounts of spam after you register for it.

Do you personally know anyone reporting these amazing results? I only really know one person who claims any results even close to amazing, and they're kind of well known for taking ill-advised shortcuts and being terrible at estimating stuff. And even they concede that LLMs are only really good for particular types of tasks.

Having more or less grown up on 4chan, I just assume that anything posted on the internet is a complete fabrication until proven otherwise, so I don't really put much stock in self-reported internet success stories. 😅

@RosaCtrl I feel you! When "everyone" is gushing over how LLMs are going to obsolete everything from accountants to software to courts, it can be pretty hard to not get swept up in it. Especially since so much of the marketing is disguised as worry about "the consequences of the next industrial revolution" (remember that open letter from "AI" companies about "AI" being "too dangerous", asking for a six month moratorium on new models? Excellent marketing right there).

I think the best way to immunize oneself is to just get a good understanding of how LLMs work and why they're not what people claim. I've spent countless hours using state of the art tools to troubleshoot problems that were too hard to figure out myself in a few minutes, and the main takeaway from those sessions is the sheer insidiousness of how working with an LLM feels like you're making progress while you're really just being led around in circles. Not once has any of these sessions contributed to me solving the problem at hand. Instead, I've wasted my time being tricked into believing that the solution is just around the next prompt.

Same with code generation. LLMs are great for templating trivial things ("give me a set of python dataclasses matching this openapi spec/example request"), but completely fall down when it comes to anything that more complex than that which also needs to be maintained. Seeing the terrible PRs submitted by colleagues who don't even stop to think "is this code even necessary" before generating 300 lines of technical debt really reinforces that it's not you who missed something - it's them.

I can really recommend the book "building a large language model from scratch", which walks you through building GPT2. Aside from making a pet bullshit generator being a fun exercise, it really pulls back the curtain on the whole "LLMs are just one step from being a self-aware super intelligence" spiel. The thing that separates OpenAI and friends from anyone with a bit of Python experience and basic understanding of matrix multiplications isn't some mythical AI secret sauce; it's just having access to more bandwidth and GPU compute.

That said, I still worry sometimes about the effect the AI hype will have on the world, but the realization that the danger is just the same old large scale irrational hype capitalism that's already fucking us, and not some new scary alien technology with the power to reshape reality, makes it infinitely less anxiety-inducing.

EDIT: and before some smartass comes along to crow about how more recent LLMs are nothing like GPT2: they're just minor iterations on the same concept with a few improvements that are so obvious they could have been a BSc thesis, if BSc theses had access to infinite GPUs.

@RosaCtrl solution: stop respecting their opinion on technical matters. It's jarring AF at first (I almost had a stroke when my PhD advisor started working on a fucking metaverse product), but the catharsis is so worth it.

Just like crypto didn't make actual money obsolete, LLMs won't make actual software development obsolete.