This from Ingold (being alive 2011, pg. 62) reads as more optimistic than it used to. LLMs are closed rather than open, and can we really say its users aren't losing skill as "the ‘living appendages’ of lifeless mechanism"?

@yaxu i see people describe LLMs as algorithms quite often but i'm pretty sure that's not the right word

i suppose they can *mimic* algorithms though

@sean_ae Yep I'd generally push back on LLMs being described as algorithms, if noone can see how they work there's no algorithm to talk about, it's just black box statistics. I think here though they are just talking about algorithms in terms of systems/machine-making in general, and LLMs fit that.

@yaxu @sean_ae tbh we could describe the function of an llm (or any "ai") in algorithmic terms, it's just a stochastic and fractal algorithm... so we can't trace the exact path of input to output, but we can describe in abstract detail what happens to the input to transform it into the output....

and yes, there was a lot of optimistic (naive tbh) writing about tech from a leftist perspective over the preceding 40 years :)

@emenel @yaxu i suppose if you ignore the data you can do that, but it's a bit like saying a person exists if you ignore their memories

(yes i am out on a limb here again)

@sean_ae @yaxu hah... i just mean that we can describe the exact process that happens in training and using the model, in details. we just can't predict the output from the input because of the stochastic aspect of the output... but the workings of the process itself are not a mystery, and could be describe algorithmically ... we can make simpler examples where we could trace the data through it, but not at the scale of the big models in production.
@emenel @yaxu yeah i suppose it's like a markov chain, the process can be described algorithmically but not the chain itself
@sean_ae @yaxu yeah, exactly... or l-systems, fractals, etc...
@emenel @yaxu @sean_ae there’s no world in which Ingold ever intentionally spoke optimistically about technology i feel
@Holly @emenel @sean_ae Heh I spent a day with him once and he hated everything I showed him. He loves his cello !
@yaxu @emenel @sean_ae hahaha i also presented to him at a panel once… it did not go down super well😅

@yaxu idk - do they even qualify as technology?

they're weird, statistical models, not designed or controlled, they just kind of *are*

@sean_ae I haven't really looked into them but surely they are designed, built and controlled by people? The people making them are all about power, so don't know why they wouldn't use their ownership of LLMs to change political opinion etc. Musk has talked about this a lot
@sean_ae I like Ursula Franklin's definition of technology
@yaxu @sean_ae ursula franklin is amazing, and tragically under appreciated ... she taught at my university, there's a street named after her and an annual lecture on technology and culture called the Ursula Franklin Lecture :) close to my heart.
@yaxu @sean_ae no technology has a "state of nature" ... it's all designed and controlled even if there's an element of randomness in the system. those statistical weights don't appear from nothing...

@emenel @yaxu i think you might be drifting into intelligent design territory there

i mean lets say you trained a model entirely on pictures of dogs

@sean_ae @yaxu they did that... it was the first deep dream ... (or maybe that's the reference you're making and i'm just missing it lol)

which is a great example, because we know exactly what happens when you model images of dogs and can absolutely describe the process in detail without any black-boxes or "nature" ...

@emenel @yaxu no nature apart from all of the data constituting the model, which is the thing being discussed
@sean_ae @yaxu sure .. but that's not a "nature" ... it's a selection. and then it's processed through a very specific set of steps (algorithm?) that produces a certain kind of output (model). yes, the model is a representation of the data to a degree, but not an objective or natural one. it's a specific kind of representation that is designed and enacted.

@yaxu idk if you can say designed or controlled really, maybe slightly

you could maybe say built

it's not really a set of instructions, more a set of weights (most of which are not being set intentionally unless you only include human-generated data and even that would be a bit of a stretch imo)

not often i end up in a semantic hole, pls forgive me

@sean_ae @yaxu but there's a process to derive those weights, refine/edit the data to produce a specific kind of output (done by thousands of abused and exploited workers), manual editing of the resulting weights to make the output "safe" (aka censored and also done by exploited workers), put the model into a piece of software for specific kinds of interactions that are also guided and constrained .... and querying the model (aka writing a prompt) is also a known process ... computers can never be "just the way it is" because left on its own a computer does nothing. anything it does has to be translatable to linear instructions for the cpu.
@emenel @yaxu right, but what you're saying there is that the design rests entirely in the commercial application of [thing] - but the application of what [thing] exactly?
@sean_ae @yaxu the models themselves are intentionally created is what i'm saying. from data selection to training method to operation... there is no "natural state", it's entirely designed... even if the output is unpredictable (to a degree)
@emenel @yaxu if it's entirely designed then why are people having debates about mass scraping of art etc?
@sean_ae @yaxu it’s obviously built on massive amounts of data. but it is not an objective or natural outcome or representation of that data. it’s a specific, designed, and created artifact that uses the data in a specific and intentional way.

@emenel @yaxu yeah ok, let me reframe this

if you didn't do any curation at all - and just trained a network on some arbitrary data - could the model be described as having been designed?

@sean_ae @yaxu yes. what is it modeling? what relations between data are important? what does “training” mean in detail? those are all choices that change what is being modeled.

models are a human invention… the dictionary is also an llm, it models the alphabetical relation between words and associates them with definitions …

@emenel @yaxu hmm idk about this - you seem to have ignored the word 'arbitrary' in my question
@sean_ae @yaxu even if the data is arbitrary you have to choose how to model it. the data itself has no natural model. the model is always designed.

@emenel @yaxu can you explain a bit more what you mean there?

i don't have to specify anything when training a network locally, i can just chuck any data in (sounds/gestures/images/text), not label anything and get data out the other end

@sean_ae @yaxu yes. because you’re using a system that already determines the type of model. the code you use for training a local model is the design of the model … it’s the algorithm that takes your input data and produces a specific type of model.

@sean_ae i.e, if you wanted an alphabetical model that system can't do it... the algorithm for "training a model" as it's commonly used these days is a specific type of modelling... it's not objective or representing any "natural" state of the data. it's using the data to produce a very specific representation.

@yaxu

@emenel @yaxu ah ok, i see

when you say 'model' you mean what i would call a network

@sean_ae @yaxu what's a netword? imo it's a model of relations...

@emenel @yaxu the network is the thing you are training, the model is a result of training that network

i suppose what you are saying is that since networks are designed, some of that design is affecting the quality of the model (and it is, but i would not call the model designed - otherwise they would actually work, and they frequently don't)

@sean_ae @yaxu not even just the quality... but what the model fundamentally is. statistics are only one way to model a dataset, and not an objective or neutral way.

@sean_ae tbh it’s not the quality, it’s the ontology. what is being modeled and what does the model represent? there’s no neutral or objective answer. someone (group, org, corp, etc) designed a system that models a specific kind of data in a specific way.

any and all modeling is designed…

@sean_ae @yaxu like, the thing that is produced by the training process models the data based on the algorithm for analysis and recording (ie what metadata is derived, how is it stored, how can it be recalled). all of that is specific and is what produces the model. the model can't be an objective model of the data because someone had to write the code that does the analysis and builds the model. that code is a process that produces a specific, not generic/objective, representation.

@emenel @sean_ae @yaxu

i supsect the word "network" gets in the way a bit here. we use it in a lot of contexts. so would it be better to use a more precise term like "graph" (mathematical kind: nodes and edges) so that you can then begin to get to the next set of clarifying questions? for example, "is this a directed graph (do the edges point in one direction only)?"

also, a graph is the fundamental structure of a "neural network", but can also be both a) the structure of the input data, and b) the structure of the output data.

then if we think of it in these terms, we can start to deal with the other generic, overloaded term, the "model." is it correct to interpret an LLM "model" as the (non-neutral, designed) set of decisions or instructions for how to *traverse* the output graph produced by the LLM?

@sean_ae @yaxu or another way — the data doesn’t constitute the model. the model is the result of a specific process applied to the data. and that process is designed and known. the process is the model.

@yaxu @sean_ae @toxi re: specifically this idea of blackboxes and comprehensibility specifically: https://scholar.social/@olivia/116543457466890160

(not suggesting to reopen the conversation :), just thought this my be of interest)

@emenel @yaxu @olivia yeah I posted this earlier, interesting isn’t it?

@sean_ae @emenel @yaxu funnily enough I think this is also relevant (notice where we talk about approximation)

https://scholar.social/@olivia/116380878007444018

Olivia Guest · Ολίβια Γκεστ (@[email protected])

Attached: 2 images I think ~4 years ago, Iris and I 1st sat down and formalised various meanings of function and multiple realizability — both core concepts for any serious computationalist discussions — because in part we realised nobody has done this and/or collected these for cogsci. https://doi.org/10.5281/zenodo.19388964 @Iris thread here for more: https://scholar.social/@Iris/116359421483392573 1/

Scholar Social
@olivia @emenel @yaxu sorry for being dumb but what am I looking for?

@sean_ae @emenel @yaxu it's a really short pdf so probably easier to read in your pdf application, but here's a screenshot and sorry myself for not being clear!

Guest, O., Blokpoel, M., & van Rooij, I. (2026). What the func? Multiple Realizability need not be Vague. Zenodo. https://doi.org/10.5281/zenodo.19388964

@olivia @emenel @yaxu this looks fascinating but I can’t get the link to open (!)
@sean_ae @emenel @yaxu I think zenodo is struggling. Here's another place! https://philpapers.org/rec/GUEWTF
Olivia Guest, Mark Blokpoel & Iris van Rooij, What the Func? Multiple Realizability Need not be Vague - PhilPapers

Multiple realizability (MR) is not necessarily unclear nor does it purely operate at the computational level. To understand potential relationships between MR and other constraints, such as metabolic, we formalise possible ...

@olivia @emenel @yaxu ok yeah, I see what you’re saying, very tidy

Can an LLM like ChatGPT be said to have a functional goal though? I would guess (prob wrongly) that ‘it depends’ (?)

@sean_ae @emenel @yaxu yeah exactly it depends

@olivia @sean_ae @emenel @yaxu Between us, if its functional goal is to predict the most probably next word then it is not even able to do that for human language. This is funny to me, because not only is next word prediction not the same as "reasoning, learning, cognition, etc.", but even next word prediction is intractable if one really wants the most probably one. But no-one seems to realise (or care).

https://link.springer.com/article/10.1007/s42113-024-00217-5

Reclaiming AI as a Theoretical Tool for Cognitive Science - Computational Brain & Behavior

The idea that human cognition is, or can be understood as, a form of computation is a useful conceptual tool for cognitive science. It was a foundational assumption during the birth of cognitive science as a multidisciplinary field, with Artificial Intelligence (AI) as one of its contributing fields. One conception of AI in this context is as a provider of computational tools (frameworks, concepts, formalisms, models, proofs, simulations, etc.) that support theory building in cognitive science. The contemporary field of AI, however, has taken the theoretical possibility of explaining human cognition as a form of computation to imply the practical feasibility of realising human(-like or -level) cognition in factual computational systems, and the field frames this realisation as a short-term inevitability. Yet, as we formally prove herein, creating systems with human(-like or -level) cognition is intrinsically computationally intractable. This means that any factual AI systems created in the short-run are at best decoys. When we think these systems capture something deep about ourselves and our thinking, we induce distorted and impoverished images of ourselves and our cognition. In other words, AI in current practice is deteriorating our theoretical understanding of cognition rather than advancing and enhancing it. The situation could be remediated by releasing the grip of the currently dominant view on AI and by returning to the idea of AI as a theoretical tool for cognitive science. In reclaiming this older idea of AI, however, it is important not to repeat conceptual mistakes of the past (and present) that brought us to where we are today.

SpringerLink
@Iris @sean_ae @emenel @yaxu I was gonna type this but I was hoping you'd emerge and say it better too 🙂‍↕️

@Iris @olivia @emenel @yaxu yeah, absolutely not fit for purpose, and it’s legit strange to me how easily people are swept along

Very cool to see some formal examination of this sweeping, nice work 🙏

@yaxu this does read optimistic, yeah. even as specific skills are lost, to shovels or to spellcheck or to LLMs writing summarizations of complex texts, etc., the general human (or perhaps, most forms of life?) tendency to try and create new specific skills either around or utilizing the new mechanizations/tools will surely continue.

idk that that worst case Marx bit is really possible in the doom-y interpretation of it, in part because of that drive to create new specific skills.

cool share!

@yaxu i wouldn't read that as optimistic, as he certainly agrees with some skills being lost. the sigaut quote especially shows this, some skills are sucessfully "automated", in the sense that they tend to do what people have done before well enough that corporations can make more profit with the machines rather than with people. but "along a line of resistance" other skills are developed.

re: open vs closed, i'd assume he's talking not about open-source or "malleability" or anything of the like. he's quoting gilbert simondon before this paragraph, who also wrote a lot about the milieu or environments of technics. so in that context open would imo mean something like "interacting with the real world". so the circular saw from the prev paragraph, that is supposed to be a closed machine, sawing wood perfectly the same time every time, is actually open, because the wood that it saws is "imperfect" and has variable density etc.

similarily LLMs are used by real people, run on physical hardware with electricity and water from real sources, etc. a closed technology would be one that can perfectly enclose all parameters of that skill which it aims to automate, which is just something that no machine can ever achieve, certainly not LLMs. i suppose the "imperfections" would in this case be the differences and subleties in human language/writing, the idea to model "reasoning" as a stochastic process in general, etc.

to come back to the skills question again, it would be too early to say what skills people are developing through use of LLMs. that there is some de-skilling taking place is undeniable, the question is just what skills will take those places.