@[email protected]There are so many different technologies within the
#AI cap that it's impossible to reason about the whole.
And this over-hyped whole is toxic because of the billionaries that finance its development and narration.
The first step to build a less anti-social technology is to get rid of the billionaries in the loop.
To this aim, we should replace the anthropomorphic language we use about it so that it becomes unsellable.
No "artificial intelligence", just statistically programmed software.
No "artificial neural networks" but vector mapping (virtual) machines that can be statistically¹ programmed.
No "traning data" but source data.
No "training" but data compilation.
And so on...
With such conceptual framework, most problems of "ai" disappear:
- The source data are to the weights what the source code is to an executable binary.
- The software running the "inference" is just a virtual machine with a custom architecture running a software expressed as matrices of floats.
- The weights are indeed an executable lossy compression of the source data.
- In the case of #LLM the software extract and patch together excerpt of such archive full of decompression artifacts.
- The "data scientists" are just programmers that instead of writing code to get an x86_64 executable, collect and select the source data. And just like with the source code a classic programmer use, they need the rights or permission to use the source data.
- Because of the loss and randomization during source data compilation, the inference output has no intrinsic meaning, even when it's optimized to fool a human and make it think it has one.
- In such output there are no hallucinations ever: it has no meaning, thus it's neither right or wrong.
- Those publishing LLM outputs on their website or distributing them through API (#OpenAI and friends) are plainly violating the copyrights of authors whose text were included in the source data, because the model is just a float-vector-encoded lossy archive of such texts.
Get rid of the antropomorphism, and you really get a normal technology. Pretty boring and useful in few specific situations that do not require generalized access by untrained people.
@[email protected] @[email protected]¹ some argue that statistics imply theoretical models of a phenomenon just to decide which data to collect and analyze, while current "ai" digest unstructured data. This is generally false, even unsupervised learning assume certain hypothesis in the choice of data to use. But even so, the point is that you need tons of examples to tune the matrices, so if you don't like "statistical programming", maybe "data-driven software synthesis" or "programming by tons of examples" works better (but I'm really open to alternatives).