"Galton and his fellow eugenicist / protegee Karl Pearson were not directly involved in the development of early computational machines. But Galton’s foundational work with multidimensional modeling — a technique he used while measuring the attractiveness of African and European women — shaped Pearson’s thinking as he developed statistical tools like logistic regression, which is one of the fundamental components of modern machine learning."

https://www.theverge.com/entertainment/897923/ghost-in-the-machine-valerie-veatch-interview

The gen AI Kool-Aid tastes like eugenics

Ghost in the Machine — out on Kinema March 26th — director Valeria Veatch speaks with The Verge about gen AI’s roots in eugenics.

The Verge
@timnitGebru I'm no mathologist, but this seems like a stretch. Is the argument that mathematics used or developed by crappy people is itself suspect for that reason? Is the math not able to be valid on its own, if it was developed or associated with crappy people?
@wesdym @timnitGebru We’re not talking about theoretical mathematics here, we’re talking about technology. And yes, all technology is political. It is designed, funded, and developed (in the specific way that it is developed), shaped by the ideologies and success criteria of those who design, fund, and develop it.

@aral "Statistical tools" sounds an awful lot like mathematics to me.

But sure, whatever. Go put some more smart-sounding bumper-stickers on your car and smugly applaud yourself, I guess.

@wesdym @aral Maybe read the rest of the article. The issue is clearly not the algorithm but the way they think they can use it.
In statistic, you consider outliers to belong in the dataset, even if they are flawed. They will eventually be scrapped by their low weight. The issue with this is that :
1. some outliers are about actual discriminated cases or person.
2. Actual discrimination might not be an outlier, because of how well it is spread out in the dataset.

@wesdym @aral Now that would mean you need to curate your datasets thoroughly. But that would mean doing a lot of extra work they don't believe is relevant, compared to growing datasets size. Quantity over quality.

Now that doesn't even fix the issue of normalization, another discriminatory behavior, which pushes this algorithm towards the optimal spot in order to give a value that fits, even though the answer would be unoriginal.

@mehdi_benadel @wesdym @aral Whoa I'm gonna go ahead and block this person. I was too generous in my response before seeing this additional nonsense.
@wesdym @aral Oh lol actually let me go ahead and block you from our server. I was too generous with my response before seeing this.

@aral @wesdym @timnitGebru "all technology is political. It is designed, funded, and developed (in the specific way that it is developed), shaped by the ideologies and success criteria of those who design, fund, and develop it."

I think this is a much more compelling argument than the genealogical one. That racists invented key statistical methods, and modern applied statistics ("AI") often produces racism, are otherwise two disconnected claims. Maybe the documentary connects them like this, but the article doesn't convey that well?

It also provides a way to account for cases that have clearly eugenic origins but depart from them (e.g. Galton-Watson branching processes were invented to model 'the extinction of aristocratic families', but have applications in nuclear physics.)

And it would mean even if logistic regression was invented by Frederick Douglass [*insert Galton contemporary of your choice], you could still get racist "AI" bankrolled by fashy oligarchs. Offspring need not be faithful to their origins.

I look forward to the documentary; Galton and Pearson are sprinkled throughout statistics education with barely any critical examination (learn the method, practice the proofs, move on...). It's good to surface this history. But I think origins are less instructive than the extant constellation of interests.