There’s a lot to chew on in this short article (ht @ajsadauskas):
https://www.bbc.com/worklife/article/20240214-ai-recruiting-hiring-software-bias-discrimination

“An AI resume screener…trained on CVs of employees already at the firm” gave candidates extra marks if they listed male-associated sports, and downgraded female-associated sports.

Bias like this is enraging, but completely unsurprising to anybody who knows half a thing about how machine learning works. Which apparently doesn’t include a lot of execs and HR folks.

1/

AI hiring tools may be filtering out the best job applicants

As firms increasingly rely on artificial intelligence-driven hiring platforms, many highly qualified candidates are finding themselves on the cutting room floor.

BBC

Years ago, before the current AI craze, I helped a student prepare a talk on an AI project. Her team asked whether it’s possible to distinguish rooms with positive vs. negative affect — “That place is so nice / so depressing” — using the room’s color palette alone.

They gathered various photos of rooms on campus, and manually tagged them as having positive or negative affect. They wrote software to extract color palettes. And they trained an ML system on that dataset.

2/

Guess what? Their software succeeded!…at identifying photos taken by Macalester’s admissions dept.

It turns out that all the publicity photos, massaged and prepped for recruiting material, had more vivid colors than the photos they took. And they’d mostly used publicity photos for the “happy” rooms and their own photos for the “sad” rooms (which generally aren’t in publicity materials).

They’d encoded a bias in their dataset, and machine learning dutifully picked up the pattern.

Oops.

3/

The student had a dilemma: she had to present her research, but the results sucked! the project failed! she was embarrassed! Should she try to fix it at the last minute?? Rush a totally different project?!?

I nipped that in the bud. “You have a •great• presentation here.” Failure is fascinating. Bad results are fascinating. And people •need• to understand how these AI / ML systems break.

4/

She dutifully gave the talk on the project as is, complete with the rug pull at the end: “Here’s our results! They’re so broken! Look, it learned the bias in our dataset! Surprise!“ It got an audible reaction from the audience. People •loved• her talk.

I wish there had been some HR folks at her talk.

Train an AI on your discriminatory hiring practices, and guess what it learns? That should be a rhetorical question, but I’ll spell it out: it learns how to infer the gender of applicants.

5/

An interesting angle I’m sure someone is studying properly: when we feed these tabula rasa ML systems a bunch of data about the world as it is, and they come back puking out patterns of discrimination; can that serve as •evidence of bias• not just in AI, but in •society itself•?

If training an ML system on a company’s past hiring decision makes it think that baseball > softball for an office job, isn’t that compelling evidence of hiring discrimination?

6/

There’s an ugly question hovering over that previous post: What if the men •are• intrinsically better? What if discrimination is correct?? What if the AI, with its Perfect Machine Logic, is bypassing all the DEI woke whatever to find The Actual Truth??!?

Um, yeah…no.

A delightful tidbit from the article: a researcher studying a hiring AI “received a high rating in the interview, despite speaking nonsense German when she was supposed to be speaking English.”

These systems are garbage.

7/

I mean, maaaaaaybe AI can help with applicant screening, but I’d need to see some •damn• good evidence that the net effect is positive. Identifying and countering training set bias, evaluating results, teasing out confounders and false successes — these are •hard• problems, problems that research work long months and years to overcome.

Do I believe for a hot minute that companies selling these hiring AIs are properly doing that work? No. No, I do not.

8/

AI’s Shiny New Thing promise of “your expensive employees are suddenly replaceable” is just too much of a candy / crack cocaine / FOMO promise for business leaders desperate to cut costs. Good sense cannot survive the onslaught.

Lots of business right now are digging themselves into holes now that they’re going to spend years climbing out of.

9/

Doing sloppy, biased resume screening is the •easy• part of HR. Generating lots of sort-of-almost-working code is the •easy• part of programming. Producing text that •sounds• generally like the correct words but is a subtle mixture of obvious, empty, and flat-out wrong — that’s the •easy• part of writing.

And a bunch of folks in businesses are going to spend the coming years learning all that the hard way.

10/

At this point, I don’t think it’s even worth trying to talk down business leaders who’ve drunk the Kool-Aid. They need to make their own mistakes.

BUT

I do think there’s a competitive advantage here for companies willing to seize it. Which great candidates are getting overlooked by biased hiring? If you can identify them, hire them, and retain them — if! — then I suspect that payoff quickly outstrips the cost savings of having an AI automate your garbage hiring practices.

/end

Yeah. The spam arms race is playing out in many spheres, and it feels kind of desperate right now tbh. A defining feature of our present moment.

From @JMMaok:
https://mastodon.online/@JMMaok/111953486878595616

J Miller (@[email protected])

@[email protected] Good thread! This is all made even harder by the fact that applicants are simultaneously adopting LLMs. This reduces the effort needed to apply, resulting in larger applicant pools with different signals. Heck, the applicants will start to get advised to change softball to baseball. And in the pantheon of resume lies, that’s trivial. But this shift by applicants also means I can’t entirely blame companies for trying some machine learning.

Mastodon
More of Paul’s grumbling on the topic of the AI mania sweeping the business world:
https://hachyderm.io/@inthehands/111927598622910144
Paul Cantrell (@[email protected])

It’s like all these execs are selling their cows for 3 beans because they heard the fairy tale and now they honestly believe they’re going to climb a giant beanstalk and slay a giant

Hachyderm.io
@inthehands @JMMaok Look y’all, how about we just cut the Gordian Knot here and destroy capitalism? No adverts in a world with no money.
@inthehands Last week my own manager asked me a question and then (while screen sharing!) typed the same question into a chatbot while I was still talking. Didn't seem to see a problem with it.

@inthehands Definitely agree. The applicant screening part is a huge problem given how biased the AI systems inherently are.

I'm not sure the 'opportunity' of gathering the AI rejects is viable though. AI applicant screenings will find large numbers of qualified candidates. Just much whiter, male and homogeneous in nature.

That will *eventually* harm the companies, but we aren't starting from a fair playing field. So in the short term it's yet another block on one side of the societal unfairness balance.

@pixelpusher220
Right. What I’m describing isn’t easy. But to the extent that hiring processes are flawed, there is a competitive advantage there to be found.
@inthehands Agreed. Hopefully it can be used successfully!
@inthehands
Backing way up, how many jobs require skills that are relevant to a specific sport? (And no, "teamwork" isn't an answer. There are a million non-sports examples of teamwork that can be highlighted in the average person's work history.)
@MHowell For sure. I mean, the premise is to paint a “whole person” picture that fosters useful conversation in the interview, but I’m sure as often as not things like this become a discrimination vector. Conversely, though, I don’t think it’s possible to scrub enough personal identity characteristics from a resume to prevent discrimination.
@inthehands You can just buy the same AI, and interview only people whose resumes were rejected.
@Retreival9096
This is a fairly compelling idea.
@inthehands it's interesting because things like not using crappy hiring screening tools might be good business in the long run (no need to unfuck processes that use them, paying for them, etc)
@inthehands
Ye Old Technology Adaption Curve. There are always more losers than winners (and we'll have retro AI by next week).
@inthehands ie. you still have to do your homework, and getting something else to do it for you isn't likely to get it right

@inthehands

This is how competitive systems learn: the language of death. In this case corporate death.

Politics careens from one failure to the next. Movement death, often learning something, but it only lasts a while.

Biological evolution: same thing. Species death.

Medicine too, though we work very hard to deny it. Death.

Technology is wrong more often than right. Progress still happens because the failed bubbles guide us violently.

@EubieDrew ML training is also competitive, of course its outputs don't actually value anything.

@inthehands I think there is one exception--for a lot of people in creative fields who may have some kind of borderline ADHD condition, getting past the blank page or the digital equivalent is a real struggle. And if there's something that can push them past that step from nothing to something, they'll find it useful.

There's a powerful temptation to just use version zero, though, especially if you're not the creator but the person paying the creator.

@mattmcirvin Indeed, I ran a successful exercise much along these lines with one of my classes (see student remarks downthread):
https://hachyderm.io/@inthehands/109479808455388578

I think there really is a “there” there with LLMs; it just bears close to no resemblance to the wildly overhyped Magic Bean hysteria currently sweeping biz. Generating bullshit does actually have useful applications. But until the dust settles, how much harm will it cause?

Paul Cantrell (@[email protected])

Attached: 3 images OK, trying an experiment with my Programming Languages class! • Have an AI generate some of your writing assignment. • Critique its output. Call BS on its BS. Assignment details in screenshots below. I’ll let you know how it goes. (Here are the links from the screenshots:) Raw AI Text: https://gist.github.com/pcantrell/7b68ce7c5b2e329543e2dadd6853be21 Comments on AI Text: https://gist.github.com/pcantrell/d51bc2d4257027a6b4c64c9010d42c32 (Better) Human Text https://gist.github.com/pcantrell/f363734336e6063f61e451e2658b50a6 #ai #chatgpt #education #writing #highered #swift #proglang

Hachyderm.io
@inthehands in a discussion I saw someone noted that a "removing AI from workflow" consulting company would soon be viable.
@StevenSavage @inthehands Let's be honest, management isn't trying to add AI to workflows, they are trying to replace workflows with AI.

@inthehands

One of these is not like the others. Here’s how I think about it: many processes can be thought of as generating a large number of leads and then screening them to find the good ones. In classic AI this a generate-and-test algorithm. It’s vital that your testing works or you will get bad answers.

Using AI for the “generate” phase is not nearly as bad as using it for screening phase, provided that your tests are good. And we do know how to test our code, don’t we?

@skybrian In the cases of hiring, coding, and writing, there is a point where the number of “leads” is high enough, and the quality is low enough, where the cost of screening them is •worse• than starting from scratch. And I think a lot of people are huffing a lot of fumes right now about just how soon that point comes.

@inthehands I don’t think that’s always the case for coding. It depends on the domain, but I find Copilot’s autocomplete to be good enough to miss it when I turn it off. And lately I’ve been asking GPT4 questions about how to do simple type-level programming in TypeScript and it beats figuring it from the manual or dodgy StackOverflow answers.

Verifying that a suggestion works isn't hard since it’s the same thing I’d do to verify that my own code works. I’d have to do that anyway.

@inthehands The same applies in translation - both AI and human. It's easy to produce something almost, but not entirely like a piece of natural language, but the flow won't be quite right, the word choice clunky, the rhythm subtly off. (I was tempted to use only two examples in that last sentence as a demonstration!)
@inthehands As I say all too often, this happened about five years ago in the translation industry (I'm a technical translator). It blasted a hole in my income and even years later, managers are still just as excited about cheap work as they ever were. More so, really.

@vivtek Yeah. Same thing happening to illustrators right now.

One of the problems that we’re up against is that the problems of shoddy work aren’t always immediately apparent. Doing it the crappy way can look like cost savings for a very long time — often long enough for investors to cash out.

Don’t know what the answer to that is. It’s not just AI; it’s a deep systemic rot.

@inthehands I don't think it's AI at all! I use machine translation tools every day. It's large agencies unilaterally deciding *they*'ll collect the dividend, and *they*'ll decide how much "easier" the work is with AI support. "We'll only pay you 60% because we think you'll agree you'll be working less" - and if you don't agree, there's the door.

But they come back and demand *more* work to make up for MT errors, and are unable to conceive of improving the MT engine/workflow. Why would they?

@inthehands People are misunderstanding those 80-20-rules (80% of time will create 20% of results).
They are not a promise of an untapped source of productivity, they are a warning about an inevitable drain of productivity.