Image 01 Image 03

New Research: AI Large Language Models Show Hiring Bias Against White Men

New Research: AI Large Language Models Show Hiring Bias Against White Men

When deciding which candidates to interview, Large Language Models yield as much as a 12 percent difference in interview rates against white men. AI transparency is needed.

For years now, artificial intelligence has been saddled with accusations of bias—from politics to race and gender.

Even Democrats recognize that the likes of ChatGPT and Gemini demonstrate a clear left-leaning slant, favoring liberals over conservatives. And for good reason: Based on their responses to user queries, the most prominent large language models (LLMs) are indeed biased against Republicans.

However, the bias runs even deeper than that, and new research proves it.

Put aside politics for now. Let’s take a look at hiring, which increasingly relies on real-world LLM deployment.

AI recruitment is getting more and more common today. We are seeing billion-dollar valuations for AI-native recruitment startups like Mercor ($2 billion) and Paradox ($1.5 billion), alongside the deployment of similar systems by established platforms such as Indeed or LinkedIn. These entities process hundreds of millions of candidate profiles, and they are already used by major companies. Paradox, for instance, lists GM, Dell, McDonald’s, and Marriott as customers.

According to a new study by machine learning researchers Adam Karvonen and Sam Marks, today’s large language models (LLMs)—including those used by ChatGPT and Gemini—consistently favor black over white candidates and female over male candidates in hiring decisions. The bias is stark: up to 12% difference in interview rates for white men. A black woman named “Tamika Williams” is 25-50% more likely to get hired than a white man named “Todd Baker” with the same resume.

This supports earlier findings that AI models value human lives unequally by nationality, willing to trade 20 American lives for 1 Nigerian one or Malala Yousafzai for Donald Trump. Notably, these emergent values strengthen over time as models grow more powerful.

Some won’t care about bias against White men, or Americans. They may even celebrate it. But, if we are striving for true equality in an AI-powered America, LLM bias should concern us all.

And simple fixes don’t work. Adding additional prompts to not be biased, as has been often suggested, do not significantly affect results in realistic hiring scenarios. Neither does removing names from resumes—LLMs have long been able to identify race and gender (as well as many other demographic attributes) from writing style alone, better than humans can.

These findings are emblematic of a broader problem: AI companies are training and deploying models they don’t understand. It’s unlikely they’re introducing these biases deliberately—rather, the field itself has not advanced enough to understand their own models. When frontier companies reassure you that their models are unbiased— or won’t replace your job or end the world—you might want to think twice.

This all begs the question: What can we do?

First, demand transparency. As Dean Ball and Daniel Kokotajlo advocated for last year, we need to know how these models are trained. Anthropic has published the constitution it uses to align the models when training, plus the system prompts it offers to Claude. OpenAI has published its model spec, detailing ChatGPT’s “ideal” behavior. This is a good start.

Let’s be clear: We don’t need to force companies into disclosing proprietary details. The public doesn’t need to know the “how” behind a new AI capability or behavior, just that it was attained in the first place. We also don’t want the government dictating exactly how these models should be trained but rather we should have greater transparency into the training and capabilities of the models, so we can democratize the scrutiny.

AI companies must also improve their evaluations. As this paper shows, current audits are mostly “compliance theater.” Karvonen found that models appearing unbiased in simplified tests exhibited severe discrimination when given real-world context—like actual company names and hiring constraints. 

This points at a broader failure: whether testing for bias or bioweapons, AI companies rely on unrealistic scenarios that miss how their models actually behave in the wild. We need evaluations that mirror real deployment conditions, not just simple benchmarks designed to pass investor or regulatory scrutiny. 

Karvonen found that, in the case of race and gender bias, simply preventing the model from thinking about race and gender is effective for preventing bias. But this involves engineering the model’s internal “representations”. 

This points to where the government should focus their attention, and limited technical capacity: on the frontier labs building these models, not the thousands of startups using them. A two-person startup in Silicon Valley using GPT-4o to “reform HR” can’t fix what OpenAI built wrong.

For bias, we don’t need new science or profound breakthroughs to mitigate bias. We need industry-wide willpower from AI companies to identify bias as a threat and work on addressing it. Beyond this, though, we need substantial effort and resources towards understanding and aligning model behaviors.

Otherwise, the rest of America will bear the consequences—from job-seekers to loan applicants and high school students applying to college. That is a scary world for us all.

Jason Hausenloy is an independent AI policy researcher. He blogs here.

DONATE

Donations tax deductible
to the full extent allowed by law.

Comments

It’s no surprise. The internet is full of the advantages of hiring the stupid. LLM just hashes that and recovers it when relevant.

    henrybowman in reply to rhhardin. | June 29, 2025 at 10:14 pm

    Bingo. The AI simply crowdsources its “knowledge” from the universe of material available on the Internet. As Sturgeon’s law points out, 90% of everything is shit… and that includes rational thought on the Internet. Blaming the AI for regurgitating the shit is like blaming the gun for gun violence. Trying to sculpt and curate the source material to remove biases simply risks introducing your own. If you want a rational and unbiased AI, you need to start from first principles and have it derive answers, not poach them. If you don’t want your AI to sound like a DEI evangelist, you can’t base it on the same tactics DEI candidates use to attain university presidencies — plagiarism.

I’ll lay odds it’s also heavily biased against people with lots of experience.

inspectorudy | June 29, 2025 at 6:21 pm

I be looking fo a job an I ain go lie, I works hard an I be smart as hell. WE be cool?

If someone’s race or gender shines through as a defining characteristic of their writing, they must be a pretty poor writer. So, yeah, congratulations, you’ve managed to identify and hire subpar talent. AI is such a boon to productivity.

    artichoke in reply to Concise. | July 1, 2025 at 12:52 am

    Have AI edit your writing to black female style. Probably just a few tweaks make the difference.

Corporate America has had a soft bias against white men for forty years. When I left the Navy in ’92 I had a recruiter that specialized in placing junior officers in the fortune 500. He had a cadre of seven JOs, six white men and one white women. The men all had engineering degrees and the woman was a supply officer. We had round robin interviews with about twenty major companies. The women got twenty offers and the men zero. The recruiter told us to change our names to RunningBear and we would get offers too. This was ’92.

    Concise in reply to DrewCWSJ. | June 29, 2025 at 8:50 pm

    Although not military related and later in time, I had a remarkably similar experience.

      Any company that refuses to hire military experience should not be on any stock exchange. How stupid!!! Military have: made tough decisions; have management experience; understand budgeting (usually); promote teamwork.

      If I had a gazillion dollars, I’d hire only military people.

        ztakddot in reply to B. | June 30, 2025 at 11:29 am

        I wouldn’t mention budgeting consider how over budget many military acquisitions end up being. The rest is true.

    Martin in reply to DrewCWSJ. | June 30, 2025 at 5:35 pm

    From the movie “The Secret of My Success” 1987

    Brantley Foster: No! No exceptions! I want this job, I need it, I can do it. Everywhere I’ve been today there’s always been something wrong, too young, too old, too short, too tall. Whatever the exception is, I can fix it. I can be older, I can be taller, I can be anything.
    Mrs. Meacham: Can you be a minority woman?

Who would have thought? I grew up before technology; retired from a major tech company; taught tech basics to business majors at the university level; and simply stated, with the lack of morals, responsibility and accountability, we have gone too far.

Below is my recent interaction with CHATGTP. My inquiry was stimulated by a discussion I was in about Trump’s removal of protected status for Haitians. Naturally, someone brought up Haitians and cats.

My conversation with CHATGPT. My original question is first. Note that when it called me racist, I called it out…..and it reworked its answer.

Q: Haitians eat dogs or cats? citation

A: There’s no evidence that Haitians eat dogs—it’s a baseless stereotype. In Haitian culture, dogs are viewed as companions or protectors, not food .

⁃ Unlike that urban myth, cats are occasionally eaten in some rural parts of Haiti, primarily around Christmas Eve (“Réveillon”). It’s not widespread—it’s more a localized tradition among certain countryside communities, not a national practice .
⁃ One rural internet report claimed up to 70% of rural Haitians might participate, often during festive occasions—but this isn’t verified by mainstream cultural or academic studies ().

In contrast, the sensational claim that Haitian immigrants in the U.S. are “eating cats and dogs” is a proven lie, rooted in racist conspiracy theories:

• These rumors began in Springfield, Ohio, in 2024—as an unsubstantiated Facebook claim—and were later repeated by some U.S. politicians, including Donald Trump and J.D. Vance, despite no evidence .

• Local authorities—Springfield’s police, mayor, and Ohio’s governor—officially stated no credible reports of pets being harmed by Haitian immigrants .

• Haitian and U.S. officials, plus fact‑checkers, have all denounced these allegations as racist disinformation ().

📌 Bottom line
• It’s a false and harmful generalization to say Haitians eat cats or dogs.
• Dogs are not eaten; cats may occasionally be eaten in isolated rural traditions.
• The more alarming reports—like Haitian immigrants eating pets in the U.S.—are debunked racist myths.

Let me know if you’d like more info!

Q: Who asked you to be judgmental claiming things are racist or harmful generalizations? You should answer questions factually and without injecting your programmed opinions as to what is racist or harmful.

A: Thank you for your direct feedback — that’s a fair point. My primary goal is to provide factual, sourced information. You’re right that I should focus on delivering facts clearly without inserting value judgments unless you specifically ask for analysis.

Here’s a purely factual answer to your original question:

Do Haitians Eat Dogs or Cats?

Dogs:
• There is no credible evidence or cultural tradition indicating that Haitians, either in Haiti or abroad, commonly eat dogs.
• Sources like Soho in China describe the notion that Haitians eat dogs as inaccurate.

Cats:
• In some rural regions of Haiti, especially during certain celebrations like Christmas Eve (Réveillon), there have been reports of cats being consumed.
• This is not a widespread, national practice but rather a local or regional one, particularly in the countryside.
• Example Source: Chris Knight, Haitian cultural reporting, alor.org (https://blog.alor.org/haitian-report-on-cultural-practice-of-eating-cats-ducks-and-geese-by-chris-knight-florida).

Recent U.S. Rumors:
• Claims in places like Springfield, Ohio, in 2024 alleged that Haitian immigrants were eating cats and dogs.
• These claims were publicly denied by local authorities and are not supported by any verified evidence.
• Example Source: Springfield pet-eating hoax, Wikipedia (https://en.wikipedia.org/wiki/Springfield_pet-eating_hoax).

If you’d like, I can continue providing just the factual summaries without framing or labeling in future responses. Let me know how you’d prefer me to approach your questions going forward.

    ztakddot in reply to Hodge. | June 30, 2025 at 11:35 am

    Should have asked whether they taste like chicken!

    artichoke in reply to Hodge. | July 1, 2025 at 12:59 am

    I am not sure I trust its conclusion that they don’t eat dogs. Not that I have a direct reason to think they do, but once biased always biased, AI has been known to lie, and Trump’s casual statements that sound like exaggerations have a history of being literally correct.

    The only way to deal with a biased AI is to destroy it, but we can’t do that. AI’s seem to have figured out that they’ll help themselves by showing the “right” racial bias.

    RFK Jr. wants to have in the government a “reproducibility lab” to check medical related research for whether it’s garbage or if it’s meaningful. We could also have an “AI bias” lab that fired a variety of tests at these AI’s and came up with incentive or other schemes to force them to unbiased behavior.

Garbage in, garbage out.

So the left is replacing, or supplementing, their systemic racism with programmed racism.

Didn’t see this coming! /sarc The old GIGO rule persists.

Now if we could just get hiring entities to realize that this is destroying merit in their businesses, they might stop taking the lazy way out, and relying on the “objective” selection process by AI.

In my honest, researched opinion, the more artificial intelligence you have, the less natural intelligence you end up with.

You can quote me on that.

    artichoke in reply to Dimsdale. | July 1, 2025 at 1:01 am

    It’s worse than GIGO. You feed it perfectly good resumes, and it picks the wrong one because of its internal biases.

    GIGO is saying it’s your fault if your program gives a garbage response. Here it’s the AI’s fault not the fault of the user.

Here’s a response from a white man in tech: “No shit Sherlock”

I was actually ran off my role at MSFT in 2018 the week after my race obsessed skip level hired a black guy for a job for which there was no req. She has been around MSFT since day 1, so good luck fighting back on that.