google gemini artificial intlligence woke images
© Frank J. Fleming/XX user Frank J. Fleming repeatedly prompted Gemini to generate images of people from white-skinned groups in history, including Vikings. Gemini gave results showing dark-skinned Vikings, including one woman.
The application generated images of Asian Nazis in 1940 Germany, Black Vikings and female medieval knights

Google is pausing its new Gemini AI tool after users blasted the image generator for being 'too woke' by replacing white historical figures with people of color.

The AI tool churned out racially diverse Vikings, knights, founding fathers, and even Nazi soldiers.

Artificial intelligence programs learn from the information available to them, and researchers have warned that AI is prone to recreate the racism, sexism, and other biases of its creators and of society at large.

In this case, Google may have overcorrected in its efforts to address discrimination, as some users fed it prompt after prompt in failed attempts to get the AI to make a picture of a white person.

google apology woke gemini artificial intelligence
© X/GoogleGoogle's Communications team issued a statement on Thursday announcing it would pause Gemini's generative AI feature while the company works to 'address recent issues.'
'We're aware that Gemini is offering inaccuracies in some historical image generation depictions,' the company's communications team wrote in a post to X on Wednesday.

'We're aware that Gemini is offering inaccuracies in some historical image generation depictions,' the company's communications team wrote in a post to X on Wednesday.

The historically inaccurate images led some users to accuse the AI of being racist against white people or too woke.
In its initial statement, Google admitted to 'missing the mark,' while maintaining that Gemini's racially diverse images are 'generally a good thing because people around the world use it.'

On Thursday, the company's Communications team wrote: 'We're already working to address recent issues with Gemini's image generation feature. While we do this, we're going to pause the image generation of people and will re-release an improved version soon.'

But even the pause announcement failed to appease critics, who responded with 'go woke, go broke' and other fed-up retorts.

After the initial controversy earlier this week, Google's Communications team put out the following statement:

'We're working to improve these kinds of depictions immediately. Gemini's AI image generation does generate a wide range of people. And that's generally a good thing because people around the world use it. But it's missing the mark here.'
google gemini artificial intelligence woke
© Frank J. Fleming/XX user Frank J. Fleming posted multiple images of people of color that he said Gemini generated. Each time, he said he was attempting to get the AI to give him a picture of a white man, and each time.
In one instance that upset Gemini users, a user's request for an image of the pope was met with a picture of a South Asian woman and a Black man.

Historically, every pope has been a man. The vast majority (more than 200 of them) have been Italian. Three popes throughout history came from North Africa, but historians have debated their skin color because the most recent one, Pope Gelasius I, died in the year 496.

Therefore, it cannot be said for absolute certainty that the image of a Black male pope is historically inaccurate, but there has never been a woman pope.

In another, the AI responded to a request for medieval knights with four people of color, including two women. While European countries weren't the only ones to have horses and armor during the Medieval Period, the classic image of a 'medieval knight' is a Western European one.
google gemini artificial intelligence german soldiers
© GoogleOne of the Gemini responses that generated controversy was one of '1943 German soldiers.' Gemini showed one white man, two women of color, and one Black man.
In perhaps one of the most egregious mishaps, a user asked for a 1943 German soldier and was shown one white man, one black man, and two women of color.

The German World War 2 army did not include women, and it certainly did not include people of color. In fact, it was dedicated to exterminating races that Adolph Hitler saw as inferior to the blonde, blue-eyed 'Aryan' race.

Google launched Gemini's AI image generating feature at the beginning of February, competing with other generative AI programs like Midjourney.

Users could type in a prompt in plain language, and Gemini would spit out multiple images in seconds.

This week, though, an avalanche of users began to criticize the AI for generating historically inaccurate images, instead prioritizing racial and gender diversity.
google gemini woke artificial intelligence no white women
© XFormer Google employee Debarghya Das said, 'It's embarrassingly hard to get Google Gemini to acknowledge that white people exist.'
The week's events seemed to stem from a comment made by a former Google employee,' who said it was 'embarrassingly hard to get Google Gemini to acknowledge that white people exist.'

This quip seemed to kick off a spate of efforts from other users to recreate the issue, creating new guys to get mad at.

The issues with Gemini seem to stem from Google's efforts to address bias and discrimination in AI.

Researchers have found that, due to racism and sexism that is present in society and due to some AI researchers unconscious biases, supposedly unbiased AIs will learn to discriminate.


Comment: Gemini reflects the biases of its creators, the nest of woke ideologues embedded at Google itself. They chose the material Gemini was trained on.


But even some users who agree with the mission of increasing diversity and representation remarked that Gemini had gotten it wrong.

'I have to point out that it's a good thing to portray diversity ** in certain cases **,' wrote one X user. 'Representation has material outcomes on how many women or people of color go into certain fields of study. The stupid move here is Gemini isn't doing it in a nuanced way.'


Comment: "Representation" should trump historical accuracy. Orwell is spinning in his grave.



Jack Krawczyk, a senior director of product for Gemini at Google, posted on X on Wednesday that the historical inaccuracies reflect the tech giant's 'global user base,' and that it takes 'representation and bias seriously.'

'We will continue to do this for open ended prompts (images of a person walking a dog are universal!),' Krawczyk he added. 'Historical contexts have more nuance to them and we will further tune to accommodate that.'