Back in February, Google paused its AI-powered chatbot Gemini’s ability to generate images of people after users complained of historical inaccuracies . Told to depict “a Roman legion,” for example, Gemini would show an anachronistic group of racially diverse soldiers while rendering “Zulu warriors” as stereotypically Black.
Google CEO Sundar Pichai apologized, and Demis Hassabis, the co-founder of Google’s AI research division DeepMind, said that a fix should arrive “in very short order” — within the next couple of weeks. But we’re now well into May, and the promised fix remains elusive.
Google touted plenty of other Gemini features at its annual I/O developer conference this week, from custom chatbots to a vacation itinerary planner and integrations with Google Calendar, Keep and YouTube Music. But image generation of people continues to be switched off in Gemini apps on the web and mobile, confirmed a Google spokesperson.
So what’s the holdup? Well, the problem’s likely more complex than Hassabis alluded to.
The datasets used to train image generators like Gemini’s generally contain more images of white people than people of other races and ethnicities, and the images of non-white people in those datasets reinforce negative stereotypes . Google, in an apparent effort to correct for these biases, implemented clumsy hardcoding under the hood. And now it’s struggling to suss out some reasonable middle path that avoids repeating history .
Will Google get there? Perhaps. Perhaps not. In any event, the drawn-out affair serves as a reminder that no fix for misbehaving AI is easy — especially when bias is at the root of the misbehavior.
We’re launching an AI newsletter! Sign up here to start receiving it in your inboxes on June 5.