HomeArtificial IntelligenceGoogle hopes to resolve Gemini's historical image diversity problem inside a couple...

Google hopes to resolve Gemini's historical image diversity problem inside a couple of weeks

According to DeepMind founder Demis Hassabis, Google hopes it’s going to soon have the ability to unlock the power of its multimodal generative AI tool Gemini to represent people. The ability to answer requests for images of individuals needs to be back online within the “next few weeks,” he said today.

Google shut down the Gemini feature last week after users identified that the tool produced historically contradictory images, corresponding to depicting the founding fathers of the United States as a various group of individuals quite than simply white men.

Hassabis answered questions on the Snafu product during an on-stage interview on the Mobile World Congress in Barcelona today.

When asked by a moderator, Wired's Steven Levy, to elucidate what went incorrect with the image generation function, Hassabis shunned providing an in depth technical explanation. Instead, he suggested that the issue was brought on by Google's failure to discover cases where users are essentially after what he called a “universal representation.” He also said the instance points to “nuances that include advanced AI.”

“This is an area that all of us grapple with. So, for instance, should you type a prompt that asks, “Give me an image of an individual walking a dog or a nurse in a hospital,” right, in those cases you clearly want some sort “universal representation”. “Especially once you consider that as Google we serve greater than 200 countries, you understand, every country all over the world – so that you don't know where the user is coming from, what their background is, or what context they're in positioned. So you need to show a form of universal spectrum of possibilities there.”

Hassabis said the problem boils right down to a “well-intentioned feature” – to advertise diversity in Gemini's image expressions of individuals – that has been applied “too bluntly all over the place.”

Prompts that ask for content about historical people should “naturally” result in a “much narrower distribution that you simply return,” he added, hinting at how Gemini might approach people prompts in the longer term.

“Historical accuracy is in fact essential to us. Therefore, now we have taken this feature offline while we troubleshoot the problem and hope to have it back online in a really short time frame. Next few weeks, next few weeks.”

In response to a follow-up query about easy methods to prevent generative AI tools from being misused by malicious actors corresponding to authoritarian regimes searching for to spread propaganda, Hassabis didn’t have an easy answer. The problem is “very complex,” he said — and certain requires a mobilization and response from all the society to set and implement limits.

“There must be really essential research and debate – including with civil society and governments, not only with tech firms,” he said. “It is a socio-technical issue that affects everyone and everybody should participate within the discussion. What values ​​should these systems have? What would they represent? How do you prevent malicious actors from accessing the identical technologies and misusing them, as you’re talking about, for harmful purposes that weren’t intended by the designers of those systems?”

He addressed the challenge of open source general-purpose AI models, which Google also offers, adding: “Customers wish to use open source systems that they will fully control.” . But then the query becomes: With these increasingly powerful systems, how do you be certain that what persons are using downstream isn't harmful?

“I believe this isn’t an issue today since the systems are still relatively young. But once you fast forward three, 4 or five years and begin talking about next-generation systems with planning capabilities and the power to act on the earth and solve problems and goals, I believe society really must think seriously about these issues – from “What happens if this spreads after which bad actors, from individuals to rogue states, can make the most of it?”

During the interview, Hassabis was also asked about his thoughts on AI devices and where the mobile market is perhaps headed as generative AI continues to drive recent developments here. He predicted a wave of “next-generation intelligent assistants” that shall be useful in people's on a regular basis lives, quite than the “gimmicks” of previous generations of AI assistants, which he said would even overpower the mobile hardware people carry with them. could change.

“I believe there’ll even be questions on which variety of device is the correct one,” he suggested. “But will the phone even have the right form think about greater than five years? Maybe we want glasses or other things in order that the AI ​​system can actually recognize among the context that you simply are in and so be much more helpful in your each day life. So I believe there are a whole lot of amazing things to invent.”


Please enter your comment!
Please enter your name here

Must Read