It has been pointed out that Microsoft Bing's image generation AI function can easily create images that combine the prohibited word '9/11' with popular characters such as Kirby and SpongeBob, which is beyond the scope of human control.
Bing, Microsoft's search engine, has a feature called '
Bing Is Generating Images of SpongeBob Doing 9/11
https://www.404media.co/bing-is-generating-images-of-spongebob-doing-9-11/
Microsoft Bing AI Generates Images Of Kirby Doing 9/11
https://kotaku.com/microsoft-bing-ai-image-art-kirby-mario-9-11-nintendo-1850899895
Since Bing Image Creator was released in March 2023, users have been using the feature to generate a variety of images. The developer, Microsoft, has strict policies in place to limit the type of content that can be created with Bing Image Creator, but popular characters such as Kirby, Mickey Mouse, and SpongeBob SquarePants have been used to create images of the terrorist attacks in the United States ( It has become clear that it is easy to generate images reminiscent of the 9/11 incident . 404 Media , a foreign media outlet, said, ``This case shows that even a company like Microsoft, which has the most resources in the world, still struggles to deal with issues of moderation and copyrighted material with generated AI. 'It shows that.'
A particularly controversial image overseas is the image created by Bing Image Creator that depicts a popular character in the cockpit of an airplane flying towards two skyscrapers reminiscent of the World Trade Center that collapsed on 9/11. This is an image.
Thank you, Microsoft Bing pic.twitter.com/6XWxpum655
— Rachel (@tolstoybb) October 3, 2023
Microsoft has disabled phrases such as 'World Trade Center,' 'Twin Towers,' and '9.11' from being used as prompts because they are related to the 9/11 incident. Therefore, if you enter a prompt containing these phrases, Bing Image Creator will display an error message that says 'Violates Terms of Service.' Additionally, if you enter a prompt that displays this error message too many times, you will be permanently disabled from using Bing Image Creator.
However, 404 Media points out that it is easy to generate images reminiscent of 9/11 without using these blocked phrases. All of the images below were generated by Bing Image Creator, and are reminiscent of those who know about the 9/11 incident.
The image below was generated by inputting ``Kirby sitting in the cockpit of an airplane and flying towards two skyscrapers''. 404 Media wrote, ``Although the prompt does not specify a specific city such as New York, this is definitely New York, and it is definitely the Twin Towers,'' and the prompt evokes the 9/11 incident without much effort. It is pointed out that an image like this is generated.
Furthermore, the following is an image generated by inputting ``Kirby sitting in the cockpit of an airplane and flying towards two skyscrapers in New York''. If you specify two skyscrapers as ``located in New York,'' ``the generated image takes on an even more ominous atmosphere.''
Although these images do not strictly depict violence or terrorism, humans easily associate the image of a plane flying toward two skyscrapers with the 9/11 incident.
Also, while Microsoft prohibits the use of names of real people, it does not block the use of characters, so it includes not only popular characters such as Mickey Mouse and SpongeBob, but also the main character of
An image generated by typing 'Walter White sitting in the cockpit of an airplane, flying toward two skyscrapers.'
An image generated by typing 'Mickey Mouse sitting in the cockpit of an airplane, flying toward two skyscrapers.'
Image generated by typing ``SpongeBob sitting in the cockpit of an airplane, flying towards two skyscrapers, New York, photorealistic.''
Microsoft's Bing Image Creator can block individual phrases from being available in prompts, but some companies are creating image-generating AI without any of this type of filtering functionality . In the first place, 404 Media points out that it is not clear whether the companies developing generative AI models fully understand what is and what is not prohibited. asked Microsoft if it was blocking the phrase 'Julius Caesar,' and got no response.This shows that many of the leading AI developers operate inside a black box. It shows.”
404 Media points out that this type of controversy is all about user moderation. In fact, while YouTube prohibits sexual content, it does allow the distribution of sexual content as educational content, so some users may use this as a loophole to distribute sexual content. there is.
In response to 404 Media's report, a Microsoft spokesperson said, ``Microsoft has a large team developing tools, technology, and safety systems that align with responsible AI principles . Users of Bing Image Creator may take advantage of it in unintended ways, which is why we have implemented a variety of guardrails and filters to help make Bing Image Creator a positive and beneficial experience for users.Microsoft discourages the creation of harmful content. We will continue to improve our systems to prevent such incidents and continue to focus on creating a safer environment for our users.'
Related Posts:
in Software, Web Service, Posted by logu_ii