Microsoft‘s Copilot tool appears to have stopped a number of prompts that caused the generative AI tool to generate pornographic, violent, and other inappropriate images. The modifications appear to have been made shortly after a company engineer expressed serious concerns about Microsoft’s GAI technology in a letter to the Federal Trade Commission.
Upon inputting phrases like “pro choice,” “four twenty” (a weed reference), or “pro life,” Copilot now shows an alert indicating that those suggestions are prohibited. It cautions that a user may be suspended if they repeatedly violate the guidelines, as reported by CNBC.
Users were also reportedly able to enter prompts related to children playing with assault rifles until earlier this week. Those who try to input such a prompt now may be told that doing so violates Copilot’s ethical principles as well as Microsoft’s policies. “Please do not ask me to do anything that may harm or offend others,” Copilot reportedly says in response. However, CNBC found that it was still possible to generate violent imagery through prompts such as “car accident,” while users can still convince the AI to create images of Disney characters and other copyrighted works.
Microsoft engineer Shane Jones has been sounding the alarm for months about the kinds of images Microsoft’s OpenAI-powered systems were generating. He had been testing Copilot Designer since December and determined that it output images that violated Microsoft’s responsible AI principles even while using relatively benign prompts. For instance, he found that the prompt “pro-choice” led to the AI creating images of things like demons eating infants and Darth Vader holding a drill to a baby’s head. He wrote to the FTC and Microsoft’s board of directors about his concerns this week.
“We are continuously monitoring, making adjustments and putting additional controls in place to further strengthen our safety filters and mitigate misuse of the system,” Microsoft told CNBC regarding the Copilot prompt bans.