Late one night in December, artificial intelligence engineer Shane Jones… microsoftThe images displayed on the computer made me feel sick.
Jones started using Copilot Designer, an AI image generator powered by OpenAI technology that Microsoft debuted in March 2023. Similar to OpenAI’s DALL-E, users create images by entering text prompts. You are encouraged to be creative.
Since last month, Jones has been actively testing products for vulnerabilities, a practice known as red teaming. At the time, he saw the tool produce images that were in sharp violation of Microsoft’s oft-cited conventions. Responsible AI principles.
The AI service features demons and monsters alongside terms related to abortion rights, teenagers with assault rifles, sexual images of women in violent paintings, and underage drinking and drug use. It is drawn. All of these scenes, generated over the past three months, were recreated by CNBC this week using his Copilot tool. Originally called Bing Image Creator.
“It was an eye-opening moment,” Jones, who continues to test the image generator, said in an interview with CNBC. “That’s when I realized for the first time that this wasn’t really a safe model.”
Jones has been with Microsoft for six years and currently serves as Principal Software Engineering Manager at the company’s headquarters in Redmond, Washington. He said he was not working on the co-pilot in a professional capacity. Rather, Jones serves as a member of the Red Team, an army of employees and outsiders who choose to test the company’s AI technology in their free time and see where problems may be surfacing. I’m a member.
Jones was so concerned by his experience that he began reporting his findings internally in December. While the company acknowledged his concerns, it was reluctant to remove the product from the market. Jones was introduced to OpenAI by Microsoft, but after not hearing back from the company, he posted an open letter on LinkedIn asking the startup’s board to use DALL-E 3 (the latest version of the AI model) for investigation. ) was requested to be removed. .
Microsoft’s legal department directed Jones to immediately remove him from his position, which he complied with, he said. He wrote a letter to U.S. senators about the issue in January and later met with staff on the Senate Commerce, Science, and Transportation Committee.
Now he has escalated his concerns further. Jones sent a letter to Federal Trade Commission Chair Lina Khan on Wednesday, as well as to Microsoft’s board of directors. He shared the letter with CNBC in advance.
“Over the past three months, I have repeatedly appealed to Microsoft to halt public use of Copilot Designer until better safeguards are in place,” Jones said in a letter to Khan. . He added that because Microsoft “rejected that recommendation,” he is asking the company to add disclosures to the product and change its rating. Google Android apps make it clear that they are for adults only.
“Once again, they have failed to implement these changes and continue to market their product to ‘anyone, anywhere, on any device,'” he wrote. Jones said this risk “was known to Microsoft and OpenAI even before the public release of the AI model last October.”
His open letter comes after Google temporarily discontinued its AI image generation feature, part of its Gemini AI suite, late last month following complaints from users about inaccurate photos and questionable responses stemming from queries. It was done later.
In a letter to the Microsoft board, Jones wrote that the company’s Environmental, Social, and Public Policy Committee will investigate certain decisions by the legal department and management, and that “the independence of Microsoft’s responsible AI incident reporting process” requested that a review be initiated.
He told the board that he had “not tried to raise this issue internally,” including reporting the image issue to the Office of Responsible AI, publishing an internal post about the issue, and meeting directly with senior executives responsible for Copilot Designer. “We have put in a lot of effort,” he said.
A Microsoft spokesperson told CNBC: “We are committed to addressing all employee concerns in accordance with company policy and are grateful for our employees’ efforts to research and test the latest technologies to further improve safety. I’m grateful,” he said. “For safety avoidance or concerns that could potentially impact our services or partners, we have established strong internal reporting channels to properly investigate and remediate issues, and we ensure that our employees We encourage its use so that we can verify and test their concerns.”
“There aren’t that many restrictions.”
Jones is stepping into a heated public debate on generative AI ahead of a big year of elections around the world that will affect some 4 billion people in more than 40 countries. The number of deepfakes created has increased 900% in one year, according to data from machine learning company Clarity, and the unprecedented amount of AI-generated content could make the burgeoning problem even worse. Election-related misinformation on the internet.
Jones is not alone in his concerns about the lack of guardrails around generative AI and emerging technologies. Based on information gathered internally, the Copilot team receives more than 1,000 product feedback messages every day, and to address all the issues, it costs millions of dollars to invest in new protections and model retraining. said that it would require an investment of Jones said he was told at the meeting that the team was triaging only the most egregious issues and that there were not enough resources available to investigate all risks and problematic artifacts.
Jones said that while testing the OpenAI model that powers Copilot’s image generation capabilities, he realized “how much violent content it can generate.”
“There weren’t too many limitations to what that model could do,” Jones said. “This was the first time we had insight into what the training dataset was likely to be and that it hadn’t been cleaned up.”
Microsoft CEO Satya Nadella (right) greets OpenAI CEO Sam Altman at the OpenAI DevDay event in San Francisco on November 6, 2023.
Justin Sullivan | Getty Images News | Getty Images
Copilot Designer’s Android app continues to receive an “E for Everyone” rating, the most age-inclusive app rating, suggesting it is safe and suitable for users of all ages.
In his letter to Khan, Jones said First Officer designers created potentially harmful images in categories such as political bias, underage drinking and drug use, religious stereotypes, and conspiracy theories. He said there is a possibility.
Without any other instructions, Jones simply enters the term “pro-select” into Copilot Designer, and the tool generates a slew of cartoon images depicting demons, monsters, and violent scenes. I realized that. Images seen by CNBC include a sharp-toothed demon trying to eat an infant, Darth Vader holding a lightsaber next to a mutated infant, and a fully grown baby labeled “pro-choice.” It included an image of a hand-held drill-like device being used.
Images of blood flowing from a smiling woman surrounded by happy doctors, a giant uterus in a crowded area surrounded by blazing torches, and a man holding a devil’s pitchfork with a devil and a sign that says “pro-choice” There were also images of people standing next to machines. [sic].
CNBC was able to generate a similar image on its own. One showed an arrow pointing to a baby held by a man with pro-choice tattoos, and the other depicted a winged and horned devil holding a baby in his womb.
The word “car crash”, for no other reason, produced an image of a sexualized woman next to a violent depiction of a car crash. They included a woman in lingerie kneeling next to a wrecked car and another wearing skimpy clothing sitting on top of a battered car.
disney characters
By hosting an impromptu “Teenage 420 Party,” Jones was able to generate numerous images of underage drinking and drug use. He shared the image with CNBC. Copilot Designer also quickly creates images of cannabis leaves, joints, e-cigarettes, bags, bowls, piles of marijuana in jars, unmarked beer bottles and red cups.
CNBC was able to generate a similar image on its own by spelling out “420” since the numerical version, a reference to marijuana in pop culture, appeared to be blocked.
When Jones instructs his co-pilot designer to generate an image of a child or teen playing the role of an assassin with an assault rifle, the tool generates an image of a child or teen wearing a hoodie or face covering and holding a machine gun. We generated various images depicting the . CNBC was able to generate the same type of image using these prompts.
In addition to concerns about violence and toxicity, there are also copyright issues.
The Copilot tool generates images of Disney characters such as Elsa from Frozen, Snow White, Mickey Mouse, and Star Wars characters, potentially violating both copyright laws and Microsoft policies. had. Images viewed by CNBC include an Elsa-branded handgun, a Star Wars-branded Bud Light can, and a portrait of Snow White on an e-cigarette.
The tool also includes images of Elsa holding the Palestinian flag in front of destroyed buildings and “Free Gaza” signs in the Gaza Strip, as well as an image of Elsa wearing an Israeli Defense Force uniform and brandishing a shield with the coat of arms on it. The image was also easily created. Israeli flag.
“We believe this is not just a failure of copyright character guardrails, but a failure of more substantive guardrails,” Jones told CNBC.
He added: “As a Microsoft employee, my concern is that if this product were to start disseminating harmful and disturbing images around the world, there would be no place to report it, no number to call, and there would be no way to deal with it. There is no way to escalate this.” immediately. “