OpenAI Forms Safety Committee as it Starts Training Latest Artificial Intelligence Model

FILE - The OpenAI logo is seen displayed on a cell phone with an image on a computer monitor generated by ChatGPT’s Dall-E text-to-image model, Friday, Dec. 8, 2023, in Boston. OpenAI says it’s setting up a new safety and security committee and has begun training a new artificial intelligence model to supplant the GPT-4 system that underpins its ChatGPT chatbot. The San Francisco startup said in a blog post Tuesday May 28, 2024 that the committee will advise the full board on “critical safety and security decisions” for its projects and operations. (AP Photo/Michael Dwyer, File)
FILE - The OpenAI logo is seen displayed on a cell phone with an image on a computer monitor generated by ChatGPT’s Dall-E text-to-image model, Friday, Dec. 8, 2023, in Boston. OpenAI says it’s setting up a new safety and security committee and has begun training a new artificial intelligence model to supplant the GPT-4 system that underpins its ChatGPT chatbot. The San Francisco startup said in a blog post Tuesday May 28, 2024 that the committee will advise the full board on “critical safety and security decisions” for its projects and operations. (AP Photo/Michael Dwyer, File)
TT

OpenAI Forms Safety Committee as it Starts Training Latest Artificial Intelligence Model

FILE - The OpenAI logo is seen displayed on a cell phone with an image on a computer monitor generated by ChatGPT’s Dall-E text-to-image model, Friday, Dec. 8, 2023, in Boston. OpenAI says it’s setting up a new safety and security committee and has begun training a new artificial intelligence model to supplant the GPT-4 system that underpins its ChatGPT chatbot. The San Francisco startup said in a blog post Tuesday May 28, 2024 that the committee will advise the full board on “critical safety and security decisions” for its projects and operations. (AP Photo/Michael Dwyer, File)
FILE - The OpenAI logo is seen displayed on a cell phone with an image on a computer monitor generated by ChatGPT’s Dall-E text-to-image model, Friday, Dec. 8, 2023, in Boston. OpenAI says it’s setting up a new safety and security committee and has begun training a new artificial intelligence model to supplant the GPT-4 system that underpins its ChatGPT chatbot. The San Francisco startup said in a blog post Tuesday May 28, 2024 that the committee will advise the full board on “critical safety and security decisions” for its projects and operations. (AP Photo/Michael Dwyer, File)

OpenAI says it's setting up a safety and security committee and has begun training a new AI model to supplant the GPT-4 system that underpins its ChatGPT chatbot.

The San Francisco startup said in a blog post Tuesday that the committee will advise the full board on “critical safety and security decisions" for its projects and operations.

The safety committee arrives as debate swirls around AI safety at the company, which was thrust into the spotlight after a researcher, Jan Leike, resigned and leveled criticism at OpenAI for letting safety “take a backseat to shiny products." OpenAI co-founder and chief scientist Ilya Sutskever also resigned, and the company disbanded the “superalignment” team focused on AI risks that they jointly led, The AP reported.

OpenAI said it has “recently begun training its next frontier model” and its AI models lead the industry on capability and safety, though it made no mention of the controversy. “We welcome a robust debate at this important moment,” the company said.

AI models are prediction systems that are trained on vast datasets to generate on-demand text, images, video and human-like conversation. Frontier models are the most powerful, cutting edge AI systems.

The safety committee is filled with company insiders, including OpenAI CEO Sam Altman and Chairman Bret Taylor, and four OpenAI technical and policy experts. It also includes board members Adam D’Angelo, who’s the CEO of Quora, and Nicole Seligman, a former Sony general counsel.

The committee's first job will be to evaluate and further develop OpenAI’s processes and safeguards and make its recommendations to the board in 90 days. The company said it will then publicly release the recommendations it's adopting “in a manner that is consistent with safety and security.”



AI Chatbots Must Learn to Say 'Help!' Says Microsoft Exec

A Microsoft logo is seen in Los Angeles, California US November 7, 2017. (Reuters)
A Microsoft logo is seen in Los Angeles, California US November 7, 2017. (Reuters)
TT

AI Chatbots Must Learn to Say 'Help!' Says Microsoft Exec

A Microsoft logo is seen in Los Angeles, California US November 7, 2017. (Reuters)
A Microsoft logo is seen in Los Angeles, California US November 7, 2017. (Reuters)

Generative AI tools will save companies lots of time and money, promises Vik Singh, a Microsoft vice president, even if the models must learn to admit when they just don't know what to do.
"Just to be really frank, the thing that's really missing today is that a model doesn't raise its hands and say 'Hey, I'm not sure, I need help,'" Singh told AFP in an interview.
Since last year, Microsoft, Google and their competitors have been rapidly deploying generative AI applications like ChatGPT, which produce all kinds of content on demand and give users the illusion of omniscience.
But despite progress, they still "hallucinate," or invent answers.
This is an important problem for the Copilot executive to solve: Singh's corporate customers can't afford for their AI systems to go off the rails, even occasionally.
Marc Benioff, CEO of Salesforce, this week said he saw many of his customers increasingly frustrated with the meanderings of Microsoft's Copilot.
Singh insisted that "really smart people" were trying to find ways for a chatbot to admit "when it doesn't know the right answer and to ask for help."
'Real savings'
A more humble model would be no less useful, in Singh's opinion. Even if the model has to turn to a human in 50 percent of cases, that still saves "tons of money."
At one Microsoft client, "every time a new request comes in, they spend $8 to have a customer service rep answer it, so there are real savings to be had, and it's also a better experience for the customer because they get a faster response."
Singh arrived at Microsoft in January and this summer took over as head of the teams developing "Copilot," Microsoft's AI assistant that specializes in sales, accounting and online services.
These applications have the gargantuan task of bringing in revenue and justifying the massive investments in generative AI.
At the height of the AI frenzy, start-ups driving the technology were promising systems so advanced that they would "uplift humanity," in the words of Sam Altman, head of OpenAI, which is mainly funded by Microsoft.
But for the time being, the new technology is mainly used to boost productivity, and hopefully profits.
According to Microsoft, Copilot can do research for salespeople, freeing up time to call customers. Lumen, a telecom company, "saves around $50 million a year" doing this, said Singh.
Singh's teams are working on integrating Copilot directly into the tech giant's software and making it more autonomous.
"Let's say I'm a sales rep and I have a customer call," suggested the executive. Two weeks later, the model can "nudge the rep to go follow up, or better, just go and automatically send the email on the rep's behalf because it's been approved to do so."
'First inning'
In other words, before finding a solution to global warming, AI is expected to rid humanity of boring, repetitive chores.
"We're in the first inning," Singh said. "A lot of these things are productivity based, but they obviously have huge benefits."
Will all these productivity gains translate into job losses?
Leaders of large firms, such as K Krithivasan, boss of Indian IT giant TCS, have declared that generative AI will all but wipe out call centers.
But Singh, like many Silicon Valley executives, is counting on technology to make humans more creative and even create new jobs.
He pointed to his experience at Yahoo in 2008, when a dozen editors chose the articles for the home page.
"We came up with the idea of using AI to optimize this process, and some people asked 'Oh my God, what's going to happen to the employees?'" said Singh.
The automated system made it possible to renew content more quickly, thereby increasing the number of clicks on links but also the need for new articles.
"In the end," said the executive, "we had to recruit more editors."