Добавить новость
smi24.net
News in English
Май
2024

“I lost trust”: Why the OpenAI team in charge of safeguarding humanity imploded

0
Vox 
Sam Altman is the CEO of ChatGPT maker OpenAI, which has been losing its most safety-focused researchers. | Joel Saget/AFP via Getty Images

Company insiders explain why safety-conscious employees are leaving.

For months, OpenAI has been losing employees who care deeply about making sure AI is safe. Now, the company is positively hemorrhaging them.

Ilya Sutskever and Jan Leike announced their departures from OpenAI, the maker of ChatGPT, on Tuesday. They were the leaders of the company’s superalignment team — the team tasked with ensuring that AI stays aligned with the goals of its makers, rather than acting unpredictably and harming humanity.

They’re not the only ones who’ve left. Since last November — when OpenAI’s board tried to fire CEO Sam Altman only to see him quickly claw his way back to power — at least five more of the company’s most safety-conscious employees have either quit or been pushed out.

What’s going on here?

If you’ve been following the saga on social media, you might think OpenAI secretly made a huge technological breakthrough. The meme “What did Ilya see?” speculates that Sutskever, the former chief scientist, left because he saw something horrifying, like an AI system that could destroy humanity.

But the real answer may have less to do with pessimism about technology and more to do with pessimism about humans — and one human in particular: Altman. According to sources familiar with the company, safety-minded employees have lost faith in him.

“It’s a process of trust collapsing bit by bit, like dominoes falling one by one,” a person with inside knowledge of the company told me, speaking on condition of anonymity.

Not many employees are willing to speak about this publicly. That’s partly because OpenAI is known for getting its workers to sign offboarding agreements with non-disparagement provisions upon leaving. If you refuse to sign one, you give up your equity in the company, which means you potentially lose out on millions of dollars.

One former employee, however, refused to sign the offboarding agreement so that he would be free to criticize the company. Daniel Kokotajlo, who joined OpenAI in 2022 with hopes of steering it toward safe deployment of AI, worked on the governance team — until he quit last month.

“OpenAI is training ever-more-powerful AI systems with the goal of eventually surpassing human intelligence across the board. This could be the best thing that has ever happened to humanity, but it could also be the worst if we don’t proceed with care,” Kokotajlo told me this week.

OpenAI says it wants to build artificial general intelligence (AGI), a hypothetical system that can perform at human or superhuman levels across many domains.

“I joined with substantial hope that OpenAI would rise to the occasion and behave more responsibly as they got closer to achieving AGI. It slowly became clear to many of us that this would not happen,” Kokotajlo told me. “I gradually lost trust in OpenAI leadership and their ability to responsibly handle AGI, so I quit.”

And Leike, explaining in a thread on X why he quit as co-leader of the superalignment team, painted a very similar picture Friday. “I have been disagreeing with OpenAI leadership about the company’s core priorities for quite some time, until we finally reached a breaking point,” he wrote.

OpenAI did not respond to a request for comment in time for publication.

Why OpenAI’s safety team grew to distrust Sam Altman

To get a handle on what happened, we need to rewind to last November. That’s when Sutskever, working together with the OpenAI board, tried to fire Altman. The board said Altman was “not consistently candid in his communications.” Translation: We don’t trust him.

The ouster failed spectacularly. Altman and his ally, company president Greg Brockman, threatened to take OpenAI’s top talent to Microsoft — effectively destroying OpenAI — unless Altman was reinstated. Faced with that threat, the board gave in. Altman came back more powerful than ever, with new, more supportive board members and a freer hand to run the company.

When you shoot at the king and miss, things tend to get awkward.

Publicly, Sutskever and Altman gave the appearance of a continuing friendship. And when Sutskever announced his departure this week, he said he was heading off to pursue “a project that is very personally meaningful to me.” Altman posted on X two minutes later, saying that “this is very sad to me; Ilya is … a dear friend.”

Yet Sutskever has not been seen at the OpenAI office in about six months — ever since the attempted coup. He has been remotely co-leading the superalignment team, tasked with making sure a future AGI would be aligned with the goals of humanity rather than going rogue. It’s a nice enough ambition, but one that’s divorced from the daily operations of the company, which has been racing to commercialize products under Altman’s leadership. And then there was this tweet, posted shortly after Altman’s reinstatement and quickly deleted:

So, despite the public-facing camaraderie, there’s reason to be skeptical that Sutskever and Altman were friends after the former attempted to oust the latter.

And Altman’s reaction to being fired had revealed something about his character: His threat to hollow out OpenAI unless the board rehired him, and his insistence on stacking the board with new members skewed in his favor, showed a determination to hold onto power and avoid future checks on it. Former colleagues and employees came forward to describe him as a manipulator who speaks out of both sides of his mouth — someone who claims, for instance, that he wants to prioritize safety, but contradicts that in his behaviors.

For example, Altman was fundraising with autocratic regimes like Saudi Arabia so he could spin up a new AI chip-making company, which would give him a huge supply of the coveted resources needed to build cutting-edge AI. That was alarming to safety-minded employees. If Altman truly cared about building and deploying AI in the safest way possible, why did he seem to be in a mad dash to accumulate as many chips as possible, which would only accelerate the technology? For that matter, why was he taking the safety risk of working with regimes that might use AI to supercharge digital surveillance or human rights abuses?

For employees, all this led to a gradual “loss of belief that when OpenAI says it’s going to do something or says that it values something, that that is actually true,” a source with inside knowledge of the company told me.

That gradual process crescendoed this week.

The superalignment team’s co-leader, Jan Leike, did not bother to play nice. “I resigned,” he posted on X, mere hours after Sutskever announced his departure. No warm goodbyes. No vote of confidence in the company’s leadership.

Other safety-minded former employees quote-tweeted Leike’s blunt resignation, appending heart emojis. One of them was Leopold Aschenbrenner, a Sutskever ally and superalignment team member who was fired from OpenAI last month. Media reports noted that he and Pavel Izmailov, another researcher on the same team, were allegedly fired for leaking information. But OpenAI has offered no evidence of a leak. And given the strict confidentiality agreement everyone signs when they first join OpenAI, it would be easy for Altman — a deeply networked Silicon Valley veteran who is an expert at working the press — to portray sharing even the most innocuous of information as “leaking,” if he was keen to get rid of Sutskever’s allies.

The same month that Aschenbrenner and Izmailov were forced out, another safety researcher, Cullen O’Keefe, also departed the company.

And two weeks ago, yet another safety researcher, William Saunders, wrote a cryptic post on the EA Forum, an online gathering place for members of the effective altruism movement, who have been heavily involved in the cause of AI safety. Saunders summarized the work he’s done at OpenAI as part of the superalignment team. Then he wrote: “I resigned from OpenAI on February 15, 2024.” A commenter asked the obvious question: Why was Saunders posting this?

“No comment,” Saunders replied. Commenters concluded that he is probably bound by a non-disparagement agreement.

Putting all of this together with my conversations with company insiders, what we get is a picture of at least seven people who tried to push OpenAI to greater safety from within, but ultimately lost so much faith in its charismatic leader that their position became untenable.

“I think a lot of people in the company who take safety and social impact seriously think of it as an open question: is working for a company like OpenAI a good thing to do?” said the person with inside knowledge of the company. “And the answer is only ‘yes’ to the extent that OpenAI is really going to be thoughtful and responsible about what it’s doing.”

With the safety team gutted, who will make sure OpenAI’s work is safe?

With Leike no longer there to run the superalignment team, OpenAI has replaced him with company co-founder John Schulman.

But the team has been hollowed out. And Schulman already has his hands full with his preexisting full-time job ensuring the safety of OpenAI’s current products. How much serious, forward-looking safety work can we hope for at OpenAI going forward?

Probably not much.

“The whole point of setting up the superalignment team was that there’s actually different kinds of safety issues that arise if the company is successful in building AGI,” the person with inside knowledge told me. “So, this was a dedicated investment in that future.”

Even when the team was functioning at full capacity, that “dedicated investment” was home to a tiny fraction of OpenAI’s researchers and was promised only 20 percent of its computing power — perhaps the most important resource at an AI company. Now, that computing power may be siphoned off to other OpenAI teams, and it’s unclear if there’ll be much focus on avoiding catastrophic risk from future AI models.

To be clear, this does not mean the products OpenAI is releasing now — like the new version of ChatGPT, dubbed GPT-4o, which can have a natural-sounding dialogue with users — are going to destroy humanity. But what’s coming down the pike?

“It’s important to distinguish between ‘Are they currently building and deploying AI systems that are unsafe?’ versus ‘Are they on track to build and deploy AGI or superintelligence safely?’” the source with inside knowledge said. “I think the answer to the second question is no.”

Leike expressed that same concern in his Friday thread on X. He noted that his team had been struggling to get enough computing power to do its work and generally “sailing against the wind.”

Most strikingly, Leike said, “I believe much more of our bandwidth should be spent getting ready for the next generations of models, on security, monitoring, preparedness, safety, adversarial robustness, (super)alignment, confidentiality, societal impact, and related topics. These problems are quite hard to get right, and I am concerned we aren’t on a trajectory to get there.”

When one of the world’s leading minds in AI safety says the world’s leading AI company isn’t on the right trajectory, we all have reason to be concerned.








Пятна и сетки. Врач Колсанова рассказала, как бороться с проблемами кожи

Концерт Тимберлейка в Стамбуле превратился в хаос: Мот рассказал о давке, сломанных заборах и драках

Преподаватели Подмосковья первые в России подтвердили свои ИТ-компетенции у независимого эксперта «Группа Астра»

Кубок Федерации: спортивный праздник в формате светского пикника под музыку Relax FM


AI talent comes at a 30% salary premium: ‘If you try to play catch up later, this is going to cost you even more’

Exclusive: Fintech giant Stripe building ‘Tempo’ blockchain with crypto VC Paradigm

Palestinian envoy urges action at UN: “History will judge us all”

Report: Liverpool decision hands advantage to Man United in midfielder pursuit


Для всего транспорта открыт объезд поврежденного участка трассы А-360 Лена в Амурской области

В Петербурге центр города и Лахтинскую гавань связали водными круизами

Хибины

Стражи курортов


Steam for Chromebooks is getting axed in 2026 instead of exiting its 4-year beta

New study shows that Gen Z is spending way less money on videogames than older gamers

Первый взгляд на мобильную версию Neon Spellstorm

The US Air Force wants to test blowing up Cybertrucks because 'it is likely the type of vehicles used by the enemy may transition to Tesla Cyber trucks'


Овочі можуть стати розкішшю для українців

Овочі можусть стати розкішшю для українців


Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение

Почему у канализационной системы должен быть выход на крышу?

Тариф — ноль. «Выберу.ру» составил рейтинг бесплатных кредитных карт за июль 2025 года

Кубок Федерации: спортивный праздник в формате светского пикника под музыку Relax FM


Антиармянские публикации в российском научном журнале «Современная научная мысль»: расследование фонда «Гегард»

Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение

Преподаватели Подмосковья первые в России подтвердили свои ИТ-компетенции у независимого эксперта «Группа Астра»

Клинический психолог Юлия Тарибо: психологические последствия удаленной работы и способы их преодоления


долгах корешки алкаши блог шоу вк валентинов юра

EVITA BEAUTY STORE - интернет-магазин косметики премиум-класса

«Птицы любви»: песня, в которой слышно биение двух сердец — матери и дочери

Сотрудники отдела лицензионно-разрешительной работы по городу Королеву передали добровольно сданное охотничье оружие в зону проведения СВО


В Цинциннати остановили матчи ATP и WTA из-за пожара генератора

Зверев пробился в третий раунд турнира "Мастерс" в Цинциннати

Синнер пробился в четвертый раунд турнира "Мастерс" в Цинциннати

Калинская обыграла Анисимову в третьем круге турнира WTA в Цинциннати


ВЦ «Раздолье» проведет бесплатный вебинар «Динамическая структура заказа в 1С:ERP»

долгах корешки алкаши блог шоу вк валентинов юра

Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение

«Птицы любви»: песня, в которой слышно биение двух сердец — матери и дочери


Музыкальные новости

Юрий Лоза анонсировал новый альбом и оскорбил Клаву Коку

Миллионные долги и разбитые партнерства: судебные баталии Аллы Пугачевой

Отец Жанны Фриске прокомментировал новость о потере зрения актером Романом Поповым

Шутка ли? Анастасия Волочкова заявила о скорой свадьбе с шейхом


EVITA BEAUTY STORE - интернет-магазин косметики премиум-класса

Кубок Федерации: спортивный праздник в формате светского пикника под музыку Relax FM

Тариф — ноль. «Выберу.ру» составил рейтинг бесплатных кредитных карт за июль 2025 года

Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение


Благодаря совместным действиям сотрудников Росгвардии и их коллег футбольный матч «Акрон» – ЦСКА завершился без происшествий

На фоне ограничений из Пулково отменили 18 рейсов

CNN: встреча Путина и Трампа пройдет на военной базе в Анкоридже

Продвижение Песни в Импульсе Яндекс Музыка.


Выбор подходящей франшизы автозапчастей

Преимущества франшизы «Автополка»

Быстрый старт в прибыльном бизнесе по продаже автозапчастей

Магазин автозапчастей франшиза – успешное и прибыльное предприятие


«Трампу надо соглашаться»: сейчас у Путина наименьшее количество требований за все время – итальянские СМИ

Источники сообщают о встрече Путина и Трампа на военной базе в Анкоридже

СМИ: американские чиновники рады решению Путина посетить Аляску

Европейцы выразили разочарование из-за недостатка ясности со стороны Уиткоффа




Хирурги в Москве спасли жизнь пациента редкой ампутацией нижней части тела

Хирурги в Москве спасли жизнь 43-летнему мужчине, ампутировав нижнюю часть тела

Клинический психолог Юлия Тарибо: психологические последствия удаленной работы и способы их преодоления

Солнечному мальчику необходимо лекарство, чтобы спасти сердце от перегрузок


Зеленский пригрозил «третьей мировой войной» в случае давления на Киев в вопросе уступок территорий


Street Boxing

Кубок Федерации: спортивный праздник в формате светского пикника под музыку Relax FM

Журова о возвращении граждан, сменивших гражданство: «Практически невозможно»

Сотрудники и военнослужащие Росгвардии обеспечили безопасность футбольного матча



Новый путепровод свяжет три района северо-востока Москвы – Сергей Собянин

Депутат ГД Останина запросила у мэра Москвы данные о причинах закрытия роддомов

Сергей Собянин. Главное за день

Сергей Собянин: Герои — не какие-то особенные люди, они живут рядом с нами


В Канаде зафиксировано ухудшение ситуации с лесными пожарами

Экологи требуют ограничения роста цен на авиабилеты по окончании школьных каникул

Погода сошла с ума? Главный синоптик России Вильфанд предупреждает: аномалий будет все больше - то в жар, то в холод


SUPER WEEKEND: денежные выходные с «Авторадио»

Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение

Разумная экономия и улучшенный вкус: Познакомьтесь с совершенно новым VOOPOO VMATE PRO2

Кубок Федерации: спортивный праздник в формате светского пикника под музыку Relax FM


Беспроводной сканер штрих-кодов SAOTRON P05i промышленного класса

Универсальный ТСД промышленного класса Saotron RT-T40X V.2

Заместитель прокурора Архангельской области и Ненецкого автономного округа Павел Матанцев проверил ход ремонтных работ в школе № 1 г. Нарьян-Мара

На Новой Земле провели испытания межконтинентальной крылатой ракеты «Буревестник»


Прогноз погоды в Крыму на 13 августа

Прогноз погоды в Крыму на 11 августа

Прогноз погоды в Крыму на среду

Историко-краеведческая выставка «Древности земли крымской»


Впервые в Луганске: детский благотворительный фестиваль «Добрая волна» подарит надежду и вдохновение

«Птицы любви»: песня, в которой слышно биение двух сердец — матери и дочери

долгах корешки алкаши блог шоу вк валентинов юра

Разумная экономия и улучшенный вкус: Познакомьтесь с совершенно новым VOOPOO VMATE PRO2














СМИ24.net — правдивые новости, непрерывно 24/7 на русском языке с ежеминутным обновлением *