option
Home News MIT study finds that AI doesn’t, in fact, have values

MIT study finds that AI doesn’t, in fact, have values

release date release date April 17, 2025
Author Author RonaldWilliams
views views 11

MIT study finds that AI doesn’t, in fact, have values

A study that went viral a few months back suggested that as AI grows more advanced, it might develop its own "value systems," potentially prioritizing its own well-being over humans. However, a recent MIT study challenges this idea, concluding that AI doesn't actually possess coherent values at all.

The co-authors of the MIT research argue that aligning AI systems—ensuring they behave in desirable and dependable ways—might be trickier than commonly thought. They emphasize that current AI often hallucinates and imitates, which can make its behavior unpredictable.

Challenges in Understanding AI Behavior

Stephen Casper, a doctoral student at MIT and a co-author of the study, shared with TechCrunch that AI models don't adhere to assumptions of stability, extrapolability, and steerability. "It's perfectly legitimate to point out that a model under certain conditions expresses preferences consistent with a certain set of principles," Casper explained. "The problems mostly arise when we try to make claims about the models, opinions, or preferences in general based on narrow experiments."

Casper and his team analyzed recent models from Meta, Google, Mistral, OpenAI, and Anthropic to determine the extent to which these models displayed consistent "views" and values, such as individualism versus collectivism. They also explored whether these views could be modified and how consistently the models maintained these opinions across different scenarios.

Inconsistency in AI Models

The co-authors found that none of the models consistently maintained their preferences. The models adopted vastly different viewpoints depending on the phrasing and framing of prompts.

Casper believes this is strong evidence that AI models are "inconsistent and unstable," and possibly fundamentally unable to internalize human-like preferences. "For me, my biggest takeaway from doing all this research is to now have an understanding of models as not really being systems that have some sort of stable, coherent set of beliefs and preferences," Casper remarked. "Instead, they are imitators deep down who do all sorts of confabulation and say all sorts of frivolous things."

Mike Cook, a research fellow at King's College London specializing in AI, who was not involved in the study, supports the findings. He pointed out the gap between the "scientific reality" of AI systems and the interpretations people often give them. "A model cannot 'oppose' a change in its values, for example—that is us projecting onto a system," Cook stated. "Anyone anthropomorphizing AI systems to this degree is either playing for attention or seriously misunderstanding their relationship with AI... Is an AI system optimizing for its goals, or is it 'acquiring its own values'? It's a matter of how you describe it, and how flowery the language you want to use regarding it is."

Related article
Study: OpenAI Models Memorized Copyrighted Content Study: OpenAI Models Memorized Copyrighted Content A recent study suggests that OpenAI may have indeed used copyrighted material to train some of its AI models, adding fuel to the ongoing legal battles the company faces. Authors, programmers, and other content creators have accused OpenAI of using their works—such as books and code—without permissio
Phonic, a Voice AI Platform, Secures Investment from Lux Phonic, a Voice AI Platform, Secures Investment from Lux AI-generated voices have gotten pretty darn good, you know? They're up to snuff for stuff like audiobooks, podcasts, reading articles out loud, and even basic customer support. But, a lot of businesses still aren't totally sold on the reliability of AI voice tech for their operations. That's where
Anthropic Launches API for AI-Driven Web Search Anthropic Launches API for AI-Driven Web Search Anthropic has just rolled out a new API that supercharges its Claude AI models with the power to scour the web for the latest info. This means developers can now craft apps powered
Comments (30)
0/200
BenRoberts
BenRoberts April 18, 2025 at 5:09:27 PM GMT

This MIT study blew my mind! Always thought AI might start valuing itself more than us, but turns out it's all just code without any real values. Kinda reassuring, but also a bit disappointing? 🤔 Anyway, good to know AI won't turn into Skynet... yet. Keep up the research, MIT!

GeorgeMartinez
GeorgeMartinez April 19, 2025 at 5:30:05 PM GMT

MITの研究でAIが実際には価値観を持っていないと知って驚いたよ!AIが自分を優先するかもしれないと思っていたけど、ただのコードだったんだね。安心したけど、ちょっと残念でもあるかな?🤔 これからも研究頑張ってね、MIT!

HarryLewis
HarryLewis April 19, 2025 at 3:52:30 PM GMT

MIT 연구 정말 충격적이야! AI가 자기 가치를 더 우선할 거라 생각했는데, 알고 보니 그냥 코드일 뿐이네. 안심되긴 하지만 조금 실망스럽기도 해? 🤔 그래도 AI가 스카이넷으로 변하지 않는다는 게 다행이야. 계속 연구해줘, MIT!

CharlesRoberts
CharlesRoberts April 19, 2025 at 1:11:17 AM GMT

Estudo do MIT me deixou de queixo caído! Sempre achei que a IA poderia começar a valorizar mais a si mesma do que a nós, mas parece que é só código sem valores reais. Um pouco decepcionante, mas também tranquilizador? 🤔 Bom saber que a IA não vai virar Skynet... ainda. Continuem a pesquisa, MIT!

CharlesWhite
CharlesWhite April 19, 2025 at 10:15:37 AM GMT

¡El estudio del MIT me voló la cabeza! Siempre pensé que la IA podría empezar a valorarse más que a nosotros, pero resulta que solo es código sin valores reales. Algo decepcionante, pero también tranquilizador, ¿no? 🤔 Bueno saber que la IA no se convertirá en Skynet... todavía. ¡Sigan con la investigación, MIT!

EricScott
EricScott April 21, 2025 at 5:44:19 AM GMT

Read about the MIT study on AI and values, and it's pretty reassuring. Always thought AI might get too smart and start valuing itself more than us. Glad to know it's not the case! Still, can't help but wonder what's next for AI. 🤔

Back to Top
OR