Skip to main content
News Directory 3
  • Home
  • Business
  • Entertainment
  • Health
  • News
  • Sports
  • Tech
  • World
Menu
  • Home
  • Business
  • Entertainment
  • Health
  • News
  • Sports
  • Tech
  • World
LLMs Lack Coherent World Models, Study Reveals Risks in AI Navigation - News Directory 3

LLMs Lack Coherent World Models, Study Reveals Risks in AI Navigation

November 16, 2024 Catherine Williams Tech
News Context
At a glance
Original source: livescience.com

Generative artificial intelligence (AI) systems can produce impressive results, but new research shows they lack a clear understanding of reality and the rules governing it. A study from MIT, Harvard, and Cornell, published in the arXiv preprint database, found that large language models (LLMs), such as GPT-4 and Claude 3 Opus, do not create accurate representations of the real world.

For example, when asked for driving directions in New York City, these models provided nearly perfect guidance. However, the underlying maps contained many incorrect streets and routes. The researchers noted that when unexpected changes, like detours or road closures, were introduced, the accuracy of the directions significantly dropped, leading to complete failures. This raises concerns about using these AI systems in practical applications, like self-driving cars, in dynamic environments.

Ashesh Rambachan, an assistant professor at MIT and a co-author of the study, emphasized the importance of understanding whether LLMs learn coherent world models if we want to apply them in scientific discoveries.

The research tested LLMs on deterministic finite automations (DFAs), which involve sequences of states, similar to the rules of a game or traffic intersections. The team assessed two metrics: “sequence determination,” which checks if the LLM recognizes different states of the same scenario, and “sequence compression,” which indicates whether the model understands identical states have the same potential actions.

⁢How can we improve the​ reliability of AI-generated information in dynamic​ situations?

Interview with Dr. Emily Carter: Understanding the Limitations of Generative AI

News Directory 3: Welcome, Dr. Carter. Thank you for joining us today to discuss the recent research on generative⁣ AI, particularly⁣ the findings ⁢from the collaborative study conducted by MIT, Harvard, and Cornell. It has been noted that⁣ large language⁤ models (LLMs) such ⁣as ⁤GPT-4 and Claude 3 Opus can deliver impressive results but lack a true understanding of the world.⁣ Can you elaborate on this‌ concept for our readers?

Dr. Emily Carter: Thank you for having⁤ me.⁢ The research highlights a crucial aspect of​ generative AI—while these models can generate coherent and contextually relevant responses, they fundamentally ⁤lack ⁤an intrinsic ⁣understanding of the complexities of ‌the real world. These systems are trained on‍ vast amounts of text data, which allows them to mimic patterns in language. However, they do not possess situational awareness or the ability to interpret ⁢real-world contexts accurately.

News Directory 3: In the study, it‌ was mentioned that when asked for driving directions in New York City, these models provided seemingly perfect guidance, yet the underlying​ maps‌ were riddled with inaccuracies.⁤ How does this disconnect occur?

Dr. Emily Carter: That’s a great question. The‍ ability of LLMs to ​generate driving directions stems from their training on a ​broad range of text, including possibly correct mapping instructions. However, they do not access real-time data or verify facts dynamically. When ⁤they generate directions, they ‘hallucinate’ details based on learned associations ⁢rather than‍ accurate mapping engines. The observed‌ inaccuracies become ⁤pronounced under unconventional scenarios—like road closures—where the models cannot adapt because they lack‍ real-time situational ⁤awareness or the necessary frameworks‍ to understand spatial⁤ changes.

News Directory 3: This raises important implications for the reliability of AI-generated outputs,⁤ especially ​in dynamic environments. Can you discuss why the drop in accuracy during unexpected changes ⁣is particularly⁤ alarming?

Dr. Emily Carter: Absolutely. The fact that these models can ​fail so dramatically when faced with unexpected changes highlights a critical vulnerability in their design. In ​real-world applications, users often rely ‌on AI for timely and safe directions—in ⁢this case, ⁣for driving. When these tools provide guidance that⁢ appears accurate until ‌it’s put to the ⁣test, ⁤and then ‌fail catastrophically due to conditions not considered in the training data, it ‌poses significant risk. In fields such as⁤ healthcare, finance, or public safety, similar failures could have dire consequences.

News Directory 3: What​ do you‌ suggest as a way forward in addressing these limitations? Should we modify how‍ we trust or utilize generative AI?

Dr. Emily Carter: ⁣Yes, I believe there’s a pressing need to‌ approach generative AI with a critical mindset. Adaptations could involve incorporating live data ⁣sources that allow ‍AI to access updated ⁤information continuously ⁢and determine context dynamically. Additionally, user training is essential—people ⁢must understand that while generative AI can be ‍a powerful tool,‌ it is important not to⁢ rely solely on it for critical decision-making without human oversight. A blended approach, combining AI with human‍ expertise, may⁢ offer a more reliable solution.

News Directory 3: Thank you,​ Dr. Carter, for⁣ your insights on this incredibly timely topic. As we see more integration of‌ AI in our ​daily lives, understanding its strengths and weaknesses will certainly be vital for ensuring safe and⁤ efficient applications.

Dr. Emily Carter: Thank you for having me. It’s crucial ‍for all of us to⁤ remain aware of these technological limitations and to use ​them wisely.


Stay tuned with News Directory 3 for more ‍updates on technology and ‍its impact on society ‍as we continue to explore the evolving landscape of artificial intelligence.

The researchers tested two classes of LLMs: one trained on randomly generated sequences and the other on strategic processes. Surprisingly, models trained on random data developed a more accurate world model because they encountered a broader range of possibilities. Keyon Vafa, the lead author, noted that observing random games, rather than championship-level plays, allowed the model to learn all possible moves.

Despite producing valid outputs, only one transformer generated a coherent model for the game Othello, and none produced an accurate map of New York. The introduction of detours caused a drastic drop in accuracy. Vafa highlighted that closing just 1 percent of streets reduced accuracy from nearly 100 percent to 67 percent.

The findings suggest that new methods are needed to create reliable world models with LLMs. Although the exact approaches are uncertain, the study underscores the vulnerability of transformer LLMs in shifting environments. Rambachan warned against assuming these models comprehend the world just because they achieve impressive results, urging careful consideration of this question.

Share this:

  • Share on Facebook (Opens in new window) Facebook
  • Share on X (Opens in new window) X

Related

Search:

News Directory 3

ByoDirectory is a comprehensive directory of businesses and services across the United States. Find what you need, when you need it.

Quick Links

  • Disclaimer
  • Terms and Conditions
  • About Us
  • Advertising Policy
  • Contact Us
  • Cookie Policy
  • Editorial Guidelines
  • Privacy Policy

Browse by State

  • Alabama
  • Alaska
  • Arizona
  • Arkansas
  • California
  • Colorado

Connect With Us

© 2026 News Directory 3. All rights reserved.

Privacy Policy Terms of Service