Tuesday, December 9, 2025

Scaled-up LLMs are more prone to sensible yet wrong answers

Making it less reliable

Researchers at the Universitat Politècnica de València, in Spain, have found that as Large Language Models (LLMs) grow larger and more sophisticated, they are more likely to span sensible yet wrong answers, making them less reliable. Researchers analyzed three popular LLMs; GPT by OpenAI, the LLaMA of Meta, and the BLOOM suite developed by BigScience, and noticed that with new versions accuracy increases, so do the hedging, refusal, or evasiveness. The study also found that LLMs rarely admit to a user that they do not know an answer.

- Advertisement -
Explore more ..

Humanoid Launches UK’s First Industrial Humanoid Robot, HMND 01 Alpha

Built in record time, Alpha targets labour shortages with warehouse-ready automation.

New Calculus Method Boosts Robot Agility

Yale researchers develop a faster way for robots to compute derivatives, unlocking smoother, more proactive movements.

Volkswagen Powers Ahead with Unified Cells and Solid-State Batteries

Volkswagen's battery innovations promise longer range and faster charging for future EVs.

Honeywell Unveils Ionic™: Modular Energy Storage for Industry

A compact all-in-one BESS to cut costs, boost reliability, and integrate renewables.
- Advertisement -