Researchers at the Universitat Politècnica de València, in Spain, have found that as Large Language Models (LLMs) grow larger and more sophisticated, they are more likely to span sensible yet wrong answers, making them less reliable. Researchers analyzed three popular LLMs; GPT by OpenAI, the LLaMA of Meta, and the BLOOM suite developed by BigScience, and noticed that with new versions accuracy increases, so do the hedging, refusal, or evasiveness. The study also found that LLMs rarely admit to a user that they do not know an answer.
Scaled-up LLMs are more prone to sensible yet wrong answers
Making it less reliable
- Advertisement -
Explore more ..
Humanoid Launches UK’s First Industrial Humanoid Robot, HMND 01 Alpha
Built in record time, Alpha targets labour shortages with warehouse-ready automation.
New Calculus Method Boosts Robot Agility
Yale researchers develop a faster way for robots to compute derivatives, unlocking smoother, more proactive movements.
Volkswagen Powers Ahead with Unified Cells and Solid-State Batteries
Volkswagen's battery innovations promise longer range and faster charging for future EVs.
Honeywell Unveils Ionic™: Modular Energy Storage for Industry
A compact all-in-one BESS to cut costs, boost reliability, and integrate renewables.
TRENDING
- Advertisement -



