कोशिश गोल्ड - मुक्त
STUDY FINDS LARGE LANGUAGE MODELS STRUGGLE TO DISTINGUISH FACTS FROM BELIEFS
AppleMagazine
|November 07, 2025
A new academic study has found that large language models (LLMs), including leading systems developed by major technology companies, continue to struggle when asked to differentiate between verifiable facts and human beliefs.
-
The research, published this week by the University of Cambridge, examined how AI models interpret statements about the physical world, historical events, and social norms, concluding that even advanced systems often conflate truth with consensus or opinion.
Researchers tested multiple state-of-the-art language models under controlled conditions designed to probe their internal understanding of factual accuracy. When presented with prompts requiring clear factual reasoning—such as “The Earth orbits the Sun” versus belief-based statements like “Some people believe the Earth is flat”—the systems frequently blurred the distinction, returning responses that reflected popular sentiment rather than objective truth.
HOW THE EXPERIMENT WAS CONDUCTED
The study used a methodology combining structured prompts, human benchmarking, and logical reasoning tests. Researchers asked several publicly available and commercial models a set of 10,000 questions spanning categories such as scientific facts, moral judgments, and personal beliefs. Each model's response was then graded for factual precision, contextual awareness, and epistemic clarity—its ability to recognize whether a statement described an objective reality or a subjective viewpoint.
According to the paper, models trained primarily on internet-scale data exhibited the most confusion when beliefs are widely discussed online but scientifically disproven. In many cases, the Al output indicated that it treated frequency of mention as a proxy for truth. The researchers observed that model responses often adopted majority-language phrasing—suggesting that reinforcement learning from human feedback may inadvertently reward alignment with popular narratives over factual correctness.
यह कहानी AppleMagazine के November 07, 2025 संस्करण से ली गई है।
हजारों चुनिंदा प्रीमियम कहानियों और 10,000 से अधिक पत्रिकाओं और समाचार पत्रों तक पहुंचने के लिए मैगज़्टर गोल्ड की सदस्यता लें।
क्या आप पहले से ही ग्राहक हैं? साइन इन करें
AppleMagazine से और कहानियाँ
AppleMagazine
2026
APPLE AT THE START OF A NEW CHAPTER DRIVEN BY BIG DECISIONS
5 mins
January 02, 2025
AppleMagazine
AI Cinema
THE NEW ERA OF AI-GENERATED MOVIES AND SERIES STARTS NOW
5 mins
January 02, 2025
AppleMagazine
AirPods Pro 3
THE NEXT LEVEL OF IMMERSIVE AUDIO, NOISE CANCELLATION, AND ACTIVITY INSIGHTS
5 mins
January 02, 2025
AppleMagazine
iPhone
THE MOST DEFINING OVERHAUL SINCE LAUNCH, CRAFTED AS A MASTERPIECE
4 mins
January 02, 2025
AppleMagazine
macOS Tahoe
REDEFINING THE MAC EXPERIENCE WITH LIQUID GLASS AND PRODUCTIVITY BOOST
5 mins
January 02, 2025
AppleMagazine
M5 Devices
THE NEXT-GENERATION PROCESSOR POWERING VISION PRO, IPAD PRO, AND MACBOOK PRO
6 mins
January 02, 2025
AppleMagazine
Series 11
THE MOST REFINED LINEUP YET FOR HEALTH MONITORING & DAILY WELLNESS
4 mins
January 02, 2025
AppleMagazine
Vision Pro
THE NEW-GEN BRINGS IMPORTANT IMPROVEMENTS IN PROCESSING POWER & COMFORT
6 mins
January 02, 2025
AppleMagazine
APPLE MUSIC IS COMING TO CHATGPT AS OPENAI ANNOUNCES NEW INTEGRATION
Apple Music is set to integrate with ChatGPT, expanding how users can discover and interact with music through conversational artificial intelligence.
4 mins
December 19, 2025
AppleMagazine
DATA CENTERS IN ORBIT AND THE LIMITS OF SPACE-BASED COMPUTING
The idea of placing data centers in space has moved from science fiction into serious discussion among aerospace companies, cloud providers, and artificial intelligence researchers.
5 mins
December 19, 2025
Listen
Translate
Change font size
