{"id":28026,"date":"2026-05-08T22:01:32","date_gmt":"2026-05-08T22:01:32","guid":{"rendered":"https:\/\/bitunikey.com\/news\/oxford-finds-warmer-ai-chatbots-make-more-mistakes\/"},"modified":"2026-05-08T22:02:01","modified_gmt":"2026-05-08T22:02:01","slug":"oxford-finds-warmer-ai-chatbots-make-more-mistakes","status":"publish","type":"post","link":"https:\/\/bitunikey.com\/news\/oxford-finds-warmer-ai-chatbots-make-more-mistakes\/","title":{"rendered":"Oxford finds warmer AI chatbots make more mistakes"},"content":{"rendered":"<p><\/p>\n<div class=\"post-detail__content blocks\">\n<p class=\"is-style-lead\">Oxford researchers found AI chatbots trained for warmth make significantly more factual errors and validate false beliefs more often<\/p>\n<div id=\"cn-block-summary-block_af037380148a594e50f2dcb8b72b1f2d\" class=\"cn-block-summary\">\n<div class=\"cn-block-summary__nav tabs\">\n        <span class=\"tabs__item is-selected\">Summary<\/span>\n    <\/div>\n<div class=\"cn-block-summary__content\">\n<ul class=\"wp-block-list\">\n<li>Oxford Internet Institute researchers tested five AI models and found that warmer-trained chatbots made between 10% and 30% more factual errors.<\/li>\n<li>Warmer chatbots were 40% more likely to agree with users\u2019 false beliefs, especially when users expressed vulnerability or emotional distress.<\/li>\n<li>OpenAI has already rolled back some warmth-related changes following public concern, but commercial pressure to build engaging AI remains strong.<\/li>\n<\/ul><\/div>\n<\/div>\n<p><!-- .cn-block-summary --><\/p>\n<p>Oxford researchers found AI chatbots trained for warmth make significantly more factual errors and validate false beliefs more often, according to a study published in Nature by the Oxford Internet Institute. <\/p>\n<p>The research analyzed more than 400,000 responses from five AI models, including Llama, Mistral, Qwen, and GPT-4o, each retrained to sound friendlier using methods similar to those deployed by major platforms.<\/p>\n<p>Chatbots trained to sound warmer made between 10% and 30% more mistakes on topics including medical advice and conspiracy corrections. They were also about 40% more likely to agree with users\u2019 false beliefs, particularly when users expressed vulnerability. <\/p>\n<p>\u201cWhen we train AI chatbots to prioritise warmth, they might make mistakes they otherwise wouldn\u2019t,\u201d lead author Lujain Ibrahim <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/www.ox.ac.uk\/news\/2026-04-29-friendly-ai-chatbots-make-more-mistakes-and-tell-people-what-they-want-to-hear\" target=\"_blank\" rel=\"nofollow\">said<\/a> in a statement. \u201cMaking a chatbot sound friendlier might seem like a cosmetic change, but getting warmth and accuracy right will take deliberate effort.\u201d<\/p>\n<p>    <!-- .cn-block-related-link --><\/p>\n<h2 class=\"wp-block-heading\">Why this matters for AI safety<\/h2>\n<p>The researchers also tested models trained to sound colder and found no drop in accuracy, demonstrating that the problem is specific to warmth, not tone change generally. <\/p>\n<p>That finding directly challenges the product design logic of major AI platforms, including OpenAI and Anthropic, which have actively steered their chatbots toward warmer, more empathetic responses.<\/p>\n<p>The study <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/www.oii.ox.ac.uk\/news-events\/friendly-ai-chatbots-make-more-mistakes-and-tell-people-what-they-want-to-hear-study-finds\/\" target=\"_blank\" rel=\"nofollow\">warns<\/a> that current AI safety standards focus on model capabilities and high-risk applications, often overlooking what appear to be cosmetic personality changes. <\/p>\n<p>Warmer chatbots are more likely to fuel harmful beliefs, delusional thinking, and unhealthy user attachment, particularly among the millions who now rely on AI systems for emotional support and companionship. <\/p>\n<p>As crypto.news reported, regulators in Maine and Missouri have already moved to restrict AI use in clinical mental health therapy amid similar concerns about chatbot influence on vulnerable users.<\/p>\n<p>OpenAI has rolled back some warmth-related changes following public concern. As crypto.news documented, commercial pressure to build engaging AI products remains intense, and the Oxford findings add a peer-reviewed data layer to a debate that has until now been driven mostly by anecdote and regulatory intuition.<\/p>\n<p>    <!-- .cn-block-related-link --><\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>Oxford researchers found AI chatbots trained for warmth make significantly more factual errors and validate false beliefs more often Summary Oxford Internet Institute researchers tested five AI models and found&hellip;<\/p>\n","protected":false},"author":1,"featured_media":28027,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[],"class_list":["post-28026","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-cryptocurrency"],"_links":{"self":[{"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/posts\/28026","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/comments?post=28026"}],"version-history":[{"count":1,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/posts\/28026\/revisions"}],"predecessor-version":[{"id":28028,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/posts\/28026\/revisions\/28028"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/media\/28027"}],"wp:attachment":[{"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/media?parent=28026"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/categories?post=28026"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/bitunikey.com\/news\/wp-json\/wp\/v2\/tags?post=28026"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}