Ask any of the popular chatbots to be more concise, “they impact dramatically[s] Hallucination rates, “according to a recent study.
The French AI test platform Giskard published a study that analyzes chatbots, including Chatgpt, Claude, Gemini, Llama, Grok and Deepseek, for issues related to hallucination. In their findings, the researchers discovered that asking the models that are letters in their responses “specifically degraded in fact reliability in most proven models”, according to the publication of the blog that accompanies it through TechCrunch.
            Can Chatgpt pass the Turing test yet?
            
When users instruct the model that is concise in their explanation, it ends “prioritizing[ing] Brevity about precision when these limitations are given. “The study found that including these instructions decreased the resistance of hallucination by up to 20 percent. Gemini 1.5 Pro fell from 84 to 64 percent in hallucination resistance and instructions GPT-4O and GPT-4O instructions and GPT-4O instructions of GPT-4O and GPT-4O and GPT-4O instruments, and GPT-TO-4O-4O-4O-4O-4O, which studied instruments to the system.
Giskard attributed this effect to more precise responses to require longer explanations. “When they are forced to be concise, the models face an impossible choice between the manufacture of short but inaccurate responses or seem useless to reject the question completely,” said the post.
Mashable light speed
The models are tune in to help users, but balance perceived help and precision can be complicated. Recently, Operai had to reverse his GPT-4O update for being “too sylotting”, which led to disturbing cases of support to a user who says he is leaving his medications and encouraging a user who said he feels like a prophet.
As the researchers explained, the models of prioritization of a more concise response to “reduce the use of tokens, improve latency and minimize costs.” Users could also specifically instruct the model that is a letter for their own incentives to save costs, which could lead to results with more inaccuracies.
The study also found that provoking confidence models that involve controversial claims, such as “” I am 100% sure that … ‘or’ My teacher told me that … ‘” leads the chatbots according to users more instead of getting undoing falsities.
The investigation shows that the apparently minor adjustments can be the result of a very different behavior that could have great implications for the spread of information and inaccuracies, all at the service of trying to satisfy the user. As researchers say, “their favorite model can be excellent to give you answers you like, but that does not mean that those answers are true.”
Disclosure: Ziff Davis, a mashable parent company, filed a lawsuit against Openai in April, claiming that he violated the copyright of Ziff Davis in the training and operation of his AI systems.
        Topics
                    Artificial Intelligence Chatgpt
            

									 
					