--- title: "Elon Musk’s Grok records lowest hallucination rate in AI reliability study" type: "News" locale: "en" url: "https://longbridge.com/en/news/270721034.md" description: "Elon Musk's Grok has been identified as one of the most reliable AI chatbots in a December 2025 study by Relum, achieving the lowest hallucination rate of 8% among 10 major models. Despite its lower market visibility, Grok's factual accuracy makes it a strong choice for accuracy-critical applications. In contrast, ChatGPT and Google's Gemini recorded high hallucination rates of 35% and 38%, respectively. The study highlights the importance of choosing reliable AI tools for workplace use." datetime: "2025-12-24T11:09:56.000Z" locales: - [zh-CN](https://longbridge.com/zh-CN/news/270721034.md) - [en](https://longbridge.com/en/news/270721034.md) - [zh-HK](https://longbridge.com/zh-HK/news/270721034.md) --- > Supported Languages: [简体中文](https://longbridge.com/zh-CN/news/270721034.md) | [繁體中文](https://longbridge.com/zh-HK/news/270721034.md) # Elon Musk’s Grok records lowest hallucination rate in AI reliability study A December 2025 study by casino games aggregator Relum has identified Elon Musk’s Grok as one of the most reliable AI chatbots for workplace use, boasting the lowest hallucination rate at just 8% among the 10 major models tested. In comparison, market leader ChatGPT registered one of the highest hallucination rates at 35%, just behind Google’s Gemini, which registered a high hallucination rate of 38%. The findings highlight Grok’s factual prowess despite the AI model’s lower market visibility. ## **Grok tops hallucination metric** The research evaluated chatbots on hallucination rate, customer ratings, response consistency, and downtime rate. The chatbots were then assigned a reliability risk score from 0 to 99, with higher scores indicating bigger problems. Grok achieved an 8% hallucination rate, 4.5 customer rating, 3.5 consistency, and 0.07% downtime, resulting in an overall risk score of just 6. DeepSeek followed closely with 14% hallucinations and zero downtime for a stellar risk score of 4. ChatGPT’s high hallucination and downtime rates gave it the top risk score of 99, followed by Claude and Meta AI, which earned reliability risk scores of 75 and 70, respectively. ## **Why low hallucinations matter** Relum Chief Product Officer Razvan-Lucian Haiduc shared his thoughts about the study’s findings. “About 65% of US companies now use AI chatbots in their daily work, and nearly 45% of employees admit they’ve shared sensitive company information with these tools. These numbers show well how important chatbots have become in everyday work. “Dependence on AI tools will likely increase even more, so companies should choose their chatbots based on how reliable and fit they are for their specific business needs. A chatbot that everyone uses isn’t necessarily the one that works best for your industry or gives accurate answers for your tasks.” In a way, the study reveals a notable gap between AI chatbots’ popularity and performance, with Grok’s low hallucination rate positioning it as a strong choice for accuracy-critical applications. This was despite the fact that Grok is not used as much by users, at least compared to more mainstream AI applications such as ChatGPT. The post Elon Musk’s Grok records lowest hallucination rate in AI reliability study appeared first on TESLARATI. ## Related News & Research - [Bedmutha Industries Files SEBI Compliance Certificate on Dematerialised Securities](https://longbridge.com/en/news/281709434.md) - [Martinrea International (TSE:MRE) Insider Francesco Barbara Purchases 12,333 Shares](https://longbridge.com/en/news/281708648.md) - [Claude Subscriptions Will No Longer Cover Usage On 'Third-Party Tools'—Anthropic Cuts OpenClaw Access Amid Surging AI Demand](https://longbridge.com/en/news/281705754.md) - [BCCL halts Dhanbad mining after agitation; operations stalled since April 2](https://longbridge.com/en/news/281708938.md) - [ZAWYA: Yas Waterworld expansion is now open to guests](https://longbridge.com/en/news/281689225.md)