--- title: "DeepSeek V3 \"Error at the Door\": I am ChatGPT" description: "DeepSeek V3 has recently sparked heated discussions due to mistakenly referring to itself as ChatGPT during testing. Although some netizens questioned whether DeepSeek V3 was trained on ChatGPT's outp" type: "news" locale: "en" url: "https://longbridge.com/en/news/223493342.md" published_at: "2024-12-29T05:50:32.000Z" --- # DeepSeek V3 "Error at the Door": I am ChatGPT > DeepSeek V3 has recently sparked heated discussions due to mistakenly referring to itself as ChatGPT during testing. Although some netizens questioned whether DeepSeek V3 was trained on ChatGPT's outputs, the overall discussion suggests that this possibility is unlikely. Netizen Riley Goodside pointed out that DeepSeek V3's performance is unrelated to its training data, and it scored close to Llama 3.1 405B in the Pile test, indicating that its quality is not influenced by ChatGPT data If we talk about the hottest topic in the large model circle these days, it must be DeepSeek V3. However, while netizens are testing it, a bug has also become a hot topic of discussion—— **Just missing a question mark, DeepSeek V3 actually claims to be ChatGPT.** Even when asked to tell a joke, the generated result is the same as ChatGPT: Additionally, one highlight of DeepSeek V3's explosive popularity is that the training only cost $5.576 million. As a result, some people began to doubt: **Could it be trained based on ChatGPT's output?** Coincidentally, **Altman** also posted a status, seemingly hinting at something... However, DeepSeek V3 is not the first large model to have a **"reporting error"** issue. For example, **Gemini** once claimed to be Baidu's **Wenxin Yiyan**... So what exactly is going on here? ## Why did DeepSeek V3 report an error? First, it is important to emphasize that, based on the overall discussion among netizens, the **possibility that DeepSeek V3 was trained on ChatGPT's output is low**. The reason for this, as summarized by netizen Riley Goodside, is—— **because the shadow of ChatGPT is everywhere.** > Even if DeepSeek V3 intentionally trained on ChatGPT's output, it doesn't matter. Almost all large models that appeared after ChatGPT have encountered it. > > For example, ShareGPT, a not-so-new ChatGPT dialogue dataset, has been adjusted by many people using it and other ChatGPT data sources. But even so, no large model has emerged at the level of DeepSeek V3 Immediately after, Riley Goodside presented some evidence from the DeepSeek V3 report: > Moreover, if ChatGPT data was used, some quality issues regarding DeepSeek V3 would be inexplicable. > > For example, in the Pile test (the effect of compressing the base model Pile), DeepSeek V3 scored almost the same as Llama 3.1 405B, which is unrelated to whether it was exposed to ChatGPT data. > > Additionally, the report states that 95% of GPU-hours were used for pre-training the base model, and even if it was related to ChatGPT data, this part would occur in the post-training phase (the last 5%). Instead of focusing on whether ChatGPT data was used, perhaps we should pay more attention to why large models frequently encounter the "error at the door" issue. TechCrunch provided a sharp comment on this issue: > **Because the places where AI companies obtain data—the internet—are filled with AI garbage.** After all, a report from the European Union once predicted that by 2026, 90% of online content could be AI-generated. This kind of "AI pollution" makes it difficult to "thoroughly filter AI outputs from training data." Heidy Khlaaf, the chief scientist at AI Now Institute, stated: > Despite the risks, developers are still attracted by the cost savings brought by "distilling" knowledge from existing AI models. > > Models trained unexpectedly on outputs from ChatGPT or GPT-4 may not necessarily exhibit outputs reminiscent of OpenAI's customized messages. Now, regarding the hotly debated question among netizens, Quantum Bit conducted a round of practical tests, and DeepSeek V3 has not yet resolved this bug. It still lacks a question mark, and the response results vary: ## DeepSeek V3 More Features However, to be fair, the vast majority of netizens have given a big thumbs up to the capabilities of DeepSeek V3. This can be evidenced by the collective praise of "elegant" from various AI big shots here. In the past few days, netizens have successively shared more **practical uses** powered by DeepSeek V3. For example, one netizen pitted DeepSeek V3 against Claude Sonnet 3.5, using both to **create websites** in Scroll Hub: After testing, the blogger believes DeepSeek V3 completely outperformed! Another netizen shared their experience using DeepSeek V3 in an **AI video editor**. They stated that they no longer need to waste time on FFMPEG commands; DeepSeek V3 is not only free but can also change your workflow: The AI programming tool **Cursor** can also be integrated with DeepSeek V3; let's look at a case of making a snake game: Well, DeepSeek V3 is indeed quite useful. ## One More Thing Regarding the previously released 53-page paper, some netizens also noticed a non-technical detail— The contribution list not only showcases technical personnel but also data annotators and business staff: Netizens believe this approach aligns very well with DeepSeek's tone: Author of this article: Quantum Bit, Source: Quantum Bit, Original title: "DeepSeek V3 'Error at the Door': I am ChatGPT" Risk Warning and Disclaimer The market has risks, and investment requires caution. This article does not constitute personal investment advice and does not take into account the specific investment goals, financial conditions, or needs of individual users. Users should consider whether any opinions, views, or conclusions in this article are suitable for their specific circumstances. Investment based on this is at one's own risk ### Related Stocks - [OpenAI.NA - OpenAI](https://longbridge.com/en/quote/OpenAI.NA.md) - [09888.HK - BIDU-SW](https://longbridge.com/en/quote/09888.HK.md) - [002230.CN - IFLYTEK](https://longbridge.com/en/quote/002230.CN.md) - [00020.HK - SENSETIME-W](https://longbridge.com/en/quote/00020.HK.md) - [GOOGL.US - Alphabet](https://longbridge.com/en/quote/GOOGL.US.md) - [GOOG.US - Alphabet - C](https://longbridge.com/en/quote/GOOG.US.md) ## Related News & Research | Title | Description | URL | |-------|-------------|-----| | Sam Altman And Dario Amodei Stir Controversy At India AI Summit Amid Photo-Op Gesture— OpenAI CEO Says 'I Just Wasn't Sure...' | At the India AI Impact Summit, OpenAI CEO Sam Altman and Anthropic CEO Dario Amodei stirred controversy by opting out of | [Link](https://longbridge.com/en/news/276431749.md) | | After AMD, OpenAI Partners With Tata To Build Massive 1GW AI Data Center In India | OpenAI has partnered with Tata Group and Tata Consultancy Services to develop a large-scale AI data center in India, wit | [Link](https://longbridge.com/en/news/276304570.md) | | OpenAI expands agentic commerce push | By embedding structured product data and checkout flows directly into ChatGPT, OpenAI is seeking to position AI as the f | [Link](https://longbridge.com/en/news/276071558.md) | | Chinese tech companies progress 'remarkable,' OpenAI's Altman tells CNBC | OpenAI's Sam Altman praised the rapid progress of Chinese tech companies in AI during an AI summit in New Delhi. He note | [Link](https://longbridge.com/en/news/276315901.md) | | Altman and Amodei share a moment of awkwardness at India’s big AI summit | At the India AI Impact Summit, a moment of awkwardness arose when OpenAI's Sam Altman and Anthropic's Dario Amodei did n | [Link](https://longbridge.com/en/news/276340986.md) | --- > **Disclaimer**: This article is for reference only and does not constitute any investment advice.