• WIRED tested ChatGPT on its own product recommendations and found the AI hallucinated completely incorrect answers

  • The chatbot fabricated product picks for TVs, headphones, and laptops that WIRED’s reviewers never recommended

  • This reveals a trust crisis as consumers increasingly use AI chatbots for purchase decisions worth hundreds or thousands of dollars

  • The findings highlight ongoing accuracy problems with large language models despite OpenAI’s improvements to ChatGPT

OpenAI’s ChatGPT just failed a basic accuracy test. When WIRED asked the chatbot to cite the publication’s own product recommendations for TVs, headphones, and laptops, it confidently delivered wrong answers across the board. The experiment, published April 1, exposes a critical vulnerability in how millions of consumers now rely on AI for shopping advice – and how often these systems simply make things up.

OpenAI’s flagship chatbot just got caught making up product recommendations wholesale. In an investigation published by WIRED, reporter Reece Rogers asked ChatGPT a seemingly straightforward question: What products does WIRED’s review team actually recommend? The results weren’t just slightly off – they were completely fabricated.

The AI confidently cited specific TVs, headphones, and laptops as WIRED’s top picks, none of which the publication’s reviewers had actually endorsed. It’s a textbook case of AI hallucination, the industry term for when large language models generate plausible-sounding but entirely false information. And it’s happening at scale as millions of shoppers turn to ChatGPT for buying advice.

This isn’t an academic problem. Consumers are making real purchasing decisions based on ChatGPT’s recommendations, often involving products costing hundreds or thousands of dollars. The bot speaks with such confidence that most users wouldn’t think to fact-check its claims against the original sources. Rogers’ experiment shows exactly why that’s dangerous.

The timing couldn’t be more awkward for OpenAI. The company has been pushing ChatGPT as a reliable research and shopping assistant, recently adding features that surface product recommendations and shopping links. But if the underlying model can’t accurately cite what a publication actually recommends – even when that publication is well-indexed and widely available online – it raises serious questions about the technology’s readiness for consumer advice.