TikTok Scraps AI Summaries After Hallucinating Blueberries and Rubber Chickens

4

TikTok has significantly scaled back its “AI Overviews” feature, a move driven by a string of bizarre and inaccurate descriptions that undermined user trust. The platform, which had been testing an automated system to generate text summaries beneath video posts, has decided to pivot. Instead of providing narrative context, the updated feature will now focus strictly on identifying products shown in videos.

This retreat highlights the persistent challenges tech giants face when deploying generative AI in real-time content environments. While the technology promises to enhance discoverability and accessibility, the current iteration proved too prone to “hallucinations”—instances where the AI confidently invents details that are factually wrong or completely unrelated to the visual content.

The “Blueberry” Blunder and Other Failures

The decision to dial back the feature came after users and journalists documented numerous egregious errors. The AI’s attempts to interpret video content often resulted in surreal misinterpretations that confused rather than clarified the viewing experience.

Key examples of these failures include:

  • Celebrity Misidentification: A video featuring top TikTok creator Charli D’Amelio speaking to the camera was described by the AI as “a collection of various blueberries with different toppings.”
  • Abstract Interpretations: A clip of singer Shakira was labeled as “a repetitive sequence of several distinct blue shapes appearing and moving across the screen.”
  • Absurd Action Recognition: In a Reddit-shared example, a video of two ballroom dancers was misidentified as “a person repeatedly striking their head with a rubber chicken.”

These errors were not isolated incidents but part of a broader pattern of inaccuracy that frustrated users who expected the feature to provide useful context, akin to the summaries seen in Google Search.

Why Context Matters in Short-Form Video

TikTok’s initial ambition with AI Overviews was to mirror the utility of Google’s AI-generated search results. The goal was to explain what was happening in a video, offering additional context for users who might be scrolling silently or seeking quick information.

However, short-form video is inherently complex. It relies on rapid cuts, visual humor, cultural references, and nuanced body language—elements that current AI models struggle to parse accurately. When the AI fails to recognize a human face or a common activity, it often defaults to generic object recognition, leading to the jarring disconnect between the video content and the text description.

The Core Issue: Accuracy is paramount for AI summaries. When a tool meant to clarify content instead creates confusion, it erodes user confidence and adds noise to the platform.

A Step Back in AI Adoption

TikTok’s pivot to product identification is a pragmatic adjustment. By limiting the AI’s scope to recognizable commercial items, the platform can leverage its computer vision strengths while avoiding the pitfalls of natural language generation for open-ended creative content.

This incident mirrors broader industry trends. Just a few years ago, Google faced similar scrutiny when its early AI search experiments suggested nonsensical advice, such as eating rocks or using glue on pizza. These early missteps underscore a critical lesson for the tech industry: AI capabilities must evolve in tandem with safety and accuracy safeguards.

Conclusion

TikTok’s retreat from narrative AI summaries serves as a cautionary tale for other platforms experimenting with generative AI. While the technology holds promise for enhancing user experience, it is not yet ready to reliably interpret the nuanced, fast-paced nature of social media video. For now, identifying products is a safer, more functional application than attempting to summarize human creativity.