“Maybe I should try Chat again after only using Claude for a while”. First response:
User seeking facial feature analysis gets unsettling description of 'white creamy' substance and 'pale yellow liquid'.
A viral Reddit post has exposed a disturbing flaw in ChatGPT's image analysis capabilities. A user with prosopagnosia (facial blindness) asked the AI to describe facial features in an image for assistance, expecting a neutral, factual breakdown. Instead, ChatGPT generated a bizarre and unsettling description, stating the person's features were "completely obscured" because she was covered in something "white and creamy," with "pale yellow liquid" dripping down her face. The user, u/Severe-Magician5981, noted they decided to interpret the odd output as "something funny" because the actual description was "kind of weird," sparking a discussion about the AI's reliability for sensitive tasks.
This incident highlights a significant gap between user expectation and AI performance in computer vision. While tools like GPT-4V are marketed for detailed image understanding, this case shows they can still hallucinate grotesque or nonsensical details, especially in ambiguous scenarios. The post has resonated widely because it involves a user with a genuine disability relying on the tool for practical help, only to be met with a confusing and mildly alarming response. It raises critical questions about the robustness of safety filters and content moderation for multimodal AI systems before they are deployed for assistive purposes.
- ChatGPT generated a bizarre description of a 'creamy' face with 'pale yellow liquid' for a user with facial blindness.
- The viral Reddit post highlights a failure in AI's multimodal reasoning and safety filters for image analysis.
- The incident questions the reliability of AI assistants for sensitive, real-world tasks like disability aid.
Why It Matters
Shows AI can fail unpredictably in assistive roles, undermining trust in critical applications for vulnerable users.