Image & Video

My humble study on the effects of prompting nonexistent words on CLIP-based diffusion models.

Fake words can manipulate AI art models in predictable, powerful ways.

Deep Dive

A 2.5-year study reveals 'undictionaries'—nonexistent words that produce consistent, specific effects in CLIP-based diffusion models like Stable Diffusion. The researcher has documented the phenomenon, classified outputs, and created a system for finding more of these hidden prompt triggers. This backdoor exists because users can directly probe the model's latent space. The technique may become obsolete if future models use LLM encoders or add protective layers.

Why It Matters

This exposes a fundamental vulnerability in how AI art models interpret language, allowing hidden manipulation.