Research & Papers

Learning Illumination Control in Diffusion Models

Just describe the lighting you want, and this open-source model delivers...

Deep Dive

Researchers from the University of Maryland and IIT Delhi have developed a fully open-source pipeline for controlling illumination in images using diffusion models. The team—Nishit Anand, Manan Suri, Christopher Metzler, Dinesh Manocha, and Ramani Duraiswami—built a data engine that transforms well-lit images into supervised training triplets. Each triplet consists of a poorly-illuminated input image, a natural language lighting instruction (e.g., "make the room brighter with warm tones"), and a well-illuminated output image.

When fine-tuned on this data, their diffusion model achieves significant improvements over baseline models like Stable Diffusion 1.5, SDXL, and FLUX.1-dev in perceptual similarity, structural similarity, and identity preservation. The work was accepted to ICLR 2026's ReALM-GEN Workshop on Diffusion Models. By releasing all code, data, and model weights publicly, the team provides a reproducible solution that eliminates the need for heavy control inputs like depth maps, making advanced illumination control accessible to the open-source community.

Key Points
  • Fully open-source pipeline for illumination control in diffusion models using natural language prompts
  • Data engine generates supervised training triplets from well-lit images without requiring depth maps
  • Outperforms SD 1.5, SDXL, and FLUX.1-dev on perceptual similarity, structural similarity, and identity preservation

Why It Matters

Democratizes professional-grade lighting control for image generation, enabling precise edits via simple text instructions.