Learning Illumination Control in Diffusion Models
Just describe the lighting you want, and this open-source model delivers...
Researchers from the University of Maryland and IIT Delhi have developed a fully open-source pipeline for controlling illumination in images using diffusion models. The team—Nishit Anand, Manan Suri, Christopher Metzler, Dinesh Manocha, and Ramani Duraiswami—built a data engine that transforms well-lit images into supervised training triplets. Each triplet consists of a poorly-illuminated input image, a natural language lighting instruction (e.g., "make the room brighter with warm tones"), and a well-illuminated output image.
When fine-tuned on this data, their diffusion model achieves significant improvements over baseline models like Stable Diffusion 1.5, SDXL, and FLUX.1-dev in perceptual similarity, structural similarity, and identity preservation. The work was accepted to ICLR 2026's ReALM-GEN Workshop on Diffusion Models. By releasing all code, data, and model weights publicly, the team provides a reproducible solution that eliminates the need for heavy control inputs like depth maps, making advanced illumination control accessible to the open-source community.
- Fully open-source pipeline for illumination control in diffusion models using natural language prompts
- Data engine generates supervised training triplets from well-lit images without requiring depth maps
- Outperforms SD 1.5, SDXL, and FLUX.1-dev on perceptual similarity, structural similarity, and identity preservation
Why It Matters
Democratizes professional-grade lighting control for image generation, enabling precise edits via simple text instructions.