Research Unveils Text2Relight: Text-Guided Portrait Lighting

Abstract

We present a lighting-aware image editing pipeline that, given a portrait image and a text prompt, performs single image relighting. Our model modifies the lighting and color of both the foreground and background to align with the provided text description. The unbounded nature in creativeness of a text allows us to describe the lighting of a scene with any sensory features including temperature, emotion, smell, time, and so on. However, the modeling of such mapping between the unbounded text and lighting is extremely challenging due to the lack of dataset where there exists no scalable data that provides large pairs of text and relighting, and therefore, current text-driven image editing models does not generalize to lighting-specific use cases. We overcome this problem by introducing a novel data synthesis pipeline: First, diverse and creative text prompts that describe the scenes with various lighting are automatically generated under a crafted hierarchy using a large language model (*e.g.,* ChatGPT). A text-guided image generation model creates a lighting image that best matches the text. As a condition of the lighting images, we perform image-based relighting for both foreground and background using a single portrait image or a set of OLAT (One-Light-at-A-Time) images captured from lightstage system. Particularly for the background relighting, we represent the lighting image as a set of point lights and transfer them to other background images. A generative diffusion model learns the synthesized large-scale data with auxiliary task augmentation (*e.g.,* portrait delighting and light positioning) to correlate the latent text and lighting distribution for text-guided portrait relighting.

An innovative AI model has been developed to create dynamic lighting effects in portrait images and videos using only text input. This technology allows users to adjust colors easily with descriptive prompts, such as 'warm, freshly cooked chicken' and 'icy blue light,' eliminating the need for complicated editing tools.

Professor Seungryul Bak and his team at the UNIST Artificial Intelligence Graduate School introduced Text2Relight, an AI-driven lighting-specific foundational model that can perform relighting of a single portrait image driven by a creative text prompt as shown in Figure 1.

This groundbreaking study, conducted in collaboration with Adobe, will be showcased at the 39th Annual AAAI Conference on Artificial Intelligence (AAAI-25) in Philadelphia, taking place at the Pennsylvania Convention Center from February 25 to March 4, 2025. The research has also been accepted by the Association for the Advancement of Artificial Intelligence (AAAI), one of the leading conferences in the field.

Figure 1 Text2Relight generates the image of a relighted portrait (right) as a condition of a text prompt while keeping original Figure 1: Text2Relight generates the image of a relighted portrait (right) as a condition of a text prompt while keeping original contents in an input image (left).

The new model excels in expressing diverse lighting characteristics, such as emotional ambiance, alongside color and brightness, all through natural language inputs. Notably, it adjusts the colors of both the subject and the background simultaneously, maintaining the integrity of the original image. In contrast to existing text-based image editing models that lack specialization in lighting data and often result in image distortion or limited lighting control, Text2Relight provides a more refined solution.

To enable the AI to learn the correlation between creative texts and lighting, the research team developed a large-scale synthetic dataset. They utilized ChatGPT and text-based diffusion models to generate lighting data, while also implementing OLAT (One-Light-at-A-Time) techniques and Lighting Transfer methods to explore various lighting conditions.

In addition, the team further enhanced the model's functionality by training auxiliary datasets focused on shadow removal and illumination positioning, thereby improving visual coherence and realism in lighting effects.

Professor Bak commented, "Text2Relight holds significant potential in content creation, including reducing editing time in photo and video production and enhancing immersion in virtual and augmented reality settings."

The study was participated by Junuk Cha, a researcher at the UNIST Graduate School of Artificial Intelligence, who served as the first author. This research was supported by Adobe and the Ministry of Science and ICT (MSIT).

Journal Reference

Junuk Cha, Mengwei Ren, Krishna Kumar Singh, et al., "Text2Relight: Creative Portrait Relighting with Text Guidance", in Proc. of Annual AAAI Conference on Artificial Intelligence (AAAI), Pennsylvania, USA, 2025.

/Public Release. This material from the originating organization/author(s) might be of the point-in-time nature, and edited for clarity, style and length. Mirage.News does not take institutional positions or sides, and all views, positions, and conclusions expressed herein are solely those of the author(s).View in full here.