In this paper, we introduce Latent Bridge Matching (LBM), a new, versatile and scalable method that relies on Bridge Matching in a latent space to achieve fast image-to-image translation. We show that the method can reach state-of-the-art results for various image-to-image tasks using only a single inference step. In addition to its efficiency, we also demonstrate the versatility of the method across different image translation tasks such as object removal, normal and depth estimation, and object relighting. We also derive a conditional framework of LBM and demonstrate its effectiveness by tackling the tasks of controllable image relighting and shadow generation. We provide an open-source implementation of the method at https://github.com/gojasper/LBM.
Image-to-image translation is a task that can be considered as a transport problem where the goal is to transport the distribution of the source images (e.g. composite images) to the distribution of the target images (e.g. relighted images). In the proposed Latent Bridge Matching (LBM) method, given paired images, we propose to encode the source and target image into a latent space and then build a stochastic path called a Brownian Bridge between them. In particular, the stochasticity of these paths makes the method differ from flow matching and allows reaching a wider diversity of samples.
The training procedure is as follows and is detailed in the figure above. First, we draw a pair of images. Those samples are first encoded into the latent space using a pre-trained VAE giving the corresponding latents. We create a Brownian Bridge between these two latents. A timestep is drawn from a well chosen distribution to get the latent on the trajectory at this given timestep. This sample is then passed to the denoiser which predicts the drift of the associated Stochastic Differential Equation (SDE).
During training, we also introduce a pixel loss that consists of decoding the estimated target latent and comparing it to the real target image. We found that LPIPS works well in practice and speeds up domain shift. In order to scale with the image size, we put in place a random cropping strategy and only compute the loss on a patch if the image size is larger than a certain threshold. This limits the memory footprint of the model so it does not become a burden to the training efficiency.
Original Image
Generated Image (1 NFE)
Original Image
Generated Image (1 NFE)
@article{chadebec2025lbm,
title={LBM: Latent Bridge Matching for Fast Image-to-Image Translation},
author={Clément Chadebec and Onur Tasar and Sanjeev Sreetharan and Benjamin Aubin},
year={2025},
journal = {arXiv preprint arXiv:2503.07535},
}