Instance-Warp: Saliency Guided Image Warping for Unsupervised Domain Adaptation

WACV 2025

* indicates equal contribution
Boreas Rain
Boreas Snow
Boreas Night
Fog
Highway Fog
Highway Rain
Heavy Snow
Tunnel
Rain

Our adapted detection results on in-the-wild driving videos taken in bad weather.

Overview and Highlights

We oversample salient object regions by warping source-domain images in-place during training while performing domain adaptation.

Our approach improves adaptation across geographies, lighting and weather conditions, is agnostic to the task, domain adaptation algorithm, saliency guidance, and underlying model architecture. Our approach adds minimal memory during training and incurs no additional latency at inference time.

Highlights include:

  • +6.1 mAP50 for BDD100K Clear → DENSE Foggy
  • +3.7 mAP50 for BDD100K Day → Night
  • +3.0 mAP50 for BDD100K Clear → Rainy
  • +6.3 mIoU for Cityscapes → ACDC
  • Why Warping for Domain Adaptation?

    Domain adaptation methods often treat background and object regions uniformly. However, backgrounds occupy more pixels and exhibit large cross-domain variations, making them challenging to learn. Consequently, focusing on salient object-containing regions while reducing attention to background context can lead to more robust and adaptable recognition models.

    InstanceWarp for Domain Adaptation.

    We propose in-place warping of images based on the locations of object instances. Warped images retain the same size as the original images, but object regions are oversampled using saliency guidance. This approach compels networks to focus on object regions and reduce attention on background context when learning features.

    InstanceWarp for Domain Adaptation.

    Which image regions to warp?

    A warping guidance, specifically a saliency-based approach to oversample certain regions over others, is essential for oversampling image regions. We propose instance-level saliency guidance, which explicitly oversamples all objects during training. We could have used Static Prior Guidance [Thavamani 2021, Thavamani 2023] or Geometric Prior Guidance [Ghosh 2023], However, these methods are not designed for domain adaptation and do not explicitly oversample object instance regions, which has proven to be the most effective approach.

    Comparison of InstanceWarp for Domain Adaptation.

    Saliency-Guided Warping for Domain Adaptation

    We warp source images based on saliency guidance and then unwarp the backbone features using the same guidance before making predictions. This method can be seamlessly integrated into existing domain adaptation algorithms and is agnostic to the task, domain adaptation algorithm, saliency guidance, and underlying model architecture. Empirically, we observed that warping source domain images is more effective than warping both source and target domain images.

    Workflow of InstanceWarp for Domain Adaptation.

    Improved Backbone Features

    Grad-CAM visualization shows that the model trained with our method demonstrate a higher focus on salient objects, indicating better-learned features and improved scene comprehension.

    Feature Explanation

    Importantly, our learned features exhibit better object focus with less attention on background context. This ensures that the features are more adaptable and robust to variations in background context.

    Feature Explanation.

    Improves a variety of adaptation scenarios

    As the learned backbone features are better, our approach improves performance across various real-to-real domain adaptation tasks, including changing weather, lighting conditions, and geographies. The results below are from an adapted model pre-trained on BDD100K images taken during the day and in good weather conditions.

    Teaser.

    Method is Task, Algorithm and Backbone Agnostic

    Our method is task, adaptation algorithm, and backbone agnostic. It can be used for semantic segmentation, object detection, and other tasks with CNN or Transformer backbones.

    Workflow of InstanceWarp for Domain Adaptation.