CVPR Image Matching workshop and Transformer workshop, 2022
Xi Shen1 Alexei A. Efros2 Armand Joulin3 Mathieu Aubry1
1LIGM (UMR 8049) - Ecole des Ponts, UPE 2UC Berkeley 3Facebook AI Research
The goal of this work is to efficiently identify visually similar patterns from a pair of images, e.g. identifying an artwork detail copied between an engraving and an oil painting, or matching a night-time photograph with its daytime counterpart. Lack of training data is a key challenge for this task. We present a simple yet surprisingly effective approach to overcome this difficulty: we generate synthetic training pairs by selecting object segments in an image and copy-pasting them into another image. We then learn to predict the repeated object masks. We find that it is crucial to predict the correspondences as an auxiliary task and to use Poisson blending and style transfer on the training pairs to generalize on real data. We analyse results with two deep architectures relevant to our joint image analysis task: a transformer-based architecture and Sparse Nc-Net, a recent network designed to predict coarse correspondences using 4D convolutions. We show our approach provides clear improvements for artwork details retrieval on the Brueghel dataset and achieves competitive performance on two place recognition benchmarks, Tokyo247 and Pitts30K. We then demonstrate the potential of our approach by performing object discovery on the Internet object discovery dataset and the Brueghel dataset.
Video (11mins) |
To cite our paper,
@article{shen2021learning, title={Learning Co-segmentation by Segment Swapping for Retrieval and Discovery}, author={Shen, Xi and Efros, Alexei A and Joulin, Armand and Aubry, Mathieu}, journal={arXiv}, year={2021} }
This work was supported in part by ANR project EnHerit ANR-17-CE23-0008, project Rapid Tabasco, and IDRIS under the allocation AD011011160R1 made by GENCI.