International Journal of Computer Vision (IJCV), 2022
Xi Shen1 Robin Champenois1 Shiry Ginosar2 Ilaria Pastrolin3 Morgane Rousselot3 Oumayma Bounou4 Tom Monnier1 Spyros Gidaris5 François Bougard6 Pierre-Guillaume Raverdy4 Marie-Françoise Limon7 Christine Bénévent3 Marc Smith3 Olivier Poncet3 K. Bender8 Béatrice Joyeux-Prunel 9 Elizabeth Honig2 Alexei A. Efros2 Mathieu Aubry1
1LIGM (UMR 8049) - École des Ponts, UPE 2UC Berkeley 3École Nationale des Chartes 4Inria 5Valeo AI 6IRHT CNRS 7Archives Nationales 8Independent Researcher 9Université de Genève
Code & Data ArtMiner Code & Data Watermark PDF (springer) PDF (HAL)
Progress in the digitization of cultural assets leads to online databases that become too large for a human to analyze. Moreover, some analyses might be challenging, even for experts. In this paper, we explore two applications of computer vision to analyze historical data: watermark recognition and one-shot repeated pattern detection in artwork collections. Both problems present computer vision challenges which we believe to be representative of the ones encountered in cultural heritage applications: limited supervision is available, the tasks are fine-grained recognition, and the data comes in several different modalities. Both applications are also highly practical, as recognizing watermarks makes it possible to date and locate documents, while detecting repeated patterns allows exploring visual links between artworks. We demonstrate on both tasks the benefits of relying on deep mid-level features. More precisely, we define an image similarity score based on geometric verification of mid-level features and show how spatial consistency can be used to fine-tune out-of-the-box features for the target dataset with weak or no supervision. This paper relates and extends our previous works ArtMiner and Historical Watermark Recognition
To cite our paper,
@article{shen2021spatially, title={Spatially-consistent Feature Matching and Learning for Art Collections and Watermark Recognition}, author={Shen, Xi and Champenois, Robin and Ginosar, Shiry and Pastrolin, Ilaria and Rousselot, Morgane and Bounou, Oumayma and Monnier, Tom and Gidaris, Spyros and Bougard, Fran{\c{c}}ois and Raverdy, Pierre-Guillaume and Limon, Marie-Fran{\c{c}}oise and B{\'{e}}n{\'{e}}vent, Christine and Smith, Marc and Poncet, Olivier and Bender, K and B{\'{e}}atrice, Joyeux-Prunel and Honig, Elizabeth and Efros, Alexei A and Aubry, Mathieu}, journal = {International Journal of Computer Vision (IJCV)}, year={2022} }
This work was supported in part by ANR project EnHerit ANR-17-CE23-0008, PSL Filigrane pour tous project, project Rapid Tabasco, gifts from Adobe to Ecole des Ponts. We thank Minsu Cho, Pascal Monasse and Renaud Marlet for fruitful discussions, and Kenny Oh and Davienne Shields for thier help on defining the task and building the Brueghel dataset.