Skip to content

aranciokov/HierArtEx-MMM2025

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

20 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

HierArtEx: Hierarchical Representations and Art Experts Supporting the Retrieval of Museums in the Metaverse

This code accompanies the paper "HierArtEx: Hierarchical Representations and Art Experts Supporting the Retrieval of Museums in the Metaverse" accepted for presentation at MMM 2025. DOI preprint

Overview of HierArtEx

Abstract: The improvements in Virtual Reality technologies are bringing more attention to the Metaverse and the nearly unlimited experiences available there. Among these, digital museums have seen an increase in the number of yearly visitors, especially after the COVID-19 pandemics. However, no tools are available to support the user in the searching process. To this end, we start investigating the Text-to-Museum retrieval task, involving museums composed of many rooms enriched by multimedia elements affecting their relevance to the user query. To model this complex type of data, we design HierArtEx, which leverages hierarchical representations to model the whole museum, while combining generic and art-specific knowledge for capturing the visual contents of each single room. We validate its effectiveness on Museums3k, a large dataset that we collect, containing 3000 realistic museums each annotated by a description of its contents. Moreover, qualitative analyses confirm favorable results and their alignment with real user queries, while also highlighting the shortcomings of standard evaluation protocols in retrieval, as they fail to capture all relevant museums.

General info on the structure

The repository contains three Jupyter Notebooks for the HierArtEx method (Complex3DScenesRetrieval_HierArtEx_CameraReady.ipynb), the method without the art experts (Complex3DScenesRetrieval_HierArtEx_wo_ArtExp_CameraReady.ipynb), and the baseline (Complex3DScenesRetrieval_Baseline_CameraReady.ipynb). It further includes a Jupyter Notebook for the qualitative analysis reported in the paper (QualitativeAnalysis_CameraReady.ipynb), and a folder (ArtExp_train) containing the details on how we trained the art experts. The train/val/test split is available in indices_museum_dataset.pkl.

Data

Example from the dataset

In this work, we relied on pre-extracted features from the Museums3k dataset (available here). In each museum, twelve screenshots are taken by rotating a camera within each room. The features are extracted using a pretrained CLIP, and are found in the path notebook/tmp_museums/open_clip_features_museums3k/images. Additionally, a paragraph is created for each museum depicting its contents. The features for these are found in the path notebook/tmp_museums/open_clip_features_museums3k/descriptions/tokens_strings. Raw descriptions are also found in the folder sentences (used for analysis and visualization purposes). The features extracted from the art experts are found in the folders starting with preextracted_vectors_wikiart. They were extracted using the art experts we pretrained on WikiArt.

Citation

If you found HierArtEx or the paired data at Museums3k useful, please cite our work:

@InProceedings{10.1007/978-981-96-2061-6_5,
author="Falcon, Alex
and Abdari, Ali
and Serra, Giuseppe",
editor="Ide, Ichiro
and Kompatsiaris, Ioannis
and Xu, Changsheng
and Yanai, Keiji
and Chu, Wei-Ta
and Nitta, Naoko
and Riegler, Michael
and Yamasaki, Toshihiko",
title="HierArtEx: Hierarchical Representations and Art Experts Supporting the Retrieval of Museums in the Metaverse",
booktitle="MultiMedia Modeling",
year="2025",
publisher="Springer Nature Singapore",
address="Singapore",
pages="60--73",
abstract="The improvements in Virtual Reality technologies are bringing more attention to the Metaverse and the nearly unlimited experiences available there. Among these, digital museums have seen an increase in the number of yearly visitors, especially after the COVID-19 pandemics. However, no tools are available to support the user in the searching process. To this end, we start investigating the Text-to-Museum retrieval task, involving museums composed of many rooms enriched by multimedia elements affecting their relevance to the user query. To model this complex type of data, we design HierArtEx, which leverages hierarchical representations to model the whole museum, while combining generic and art-specific knowledge for capturing the visual contents of each single room. We validate its effectiveness on Museums3k, a large dataset that we collect, containing 3000 realistic museums each annotated by a description of its contents. Moreover, qualitative analyses confirm favorable results and their alignment with real user queries, while also highlighting the shortcomings of standard evaluation protocols in retrieval, as they fail to capture all relevant museums.",
isbn="978-981-96-2061-6"
}

Acknowledgements

This work was supported by the PRIN 2022 "MUSMA" - CUP G53D23002930006 - "Funded by EU - Next-Generation EU - M4 C2 I1.1", and by the Department Strategic Plan (PSD) of the University of Udine - Interdepartmental Project on Artificial Intelligence (2020-25).

About

Implementation of HierArtEx: Hierarchical Representations and Art Experts Supporting the Retrieval of Museums in the Metaverse (MMM 2025).

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors