DIME: An Online Tool for the Visual Comparison of Cross-modal Retrieval Models

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Cross-modal retrieval relies on accurate models to retrieve relevant results for queries across modalities such as image, text, and video. In this paper, we build upon previous work by tackling the difficulty of evaluating models both quantitatively and qualitatively quickly. We present DIME (Dataset, Index, Model, Embedding), a modality-agnostic tool that handles multimodal datasets, trained models, and data preprocessors to support straightforward model comparison with a web browser graphical user interface. DIME inherently supports building modality-agnostic queryable indexes and extraction of relevant feature embeddings, and thus effectively doubles as an efficient cross-modal tool to explore and search through datasets.

Author supplied keywords

Cite

CITATION STYLE

APA

Zhao, T., Choi, J., & Friedland, G. (2020). DIME: An Online Tool for the Visual Comparison of Cross-modal Retrieval Models. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11962 LNCS, pp. 729–733). Springer. https://doi.org/10.1007/978-3-030-37734-2_61

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free