Automatic Modeling of Social Concepts Evoked by Art Images as Multimodal Frames

Social concepts referring to non-physical objects–such as revolution, violence, or friendship–are powerful tools to describe, index, and query the content of visual data, including ever-growing collections of art images from the Cultural Heritage (CH) field. While much progress has been made towards complete image understanding in computer vision, automatic detection of social concepts evoked by images is still a challenge. This is partly due to the well-known semantic gap problem, worsened for social concepts given their lack of unique physical features, and reliance on more unspecific features than concrete concepts. In this paper, we propose the translation of recent cognitive theories about social concept representation into a software approach to represent them as multimodal frames, by integrating multisensory data. Our method focuses on the extraction, analysis, and integration of multimodal features from visual art material tagged with the concepts of interest. We define a conceptual model and present a novel ontology for formally representing social concepts as multimodal frames. Taking the Tate Gallery's collection as an empirical basis, we experiment our method on a corpus of art images to provide a proof of concept of its potential. We discuss further directions of research, and provide all software, data sources, and results.

READ FULL TEXT

page 1

page 3

page 5

page 6

page 8

page 9

page 11

page 15

research
08/21/2023

Seeing the Intangible: Surveying Automatic High-Level Visual Understanding from Still Images

The field of Computer Vision (CV) was born with the single grand goal of...
research
01/05/2018

Ontology-based Approach for Semantic Data Extraction from Social Big Data: State-of-the-art and Research Directions

A challenge of managing and extracting useful knowledge from social medi...
research
03/07/2019

Understanding Ancient Coin Images

In recent years, a range of problems within the broad umbrella of automa...
research
04/25/2020

Fuzzy Logic Based Integration of Web Contextual Linguistic Structures for Enriching Conceptual Visual Representations

Due to the difficulty of automatically mapping visual features with sema...
research
04/18/2018

Quantifying the visual concreteness of words and topics in multimodal datasets

Multimodal machine learning algorithms aim to learn visual-textual corre...
research
04/20/2023

Not Only Generative Art: Stable Diffusion for Content-Style Disentanglement in Art Analysis

The duality of content and style is inherent to the nature of art. For h...
research
05/01/2013

MATAWS: A Multimodal Approach for Automatic WS Semantic Annotation

Many recent works aim at developing methods and tools for the processing...

Please sign up or login with your details

Forgot password? Click here to reset