VICSOM: VIsual Clues from SOcial Media for psychological assessment

05/15/2019
by   Mohammad Mahdi Dehshibi, et al.
0

Sharing multimodal information (typically images, videos or text) in Social Network Sites (SNS) occupies a relevant part of our time. The particular way how users expose themselves in SNS can provide useful information to infer human behaviors. This paper proposes to use multimodal data gathered from Instagram accounts to predict the perceived prototypical needs described in Glasser's choice theory. The contribution is two-fold: (i) we provide a large multimodal database from Instagram public profiles (more than 30,000 images and text captions) annotated by expert Psychologists on each perceived behavior according to Glasser's theory, and (ii) we propose to automate the recognition of the (unconsciously) perceived needs by the users. Particularly, we propose a baseline using three different feature sets: visual descriptors based on pixel images (SURF and Visual Bag of Words), a high-level descriptor based on the automated scene description using Convolutional Neural Networks, and a text-based descriptor (Word2vec) obtained from processing the captions provided by the users. Finally, we propose a multimodal fusion of these descriptors obtaining promising results in the multi-label classification problem.

READ FULL TEXT

page 2

page 6

research
06/01/2020

A multimodal approach for multi-label movie genre classification

Movie genre classification is a challenging task that has increasingly a...
research
08/08/2016

Detecting Sarcasm in Multimodal Social Platforms

Sarcasm is a peculiar form of sentiment expression, where the surface se...
research
10/18/2017

Learning Social Image Embedding with Deep Multimodal Attention Networks

Learning social media data embedding by deep models has attracted extens...
research
10/28/2020

Fusion Models for Improved Visual Captioning

Visual captioning aims to generate textual descriptions given images. Tr...
research
03/06/2017

Cats and Captions vs. Creators and the Clock: Comparing Multimodal Content to Context in Predicting Relative Popularity

The content of today's social media is becoming more and more rich, incr...
research
08/12/2018

Multimodal Differential Network for Visual Question Generation

Generating natural questions from an image is a semantic task that requi...
research
05/24/2021

Assessing perceived organizational leadership styles through twitter text mining

We propose a text classification tool based on support vector machines f...

Please sign up or login with your details

Forgot password? Click here to reset