Rich Image Captioning in the Wild

03/30/2016
by   Kenneth Tran, et al.
0

We present an image caption system that addresses new challenges of automatically describing images in the wild. The challenges include high quality caption quality with respect to human judgments, out-of-domain data handling, and low latency required in many applications. Built on top of a state-of-the-art framework, we developed a deep vision model that detects a broad range of visual concepts, an entity recognition model that identifies celebrities and landmarks, and a confidence model for the caption output. Experimental results show that our caption engine outperforms previous state-of-the-art systems significantly on both in-domain dataset (i.e. MS COCO) and out of-domain datasets.

READ FULL TEXT

page 1

page 6

research
06/15/2018

Partially-Supervised Image Captioning

Image captioning models are becoming increasingly successful at describi...
research
05/07/2023

UIT-OpenViIC: A Novel Benchmark for Evaluating Image Captioning in Vietnamese

Image Captioning is one of the vision-language tasks that still interest...
research
05/26/2022

Prompt-based Learning for Unpaired Image Captioning

Unpaired Image Captioning (UIC) has been developed to learn image descri...
research
01/25/2019

Improving Image Captioning by Leveraging Knowledge Graphs

We explore the use of a knowledge graphs, that capture general or common...
research
10/17/2017

Describing Natural Images Containing Novel Objects with Knowledge Guided Assitance

Images in the wild encapsulate rich knowledge about varied abstract conc...
research
09/24/2021

From images in the wild to video-informed image classification

Image classifiers work effectively when applied on structured images, ye...

Please sign up or login with your details

Forgot password? Click here to reset