Generate Image Descriptions based on Deep RNN and Memory Cells for Images Features

02/05/2016
by   Shijian Tang, et al.
0

Generating natural language descriptions for images is a challenging task. The traditional way is to use the convolutional neural network (CNN) to extract image features, followed by recurrent neural network (RNN) to generate sentences. In this paper, we present a new model that added memory cells to gate the feeding of image features to the deep neural network. The intuition is enabling our model to memorize how much information from images should be fed at each stage of the RNN. Experiments on Flickr8K and Flickr30K datasets showed that our model outperforms other state-of-the-art models with higher BLEU scores.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/04/2014

Explain Images with Multimodal Recurrent Neural Networks

In this paper, we present a multimodal Recurrent Neural Network (m-RNN) ...
research
05/03/2017

Going Wider: Recurrent Neural Network With Parallel Cells

Recurrent Neural Network (RNN) has been widely applied for sequence mode...
research
12/20/2014

Deep Captioning with Multimodal Recurrent Neural Networks (m-RNN)

In this paper, we present a multimodal Recurrent Neural Network (m-RNN) ...
research
05/23/2016

Generative Choreography using Deep Learning

Recent advances in deep learning have enabled the extraction of high-lev...
research
12/09/2022

Decomposing a Recurrent Neural Network into Modules for Enabling Reusability and Replacement

Can we take a recurrent neural network (RNN) trained to translate betwee...
research
11/09/2015

Visual Language Modeling on CNN Image Representations

Measuring the naturalness of images is important to generate realistic i...
research
03/24/2020

First Investigation Into the Use of Deep Learning for Continuous Assessment of Neonatal Postoperative Pain

This paper presents the first investigation into the use of fully automa...

Please sign up or login with your details

Forgot password? Click here to reset