Bimodal network architectures for automatic generation of image annotation from text

09/05/2018
by   Mehdi Moradi, et al.
0

Medical image analysis practitioners have embraced big data methodologies. This has created a need for large annotated datasets. The source of big data is typically large image collections and clinical reports recorded for these images. In many cases, however, building algorithms aimed at segmentation and detection of disease requires a training dataset with markings of the areas of interest on the image that match with the described anomalies. This process of annotation is expensive and needs the involvement of clinicians. In this work we propose two separate deep neural network architectures for automatic marking of a region of interest (ROI) on the image best representing a finding location, given a textual report or a set of keywords. One architecture consists of LSTM and CNN components and is trained end to end with images, matching text, and markings of ROIs for those images. The output layer estimates the coordinates of the vertices of a polygonal region. The second architecture uses a network pre-trained on a large dataset of the same image types for learning feature representations of the findings of interest. We show that for a variety of findings from chest X-ray images, both proposed architectures learn to estimate the ROI, as validated by clinical annotations. There is a clear advantage obtained from the architecture with pre-trained imaging network. The centroids of the ROIs marked by this network were on average at a distance equivalent to 5.1 of the ground truth ROIs.

READ FULL TEXT
research
06/08/2022

One Hyper-Initializer for All Network Architectures in Medical Image Analysis

Pre-training is essential to deep learning model performance, especially...
research
01/12/2018

TieNet: Text-Image Embedding Network for Common Thorax Disease Classification and Reporting in Chest X-rays

Chest X-rays are one of the most common radiological examinations in dai...
research
09/04/2021

Improving Joint Learning of Chest X-Ray and Radiology Report by Word Region Alignment

Self-supervised learning provides an opportunity to explore unlabeled ch...
research
12/20/2017

Partial Labeled Gastric Tumor Segmentation via patch-based Reiterative Learning

Gastric cancer is the second leading cause of cancer-related deaths worl...
research
09/03/2020

A Comparison of Pre-trained Vision-and-Language Models for Multimodal Representation Learning across Medical Images and Reports

Joint image-text embedding extracted from medical images and associated ...
research
12/27/2021

Event-based clinical findings extraction from radiology reports with pre-trained language model

Radiology reports contain a diverse and rich set of clinical abnormaliti...
research
06/21/2019

Boosting the rule-out accuracy of deep disease detection using class weight modifiers

In many screening applications, the primary goal of a radiologist or ass...

Please sign up or login with your details

Forgot password? Click here to reset