Cross-media Structured Common Space for Multimedia Event Extraction

05/05/2020 ∙ by Manling Li, et al. ∙ University of Illinois at Urbana-Champaign Columbia University 0

We introduce a new task, MultiMedia Event Extraction (M2E2), which aims to extract events and their arguments from multimedia documents. We develop the first benchmark and collect a dataset of 245 multimedia news articles with extensively annotated events and arguments. We propose a novel method, Weakly Aligned Structured Embedding (WASE), that encodes structured representations of semantic information from textual and visual data into a common embedding space. The structures are aligned across modalities by employing a weakly supervised training strategy, which enables exploiting available resources without explicit cross-media annotation. Compared to uni-modal state-of-the-art methods, our approach achieves 4.0 event argument role labeling and visual event extraction. Compared to state-of-the-art multimedia unstructured representations, we achieve 8.3 5.0 labeling, respectively. By utilizing images, we extract 21.4 mentions than traditional text-only methods.



There are no comments yet.


page 3

page 5

page 8

page 9

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Figure 1: An example of Multimedia Event Extraction. An event mention and some event arguments (Agent and Person) are extracted from text, while the vehicle arguments can only be extracted from the image.

Traditional event extraction methods target a single modality, such as text wadden2019entity, images yatskar2016situation or videos ye2015eventnet; caba2015activitynet; soomro2012ucf101. However, the practice of contemporary journalism (Mitchell1998) distributes news via multimedia. By randomly sampling 100 multimedia news articles from the Voice of America (VOA), we find that 33% of images in the articles contain visual objects that serve as event arguments and are not mentioned in the text. Take Figure 1 as an example, we can extract the Agent and Person arguments of the Movement.Transport event from text, but can extract the Vehicle

argument only from the image. Nevertheless, event extraction is independently studied in Computer Vision (CV) and Natural Language Processing (NLP), with major differences in task definition, data domain, methodology, and terminology. Motivated by the complementary and holistic nature of multimedia data, we propose

MultiMedia Event Extraction (M2E2), a new task that aims to jointly extract events and arguments from multiple modalities. We construct the first benchmark and evaluation dataset for this task, which consists of 245 fully annotated news articles.

We propose the first method, Weakly Aligned Structured Embedding (WASE), for extracting events and arguments from multiple modalities. Complex event structures have not been covered by existing multimedia representation methods Wu2019UniVSERV; faghri2018vse++; karpathy2015deep, so we propose to learn a structured

multimedia embedding space. More specifically, given a multimedia document, we represent each image or sentence as a graph, where each node represents an event or entity and each edge represents an argument role. The node and edge embeddings are represented in a multimedia common semantic space, as they are trained to resolve event co-reference across modalities and to match images with relevant sentences. This enables us to jointly classify events and argument roles from both modalities. A major challenge is the lack of multimedia event argument annotations, which are costly to obtain due to the annotation complexity. Therefore, we propose a weakly supervised framework, which takes advantage of annotated uni-modal corpora to separately learn visual and textual event extraction, and uses an image-caption dataset to align the modalities.

We evaluate WASE on the new task of M2E2. Compared to the state-of-the-art uni-modal methods and multimedia flat representations, our method significantly outperforms on both event extraction and argument role labeling tasks in all settings. Moreover, it extracts 21.4% more event mentions than text-only baselines. The training and evaluation are done on heterogeneous data sets from multiple sources, domains and data modalities, demonstrating the scalability and transferability of the proposed model. In summary, this paper makes the following contributions:

  • We propose a new task, MultiMedia Event Extraction, and construct the first annotated news dataset as a benchmark to support deep analysis of cross-media events.

  • We develop a weakly supervised training framework, which utilizes existing single-modal annotated corpora, and enables joint inference without cross-modal annotation.

  • Our proposed method, WASE, is the first to leverage structured representations and graph-based neural networks for multimedia common space embedding.

2 Task Definition

2.1 Problem Formulation

Each input document consists of a set of images and a set of sentences . Each sentence can be represented as a sequence of tokens , where is a token from the document vocabulary . The input also includes a set of entities extracted from the document text. An entity is an individually unique object in the real world, such as a person, an organization, a facility, a location, a geopolitical entity, a weapon, or a vehicle. The objective of M2E2is twofold:

Event Extraction: Given a multimedia document, extract a set of event mentions, where each event mention has a type and is grounded on a text trigger word or an image or both, i.e.,

Note that for an event, and can both exist, which means the visual event mention and the textual event mention refer to the same event. For example in Figure 1, deploy indicates the same Movement.Transport event as the image. We consider the event as text-only event if it only has textual mention , and as image-only event if it only contains visual mention , and as multimedia event if both and exist.

Argument Extraction: The second task is to extract a set of arguments of event mention . Each argument has an argument role type , and is grounded on a text entity or an image object (represented as a bounding box), or both,

The arguments of visual and textual event mentions are merged if they refer to the same real-world event, as shown in Figure 1.

2.2 The M2E2 Dataset

We define multimedia newsworthy event types by exhaustively mapping between the event ontology in NLP community for the news domain (ACE222 and the event ontology in CV community for general domain (imSitu yatskar2016situation). They cover the largest event training resources in each community. Table 1 shows the selected complete intersection, which contains 8 ACE types (i.e., 24% of all ACE types), mapped to 98 imSitu types (i.e., 20% of all imSitu types). We expand the ACE event role set by adding visual arguments from imSitu, such as instrument, bolded in Table 1. This set encompasses 52% ACE events in a news corpus, which indicates that the selected eight types are salient in the news domain. We reuse these existing ontologies because they enable us to train event and argument classifiers for both modalities without requiring joint multimedia event annotation as training data.

Event Type Argument Role
Movement.Transport (22353) Agent (4664), Artifact (179103), Vehicle (2451), Destination (1200), Origin (660)
Conflict.Attack (32627) Attacker (19212), Target (20719), Instrument (3715), Place (1210)
Conflict.Demonstrate (15169) Entity (102184), Police (326), Instrument (0118), Place (8625)
Justice.ArrestJail (16056) Agent (64119), Person (14799), Instrument (011), Place (430)
Contact.PhoneWrite (3337) Entity (3346), Instrument (043), Place (80)
Contact.Meet (12779) Participant (119321), Place (680)
Life.Die    (24464) Agent (390), Instrument (42), Victim (165155), Place (540)
Transaction.   TransferMoney (336) Giver (193), Recipient (195),     Money (08)
Table 1: Event types and argument roles in M2E2, with expanded ones in bold. Numbers in parentheses represent the counts of textual and visual events/arguments.

We collect 108,693 multimedia news articles from the Voice of America (VOA) website 333 2006-2017, covering a wide range of newsworthy topics such as military, economy and health. We select 245 documents as the annotation set based on three criteria: (1) Informativeness: articles with more event mentions; (2) Illustration: articles with more images (); (3) Diversity: articles that balance the event type distribution regardless of true frequency. The data statistics are shown in Table 2. Among all of these events, 192 textual event mentions and 203 visual event mentions can be aligned as 309 cross-media event mention pairs. The dataset can be divided into 1,105 text-only event mentions, 188 image-only event mentions, and 395 multimedia event mentions.

Source Event Mention Argument Role
sentence image textual visual textual visual
6,167 1,014 1,297 391 1,965 1,429
Table 2: M2E2 data statistics.

We follow the ACE event annotation guidelines walker2006ace for textual event and argument annotation, and design an annotation guideline 444 for multimedia events annotation.

Figure 2: Example of bounding boxes.

One unique challenge in multimedia event annotation is to localize visual arguments in complex scenarios, where images include a crowd of people or a group of object. It is hard to delineate each of them using a bounding box. To solve this problem, we define two types of bounding boxes: (1) union bounding box: for each role, we annotate the smallest bounding box covering all constituents; and (2) instance bounding box: for each role, we annotate a set of bounding boxes, where each box is the smallest region that covers an individual participant (e.g., one person in the crowd), following the VOC2011 Annotation Guidelines555 Figure 2 shows an example. Eight NLP and CV researchers complete the annotation work with two independent passes and reach an Inter-Annotator Agreement (IAA) of 81.2%. Two expert annotators perform adjudication.

Figure 3: Approach overview. During training (left), we jointly train three tasks to establish a cross-media structured embedding space. During test (right), we jointly extract events and arguments from multimedia articles.
Figure 4: Multimedia structured common space construction. Red pixels stands for attention heatmap.

3 Method

3.1 Approach Overview

As shown in Figure 3, the training phase contains three tasks: text event extraction (Section 3.2), visual situation recognition (Section 3.3), and cross-media alignment (Section 3.4). We learn a cross-media shared encoder, a shared event classifier, and a shared argument classifier. In the testing phase (Section 3.5), given a multimedia news article, we encode the sentences and images into the structured common space, and jointly extract textual and visual events and arguments, followed by cross-modal coreference resolution.

3.2 Text Event Extraction

Text Structured Representation: As shown in Figure 4, we choose Abstract Meaning Representation (AMR) banarescu2013abstract to represent text because it includes a rich set of 150 fine-grained semantic roles. To encode each text sentence, we run the CAMR parser (wang-xue-pradhan:2015:NAACL-HLT; wang-xue-pradhan:2015:ACL-IJCNLP; wang-EtAl:2016:SemEval)

to generate an AMR graph, based on the named entity recognition and part-of-speech (POS) tagging results from Stanford CoreNLP 

(manning-EtAl:2014:P14-5). To represent each word in a sentence , we concatenate its pre-trained GloVe word embedding (pennington2014glove)

, POS embedding, entity type embedding and position embedding. We then input the word sequence to a bi-directional long short term memory (Bi-LSTM) 

graves2013speech network to encode the word order and get the representation of each word . Given the AMR graph, we apply a Graph Convolutional Network (GCN) (kipf2016semi) to encode the graph contextual information following (liu2018jointly):


where is the neighbour nodes of in the AMR graph, is the edge type between and , is the gate following (liu2018jointly), represents GCN layer number, and

is the Sigmoid function.

and denote parameters of neural layers in this paper. We take the hidden states of the last GCN layer for each word as the common-space representation , where stands for the common (multimedia) embedding space. For each entity , we obtain its representation by averaging the embeddings of its tokens.

Event and Argument Classifier: We classify each word into event types 666We use BIO tag schema to decide trigger word boundary, i.e., adding prefix B- to the type label to mark the beginning of a trigger, I- for inside, and O for none. and classify each entity into argument role :


We take ground truth text entity mentions as input following ji2008refining during training, and obtain testing entity mentions using a named entity extractor (LinACL2019).

3.3 Image Event Extraction

Image Structured Representation: To obtain image structures similar to AMR graphs, and inspired by situation recognition yatskar2016situation, we represent each image with a situation graph, that is a star-shaped graph as shown in Figure 4, where the central node is labeled as a verb (e.g., destroying), and the neighbor nodes are arguments labeled as , where is a noun (e.g., ship) derived from WordNet synsets miller1995wordnet to indicate the entity type, and indicates the role (e.g., item) played by the entity in the event, based on FrameNet fillmore2003background. We develop two methods to construct situation graphs from images and train them using the imSitu dataset yatskar2016situation as follows.

(1) Object-based Graph: Similar to extracting entities to get candidate arguments, we employ the most similar task in CV, object detection, and obtain the object bounding boxes detected by a Faster R-CNN ren2015faster model trained on Open Images kuznetsova2018open with 600 object types (classes). We employ a VGG-16 CNN (simonyan2014very) to extract visual features of an image and and another VGG-16 to encode the bounding boxes

. Then we apply a Multi-Layer Perceptron (MLP) to predict a verb embedding from

and another MLP to predict a noun embedding for each .

We compare the predicted verb embedding to all verbs in the imSitu taxonomy in order to classify the verb, and similarly compare each predicted noun embedding to all imSitu nouns

which results in probability distributions:

where and are word embeddings initialized with GloVE pennington2014glove. We use another MLP with one hidden layer followed by Softmax () to classify role for each object :

Given verb and role-noun

annotations for an image (from the imSitu corpus), we define the situation loss functions:

(2) Attention-based Graph: State-of-the-art object detection methods only cover a limited set of object types, such as 600 types defined in Open Images. Many salient objects such as bomb, stone and stretcher are not covered in these ontologies. Hence, we propose an open-vocabulary alternative to the object-based graph construction model. To this end, we construct a role-driven attention graph, where each argument node is derived by a spatially distributed attention (heatmap) conditioned on a role . More specifically, we use a VGG-16 CNN to extract a convolutional feature map for each image , which can be regarded as attention keys for local regions. Next, for each role defined in the situation recognition ontology (e.g., agent), we build an attention queryvector by concatenating role embedding with the image feature as context and apply a fully connected layer:

Then, we compute the dot product of each query with all keys, followed by Softmax, which forms a heatmap on the image, i.e.,

We use the heatmap to obtain a weighted average of the feature map to represent the argument of each role in the visual space:

Similar to the object-based model, we embed to , compare it to the imSitu noun embeddings to define a distribution, and define a classification loss function. The verb embedding and the verb prediction probability and loss are defined in the same way as in the object-based method.

Event and Argument Classifier: We use either the object-based or attention-based formulation and pre-train it on the imSitu dataset yatskar2016situation. Then we apply a GCN to obtain the structured embedding of each node in the common space, similar to Equation 1. This yields and . We use the same classifiers as defined in Equation 2 to classify each visual event and argument using the common space embedding:


3.4 Cross-Media Joint Training

In order to make the event and argument classifier shared across modalities, the image and text graph should be encoded to the same space. However, it is extremely costly to obtain the parallel text and image event annotation. Hence, we use event and argument annotations in separate modalities (i.e., ACE and imSitu datasets) to train classifiers, and simultaneously use VOA news image and caption pairs to align the two modalities. To this end, we learn to embed the nodes of each image graph close to the nodes of the corresponding caption graph, and far from those in irrelevant caption graphs. Since there is no ground truth alignment between the image nodes and caption nodes, we use image and caption pairs for weakly supervised training, to learn a soft alignment from each words to image objects and vice versa.

where indicates the word in caption sentence and represents the object of image . Then, we compute a weighted average of softly aligned nodes for each node in other modality, i.e.,


We define the alignment cost of the image-caption pair as the Euclidean distance between each node to its aligned representation,

We use a triplet loss to pull relevant image-caption pairs close while pushing irrelevant ones apart:

where is a randomly sampled negative image that does not match . Note that in order to learn the alignment between the image and the trigger word, we treat the image as a special object when learning cross-media alignment.

The common space enables the event and argument classifiers to share weights across modalities, and be trained jointly on the ACE and imSitu datasets, by minimizing the following objective functions:

All tasks are jointly optimized:

Training Model Text-Only Evaluation Image-Only Evaluation Multimedia Evaluation
Event Mention Argument Role Event Mention Argument Role Event Mention Argument Role
  Text JMEE 42.5 58.2 48.7 22.9 28.3 25.3 - - - - - - 42.1 34.6 38.1 21.1 12.6 15.8
GAIL 43.4 53.5 47.9 23.6 29.2 26.1 - - - - - - 44.0 32.4 37.3 22.7 12.8 16.4
WASE 42.3 58.4 48.2 21.4 30.1 24.9 - - - - - - 41.2 33.1 36.7 20.1 13.0 15.7
Image WASEatt - - - - - - 29.7 61.9 40.1 9.1 10.2 9.6 28.3 23.0 25.4 2.9 6.1 3.8
WASEobj - - - - - - 28.6 59.2 38.7 13.3 9.8 11.2 26.1 22.4 24.1 4.7 5.0 4.9
Multimedia VSE-C 33.5 47.8 39.4 16.6 24.7 19.8 30.3 48.9 26.4 5.6 6.1 5.7 33.3 48.2 39.3 11.1 14.9 12.8
Flatatt 34.2 63.2 44.4 20.1 27.1 23.1 27.1 57.3 36.7 4.3 8.9 5.8 33.9 59.8 42.2 12.9 17.6 14.9
Flatobj 38.3 57.9 46.1 21.8 26.6 24.0 26.4 55.8 35.8 9.1 6.5 7.6 34.1 56.4 42.5 16.3 15.9 16.1
WASEatt 37.6 66.8 48.1 27.5 33.2 30.1 32.3 63.4 42.8 9.7 11.1 10.3 38.2 67.1 49.1 18.6 21.6 19.9
WASEobj 42.8 61.9 50.6 23.5 30.3 26.4 43.1 59.2 49.9 14.5 10.1 11.9 43.0 62.1 50.8 19.5 18.9 19.2
Table 3: Event and argument extraction results (%). We compare three categories of baselines in three evaluation settings. The main contribution of the paper is joint training and joint inference on multimedia data (bottom right).

3.5 Cross-Media Joint Inference

In the test phase, our method takes a multimedia document with sentences and images as input. We first generate the structured common embedding for each sentence and each image, and then compute pairwise similarities . We pair each sentence with the closest image , and aggregate the features of each word of with the aligned representation from by weighted averaging:


where and is derived from using Equation 4. We use to classify each word into an event type and to classify each entity into a role with multimedia classifiers in Equation LABEL:eq:text_classifier. To this end, we define similar to but using and . Similarly, for each image we find the closest sentence , compute the aggregated multimedia features and , and feed into the shared classifiers (Equation 3) to predict visual event and argument roles. Finally, we corefer the cross-media events of the same event type if the similarity is higher than a threshold.

4 Experiments

4.1 Evaluation Setting

Evaluation Metrics We conduct evaluation on text-only, image-only, and multimedia event mentions in M2E2 dataset in Section 2.2. We adopt the traditional event extraction measures, i.e., Precision, Recall and F1. For text-only event mentions, we follow ji2008refining; li2013joint: a textual event mention is correct if its event type and trigger offsets match a reference trigger; and a textual event argument is correct if its event type, offsets, and role label match a reference argument. We make a similar definition for image-only event mentions: a visual event mention is correct if its event type and image match a reference visual event mention; and a visual event argument is correct if its event type, localization, and role label match a reference argument. A visual argument is correctly localized if the Intersection over Union (IoU) of the predicted bounding box with the ground truth bounding box is over 0.5. Finally, we define a multimedia event mention to be correct if its event type and trigger offsets (or the image) match the reference trigger (or the reference image). The arguments of multimedia events are either textual or visual arguments, and are evaluated accordingly. To generate bounding boxes for the attention-based model, we threshold the heatmap using the adaptive value of , where is the peak value of the heatmap. Then we compute the tightest bounding box that encloses all of the thresholded region. Examples are shown in Figure 7 and Figure 8.

Baselines The baselines include: (1) Text-only models: We use the state-of-the-art model JMEE (liu2018jointly) and GAIL Zhang2019 for comparison. We also evaluate the effectiveness of cross media joint training by including a version of our model trained only on ACE, denoted as WASE. (2) Image-only models: Since we are the first to extract newsworthy events, and the most similar work situation recognition can not localize arguments in images, we use our model trained only on image corpus as baselines. Our visual branch has two versions, object-based and attention-based, denoted as WASEobj and WASEatt. (3) Multimedia models: To show the effectiveness of structured embedding, we include a baseline by removing the text and image GCNs from our model, which is denoted as Flat. The Flat baseline ignores edges and treats images and sentences as sets of vectors. We also compare to the state-of-the-art cross-media common representation model, Contrastive Visual Semantic Embedding VSE-C shi2018learning, by training it the same way as WASE.

Parameter Settings The common space dimension is . The dimension is for image position embedding and feature map, and for word position embedding, entity type embedding, and POS tag embedding. The layer of GCN is .

4.2 Quantitative Performance

As shown in Table 3, our complete methods (WASEatt and WASEobj) outperform all baselines in the three evaluation settings in terms of F1. The comparison with other multimedia models demonstrates the effectiveness of our model architecture and training strategy. The advantage of structured embedding is shown by the better performance over the flat baseline. Our model outperforms its text-only and image-only variants on multimedia events, showing the inadequacy of single-modal information for complex news understanding. Furthermore, our model achieves better performance on text-only and image-only events, which demonstrates the effectiveness of multimedia training framework in knowledge transfer between modalities.

WASEobj and WASEatt, are both superior to the state of the art and each has its own advantages. WASEobj predicts more accurate bounding boxes since it is based on a Faster R-CNN pretrained on bounding box annotations, resulting in a higher argument precision. While WASEatt achieves a higher argument recall as it is not limited by the predefined object classes of the Faster R-CNN.

Model (%) (%) (%)
rule_based 10.1 100 18.2
VSE 31.2 74.5 44.0
Flatatt 33.1 73.5 45.6
Flatobj 34.3 76.4 47.3
WASEatt 39.5 73.5 51.5
WASEobj 40.1 75.4 52.4
Table 4: Cross-media event coreference performance.

Furthermore, to evaluate the cross-media event coreference performance, we pair textual and visual event mentions in the same document, and calculate Precision, Recall and F1 to compare with ground truth event mention pairs777We do not use coreference clustering metrics because we only focus on mention-level cross-media event coreference instead of the full coreference in all documents.. As shown in Table 4, WASEobj outperforms all multimedia embedding models, as well as the rule-based baseline using event type matching. This demonstrates the effectiveness of our cross-media soft alignment.

4.3 Qualitative Analysis

Our cross-media joint training approach successfully boosts both event extraction and argument role labeling performance. For example, in Figure 5 (a), the text-only model can not extract Justice.Arrest event, but the joint model can use the image as background to detect the event type. In Figure 5 (b), the image-only model detects the image as Conflict.Demonstration, but the sentences in the same document help our model not to label it as Conflict.Demonstration. Compared with multimedia flat embedding in Figure 6, WASE can learn structures such as Artifact is on top of Vehicle, and the person in the middle of Justice.Arrest is Entity instead of Agent.

Figure 5: Image helps textual event extraction, and surrounding sentence helps visual event extraction.
Figure 6: Comparison with multimedia flat embedding.

4.4 Remaining Challenges

One of the biggest challenges in M2E2is localizing arguments in images. Object-based models suffer from the limited object types. Attention-based method is not able to precisely localize the objects for each argument, since there is no supervision on attention extraction during training. For example, in Figure 7, the Entity argument in the Conflict.Demonstrate event is correctly predicted as troops, but its localization is incorrect because Place argument share similar attention. When one argument targets at too many instances, attention heatmaps tend to lose focus and cover the whole image, as shown in Figure 8.

Figure 7: Argument labeling error examples: correct entity name but wrong localization.
Figure 8: Attention heatmaps lose focus due to large instance candidate number.

5 Related Work

Text Event Extraction Text event extraction has been extensively studied for general news domain ji2008refining; liao2011acquiring; huang2012bootstrapped; li2013joint; chen2015event; nguyen2016joint; P18-1048; D18-1156; D18-1158; Zhang2019; liu2018jointly; wang2019open; yang2019exploring; wadden2019entity. Multimedia features has been proven to effectively improve text event extraction (zhang2017improving).

Visual Event Extraction “Events” in NLP usually refer to complex events that involve multiple entities in a large span of time (e.g. protest), while in CV chang2016bi; zhang2007semantic; ma2017joint events are less complex single-entity activities (e.g. washing dishes) or actions (e.g. jumping). Visual event ontologies focus on daily life domains, such as “dogshow” and “wedding ceremony” (perera2012trecvid). Moreover, most efforts ignore the structure of events including arguments. There are a few methods that aim to localize the agent gu2018ava; li2018recurrent; duarte2018videocapsulenet, or classify the recipient sigurdsson2016hollywood; kato2018compositional; wu2019long of events, but neither detects the complete set of arguments for an event. The most similar to our work is Situation Recognition (SR) (yatskar2016situation; mallya2017recurrent) which predicts an event and multiple arguments from an input image, but does not localize the arguments. We use SR as an auxiliary task for training our visual branch, but exploit object detection and attention to enable localization of arguments. silberer2018grounding redefine the problem of visual argument role labeling with event types and bounding boxes as input. Different from their work, we extend the problem scope to including event identification and coreference, and further advance argument localization by proposing an attention framework which does not require bounding boxes for training nor testing.

Multimedia Representation Multimedia common representation has attracted much attention recently toselli2007viterbi; weegar2015linking; hewitt2018learning; chen2019uniter; liu2019focus; Su_2019_ICCV; Sarafianos_2019_ICCV; sun2019videobert; tan2019lxmert; li2019unicoder; li2019visualbert; lu2019vilbert; sun2019contrastive; rahman2019m; su2019vl. However, previous methods focus on aligning images with their captions, or regions with words and entities, but ignore structure and semantic roles. UniVSE Wu2019UniVSERV incorporates entity attributes and relations into cross-media alignment, but does not capture graph-level structures of images or text.

6 Conclusions and Future Work

In this paper we propose a new task of multimedia event extraction and setup a new benchmark. We also develop a novel multimedia structured common space construction method to take advantage of the existing image-caption pairs and single-modal annotated data for weakly supervised training. Experiments demonstrate its effectiveness as a new step towards semantic understanding of events in multimedia data. In the future, we aim to extend our framework to extract events from videos, and make it scalable to new event types. We plan to expand our annotations by including event types from other text event ontologies, as well as new event types not in existing text ontologies. We will also apply our extraction results to downstream applications including cross-media event inference, timeline generation, etc.


This research is based upon work supported in part by U.S. DARPA AIDA Program No. FA8750-18-2-0014 and U.S. DARPA KAIROS Program No. FA8750-19-2-1004. The views and conclusions contained herein are those of the authors and should not be interpreted as necessarily representing the official policies, either expressed or implied, of DARPA, or the U.S. Government. The U.S. Government is authorized to reproduce and distribute reprints for governmental purposes notwithstanding any copyright annotation therein.