Mitigating Performance Saturation in Neural Marked Point Processes: Architectures and Loss Functions

07/07/2021
by   Tianbo Li, et al.
0

Attributed event sequences are commonly encountered in practice. A recent research line focuses on incorporating neural networks with the statistical model – marked point processes, which is the conventional tool for dealing with attributed event sequences. Neural marked point processes possess good interpretability of probabilistic models as well as the representational power of neural networks. However, we find that performance of neural marked point processes is not always increasing as the network architecture becomes more complicated and larger, which is what we call the performance saturation phenomenon. This is due to the fact that the generalization error of neural marked point processes is determined by both the network representational ability and the model specification at the same time. Therefore we can draw two major conclusions: first, simple network structures can perform no worse than complicated ones for some cases; second, using a proper probabilistic assumption is as equally, if not more, important as improving the complexity of the network. Based on this observation, we propose a simple graph-based network structure called GCHP, which utilizes only graph convolutional layers, thus it can be easily accelerated by the parallel mechanism. We directly consider the distribution of interarrival times instead of imposing a specific assumption on the conditional intensity function, and propose to use a likelihood ratio loss with a moment matching mechanism for optimization and model selection. Experimental results show that GCHP can significantly reduce training time and the likelihood ratio loss with interarrival time probability assumptions can greatly improve the model performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/21/2018

Point processes on directed linear network

In this paper we consider point processes specified on directed linear n...
research
06/01/2018

Lecture Notes: Temporal Point Processes and the Conditional Intensity Function

These short lecture notes contain a not too technical introduction to po...
research
05/23/2019

Fully Neural Network based Model for General Temporal Point Processes

A temporal point process is a mathematical model for a time series of di...
research
09/26/2019

Intensity-Free Learning of Temporal Point Processes

Temporal point processes are the dominant paradigm for modeling sequence...
research
07/15/2019

Controlling Model Complexity in Probabilistic Model-Based Dynamic Optimization of Neural Network Structures

A method of simultaneously optimizing both the structure of neural netwo...
research
06/30/2022

Modularity Optimization as a Training Criterion for Graph Neural Networks

Graph convolution is a recent scalable method for performing deep featur...
research
05/30/2019

Leveraging Simple Model Predictions for Enhancing its Performance

There has been recent interest in improving performance of simple models...

Please sign up or login with your details

Forgot password? Click here to reset