Statistical inference of random graphs with a surrogate likelihood function

by   Dingbo Wu, et al.

Spectral estimators have been broadly applied to statistical network analysis but they do not incorporate the likelihood information of the network sampling model. This paper proposes a novel surrogate likelihood function for statistical inference of a class of popular network models referred to as random dot product graphs. In contrast to the structurally complicated exact likelihood function, the surrogate likelihood function has a separable structure and is log-concave yet approximates the exact likelihood function well. From the frequentist perspective, we study the maximum surrogate likelihood estimator and establish the accompanying theory. We show its existence, uniqueness, large sample properties, and that it improves upon the baseline spectral estimator with a smaller sum of squared errors. A computationally convenient stochastic gradient descent algorithm is designed for finding the maximum surrogate likelihood estimator in practice. From the Bayesian perspective, we establish the Bernstein–von Mises theorem of the posterior distribution with the surrogate likelihood function and show that the resulting credible sets have the correct frequentist coverage. The empirical performance of the proposed surrogate-likelihood-based methods is validated through the analyses of simulation examples and a real-world Wikipedia graph dataset. An R package implementing the proposed computation algorithms is publicly available at .


page 1

page 2

page 3

page 4


Communication-Efficient Distributed Statistical Inference

We present a Communication-efficient Surrogate Likelihood (CSL) framewor...

Bayesian score calibration for approximate models

Scientists continue to develop increasingly complex mechanistic models t...

Warped Gradient-Enhanced Gaussian Process Surrogate Models for Inference with Intractable Likelihoods

Markov chain Monte Carlo methods for intractable likelihoods, such as th...

A Scheme for Molecular Computation of Maximum Likelihood Estimators for Log-Linear Models

We propose a novel molecular computing scheme for statistical inference....

Bayes Calculations from Quantile Implied Likelihood

A Bayesian model can have a likelihood function that is analytically or ...

Eigenvector-Assisted Statistical Inference for Signal-Plus-Noise Matrix Models

In this paper, we develop a generalized Bayesian inference framework for...

Using Large Language Model Annotations for Valid Downstream Statistical Inference in Social Science: Design-Based Semi-Supervised Learning

In computational social science (CSS), researchers analyze documents to ...

Please sign up or login with your details

Forgot password? Click here to reset