Generalized Bayesian Inference for Regression-Type Models with an Intractable Normalizing Constant

Abstract

Regression models with intractable normalizing constants are valuable tools for analyzing complex data structures, yet parameter inference for such models remains highly challenging—particularly when observations are discrete. In statistical inference, discrete state spaces introduce significant computational difficulties, as the normalizing constant often requires summation over extremely large or even infinite sets, which is typically infeasible in practice. These challenges are further compounded when observations are independent but not identically distributed. This paper addresses these issues by developing a novel generalized Bayesian inference approach tailored for regression models with intractable likelihoods. The key idea is to employ a specific form of generalized Fisher divergence to update beliefs about the model parameters, thereby circumventing the need to compute the normalizing constant. The resulting generalized posterior distribution can be sampled using standard computational tools, such as Markov Chain Monte Carlo (MCMC), effectively avoiding the intractability of the normalizing constant.

Share and Cite:

Gan, Q.Q. and Ye, W.Z. (2025) Generalized Bayesian Inference for Regression-Type Models with an Intractable Normalizing Constant. Advances in Pure Mathematics, 15, 319-338. doi: 10.4236/apm.2025.155016.

1. Introduction

In statistics, many models possess an intractable normalizing constant [1]. Inference for these models is complicated, because the normalizing functions of their probability distributions include the parameters of interest. In Bayesian analysis, they result in so-called doubly intractable posterior distributions which pose significant computational challenges [2].

In Bayesian inference, the posterior distribution of parameters θ given observed data y is given by:

p( θ|y )= p( y|θ )p( θ ) p( y ) , (1)

where p( y )= p( y|θ )p( θ )dθ is the marginal likelihood. The posterior is intractable when p( y ) cannot be computed analytically. The problem becomes doubly intractable when the likelihood itself contains an unknown normalizing constant:

p( y|θ )= f( y,θ ) Z( θ ) ,Z( θ )= f( y,θ )dy . (2)

The doubly intractable property introduces two fundamental challenges: the failure of standard Markov Chain Monte Carlo (MCMC) methods and the curse of dimensionality. To address these challenges, researchers have proposed a variety of strategies, including exact methods based on auxiliary variables (e.g., Exchange Algorithm [3] [4]), approximate methods (e.g., Approximate Bayesian Computation [5]), and likelihood-free alternatives (e.g., Noise Contrastive Estimation [6]). However, these approaches often suffer from limitations, such as high computational cost or restrictive model assumptions.

KSD-Bayes [7] is the first generalized Bayesian inference method proposed for handling models with intractable likelihoods. It constructs a generalized posterior distribution by minimising a loss function known as the Kernel Stein Discrepancy (KSD). This approach avoids the need to explicitly compute the normalising constant, offers robustness to model misspecification, and, under certain conditions, allows for either standard MCMC sampling or closed-form solutions. In short, KSD-Bayes serves as a more flexible and robust alternative to traditional Bayesian inference. However, the generalised posterior in KSD-Bayes is dependent on a user-specified kernel function, which poses challenges in discrete domains. In such domains, natural choices of kernel are often lacking, and even when available, they can be computationally expensive. To address this issue, DFD-Bayes [8] was developed. DFD-Bayes is a generalised Bayesian inference approach based on discrete Fisher divergence, which eliminates the need for user-specified kernels and is specifically designed for intractable likelihood problems in discrete settings. However, DFD-Bayes is currently applicable only to independent and identically distributed (i.i.d.) data and does not extend to regression models. Consequently, there is a pressing need to relax the i.i.d. assumption and develop generalised Bayesian inference methods suitable for regression models. Motivated by this gap, the aim of this paper is to propose a novel generalised Bayesian inference framework for regression models with intractable normalising constants, applicable to independent but not necessarily identically distributed (INID) discrete data.

The rest of this paper is organized as following: in Section 2, we provide background of this work. Section 3 introduces the specific model structure and presents the DSFD-Bayes methodology. Section 4 presents Monte Carlo studies and case studies on real datasets, demonstrating the effectiveness of our approach. Finally, Section 5 concludes the paper with discussion.

2. Background

The aim of this section is to briefly review Generalized Bayesian Inference.

2.1. Limitations of Standard Bayesian Inference

Standard Bayesian inference, as a core paradigm of probabilistic modeling, relies critically on three fundamental assumptions: 1) correct specification of the parametric likelihood function; 2) appropriate choice of prior distribution; and 3) rigorous application of Bayes’ theorem. However, modern statistical modeling practice demonstrates that these assumptions face significant challenges in complex data analysis scenarios.

First, in the case of model misspecification (M-open), when the true data-generating mechanism f 0 ( x ) is not within the assumed model family { p( x|θ ) } θΘ , posterior inferences may become unreliable [9]. This limitation is particularly pronounced in low-dimensional parameter inference.

Second, for models containing latent variables or complex dependency structures, the likelihood function p( x|θ ) often involves high-dimensional integration, making the normalization constant Z( θ ) analytically intractable. Even when using proxy models for approximation, parameter estimation through the KL divergence minimization remains constrained by model specification accuracy.

2.2. Generalized Bayesian Inference

Building upon these motivations, Bissiri et al. [10] proposed focusing inference exclusively on the target quantities of interest, rather than requiring full specification of the data-generating process. They produce what they refer to as a general Bayesian update—a coherent method to produce a posterior distribution over some quantity without relying on a full model for the observations, when considering the non-inferential Bayesian decision problem [11]. Given a prior distribution π( θ ) , an observed dataset { y i } i=1 n , and a positive scaling parameter w>0 , the generalized Bayesian framework replaces the likelihood with a loss function D n ( θ ) to define the posterior update, resulting in the generalized posterior:

π n D ( θ )π( θ )exp( w D n ( θ ) ). (3)

In this formulation, D n ( θ ) represents a loss function quantifying the discrepancy between the parameter θ and the observed data, while w serves as a calibration parameter governing the influence of the loss. Notably, when the loss function is chosen as the negative log-likelihood, i.e., D n ( θ )= i=1 n logp( y i |θ ) , the generalized posterior reduces to the standard Bayesian posterior, thereby recovering classical Bayesian inference as a special case.

The generalized Bayesian framework introduces the concept of a loss function into the updating process, offering a fundamentally new perspective for statistical modeling and parameter inference. Compared to the traditional Bayesian paradigm, its primary advantage lies in its remarkable flexibility. Rather than requiring full specification of a probabilistic model, the generalized Bayesian approach establishes a connection between parameters and observed data through a carefully designed loss function. This feature makes it particularly well-suited for applications where the true data-generating mechanism is unknown or excessively complex. In practice, researchers can tailor the choice of loss function to the inferential objective—for example, employing the squared loss for mean estimation or the absolute error loss for median estimation—thus allowing more precise targeting of specific quantities of interest.

Importantly, the generalized Bayesian framework exhibits strong robustness properties. Traditional Bayesian inference can suffer from substantial bias under model misspecification; in contrast, the generalized approach mitigates this issue by directly targeting the loss associated with the parameter of interest, thereby reducing reliance on the correctness of the global model specification. This robustness naturally accommodates a wide range of robust statistical methods, such as the use of the Huber loss for outlier-resistant estimation. Furthermore, the framework demonstrates powerful information integration capabilities: it is capable of incorporating not only traditional random samples but also non-random information sources, such as expert knowledge, and can effectively handle special settings like partial likelihoods commonly encountered in survival analysis.

Nevertheless, the application of the generalized Bayesian framework demands careful attention to several critical aspects. First, the choice and calibration of the loss function are of central importance, as they directly influence the reliability of the resulting inference. Researchers must thoughtfully select an appropriate loss form tailored to the specific context and cautiously determine any associated calibration parameters. Second, computational challenges can arise, particularly because the generalized posterior often lacks a closed-form expression. In such cases, inference typically relies on numerical methods such as Markov chain Monte Carlo (MCMC), and the computational burden may become substantial in high-dimensional parameter spaces.

3. Methodology

In this section, we first present a general framework for probabilistic regression-type models with an intractable normalizing constant and then present Discrete Slope Fisher Divergence. Finally, we will propose Discrete Slope Fisher Divergence. All random variables in this paper are defined on the foundational probability space ( Ω,, ) . For instance, the expectation of a random variable y is expressed as E( y )= Ω yd .

3.1. Regression-Type Models with an Intractable Normalizing Constant on a Discrete Domain

We present a specialized framework for probabilistic regression-type models with an intractable normalizing constant defined exclusively on discrete domains. Suppose we have independent observations y 1 ,, y n D from unknown distributions q( y| x 1 ),,q( y| x n ) , where D is a discrete domain (e.g., D= { 0,1 } d for binary data or D= { 0,1,2, } d for count data), with associated covariates x 1 ,, x n . We consider parametric models p( | x i ,θ ) for i=1,,n , where θ p is a vector of unknown parameters. Under the fixed design setting, the observations y 1 ,, y n are independent but not necessarily identically distributed.

Our discrete-domain framework for regression-type models with an intractable normalizing constant is given by:

p( y| x i ,θ )= 1 Z i ( θ ) p 0 ( y ) p ˜ 1 ( y| x i ,θ ), (4)

where:

p 0 ( y ) is a known function of y that encodes domain-specific information (e.g., boundary constraints or structural zeros);

p ˜ 1 ( y| x i ,θ ) models the relationship between the discrete response and covariates;

Z i ( θ ) is the normalizing constant (partition function) defined as:

Z i ( θ )= yD p 0 ( y ) p ˜ 1 ( y| x i ,θ ). (5)

If the model (4) is correctly specified, the true data-generating distribution satisfies q( y| x i ) p 0 ( y ) p ˜ 1 ( y| x i , θ 0 ) for i=1,,n , where θ 0 represents the true parameter vector.

This framework provides the foundation for developing and analyzing regression models for discrete data with intractable normalizing constants, where the computational challenges of normalization are inherent to the discrete nature of the domain.

Example (Conway-Maxwell-Poisson regression model [12]). Let y 0 , where 0 is the set of non-negative integers, with p 0 ( y )=1 and

p ˜ 1 ( y| x i ,θ )= λ i y ( y! ) ν ,

where θ= ( β ,ν ) , ν0 represents the dispersion parameter, and λ i =exp( x i β ) generalizes the Poisson mean parameter. Then model (4) yields the CMP regression with normalizing constant:

Z i ( θ )= s=0 λ i s ( s! ) ν .

The CMP distribution encompasses three important special cases:

Geometric: When ν=0 (with λ i <1 )

Poisson: When ν=1

Bernoulli: In the limit ν

For ν[ 0,1 ) , the CMP describes data that are overdispersed relative to a Poisson distribution with the same mean; For ν>1 , the CMP model is appropriate for underdispersed data.

3.2. A Discrete Slope Fisher Divergence

Fisher divergence, originally proposed for score matching [13], measures the discrepancy between two probability distributions by comparing their score functions (gradients of log-densities). However, its reliance on gradient operators limits applications to continuous and differentiable models. To overcome this limitation, Lyus [14] introduced the generalized Fisher divergence, which extends the original framework by replacing gradients with general linear operators, enabling broader applicability—including discrete data and non-gradient-based models.

Based on the Fisher divergence for data-generating distribution q( y| x i ) and parametric model distribution p( y| x i ,θ ) on d , the divergence D ( q * p * ) is defined as:

D ( q * p * )= 1 n i=1 n E[ logq( y i | x i )logp( y i | x i ,θ ) 2 ], (6)

where is the gradient operator with respect to y , E represents the expectation under the true distribution q( | x i ) .

Based on the main idea from Lyu (2012), the generalized Fisher divergence D ( q * p * ) for q * = i=1 n q( y| x i ) and p * = i=1 n p( y| x i ,θ ) is defined as:

D ( q * p * )= 1 n i=1 n E( q( y| x i ) q( y| x i ) p( y| x i ,θ ) p( y| x i ,θ ) 2 ) (7)

where denotes the Euclidean norm.

To handle discrete data, Lyu investigated a special case of the linear operator in Equation (7), called the marginalization operator . This operator is defined as:

q( y| x i ):= ( 1 q( y| x i ),, d q( y| x i ) )

where each component operator is given by:

j q( y| x i ):= y j q( y| x i )fori=1,,nandj=1,,d.

However, in the one-dimensional discrete case, this specific linear operator (namely, the marginalization operator ) fails to eliminate the influence of the normalization constant. To address this limitation, [15] proposed a new linear operator called the forward difference operator.

The traditional Fisher divergence compares the slopes of log-density functions. However, since discrete data density functions lack continuous slopes at certain points, directly comparing the log-gradients of density functions may not be appropriate for discrete data. Therefore, [15] considered a new linear operator defined as:

p( y| x i ,θ ):=p( y + | x i ,θ )p( y| x i ,θ ),i=1,,n,

where y + =y+1 . This linear operator represents the “discrete slope” of p( y| x i ,θ ) at point y , which is the discrete counterpart of the log-density gradient in continuous distributions. When y takes values from a finite set, special treatment is required at boundary points. Specifically, when y is at the upper boundary, define p( y + | x i ,θ )=0 . Similarly, define p( y | x i ,θ )=0 for y =y1 at the lower boundary.

Notably, this linear operator provides the “slope” of p( y| x i ,θ ) at point y in discrete settings. We can derive (ignoring constant terms):

p( y| x i ,θ ) p( y| x i ,θ ) = p( y+1| x i ,θ ) p( y| x i ,θ ) 1

This shows that the fundamental principle of this method is to make the ratio p( y + | x i ,θ ) p( y| x i ,θ ) as close as possible to the corresponding ratio from the data distribution, i.e., q( y + | x i ) q( y| x i ) .

To avoid division-by-zero problems when computing slopes, introducing the slope transformation proposed by [16]:

t( u )= 1 1+u

Under this transformation, if the probability density approaches zero causing the ratio to diverge to infinity, the transformed value becomes t( )=0 .

Therefore, a new generalized Fisher divergence between q * and p * is defined as:

D DSF ( q * p * )= 1 n i=1 n E{ [ t( p( y i + | x i ,θ ) p( y i | x i ,θ ) )t( q( y i + | x i ) q( y i | x i ) ) ] 2 + [ t( p( y i | x i ,θ ) p( y i | x i ,θ ) )t( q( y i | x i ) q( y i | x i ) ) ] 2 }, (8)

where, q i and p i denote q( y| x i ) and q( y| x i ,θ ) , respectively. In this paper, we refer to this divergence as the Discrete Slope Fisher Divergence.

To extend to the multivariate case, the linear operator is defined as:

p( y| x i ,θ )=( j p( y| x i ,θ ) )=( p( y ( j+ ) | x i ,θ )p( y| x i ,θ ) ),

where for j=1,,d , we define:

y ( j+ ) = ( y 1 ,, y j +1,, y d ) .

If y has bounded support, when y is at the boundary, we set:

p( y ( j+ ) | x i ,θ )=0.

Similarly, we define:

y ( j ) = ( y 1 ,, y j 1,, y d ) ,

and at the boundary of the response domain, we set:

p( y ( j ) | x i ,θ )=0.

Analogous to the univariate case, the Discrete Slope Fisher Divergence between q * and p * is defined as:

D DSF ( q * p * )= 1 n i=1 n D DSF ( q i p i ) (9)

where

D DSF ( q i p i )=E [ j=1 d { [ t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) )t( q( y i ( j+ ) | x i ) q( y i | x i ) ) ] 2 + [ t( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) )t( q( y i | x i ) q( y i ( j ) | x i ) ) ] 2 } ] (10)

Theorem 1 Equation (10) can be decomposed as:

D DSF ( q i p i )=E [ j=1 d { t ( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) 2 +t ( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) 2 2t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) } ]+C (11)

where C is a constant independent of the parameters θ .

This theorem shows that (10) is tractable. Its proof is given in the Appendix. The discrete Slope Fisher divergence between a model p θ and an empirical distribution q n = 1 n i=1 n δ y i , is computed as

D DSF ( q n p θ )= 1 n i=1 n j=1 d { t ( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) 2 +t ( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) 2 2t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) }+C( q n ), (12)

where C( q n ) is a constant depending on the empirical density p n but not on θ .

3.3. A Generalised Posterior

We are now in a position to present DSFD-Bayes.

Given a prior distribution π over the parameter space Θ and a statistical model p θ as defined in Equation (4) on a discrete domain—a regression-type model with an intractable normalization constant parameterized by θΘ —where the observed response variables are independent but not necessarily identically distributed, the DSFD-Bayes posterior distribution is defined as:

π n D ( θ )π( θ )exp( wn D DSF ( q n p θ ) ), (13)

where w>0 is a tuning constant to be specified (see Section 3.3 in [8] for guidance on selecting an appropriate weighting factor w ). The θ -independent constant term C( q n ) in D DSF ( q n p θ ) cancels out during the normalization of the DSFD-Bayes posterior. Hence, C( q n ) can be safely ignored in practical computations.

The DSFD-Bayes posterior is directly amenable to standard posterior sampling techniques, such as Markov chain Monte Carlo (MCMC) and the No-U-Turn Sampler (NUTS), as the actual computation does not involve any intractable normalizing constants. The resulting DSFD-Bayes method can be computed at cost O( nd ) linear in the size of the dataset.

4. Experiments

This paper validates the proposed method through two simulation studies—one overdispersed and one underdispersed—and presents a data analysis case involving underdispersion. All analyses were conducted on a HUAWEI computer equipped with an Intel Core i9-12900H processor (2.5 GHz) and 16 GB of RAM, running Windows 11 and Python 3.10.16.

We investigate a relatively simple model—the Conway—Maxwell—Poisson (CMP) regression model, and the CMP regression model was chosen as our experimental framework based on the following considerations: With its single dispersion parameter ν , the CMP model generalizes both Poisson and Bernoulli regression characteristics. This “parsimonious yet expressive” property makes it an ideal testbed that captures the essential complexity of real-world statistical models while preventing over-parameterization from confounding computational efficiency evaluations.

4.1. Simulation Studies

We simulated two datasets from the Conway-Maxwell-Poisson (CMP) regression model: 1) underdispersion with dispersion parameter v=1.5 ; 2) overdispersion with v=0.8 . In both simulation settings, the true regression coefficients were set to β=( 0.5,1.2,0.8 ) , where 0.5 is the intercept term.

This study employs the Conway-Maxwell-Poisson (CMP) regression model to generate simulated data, evaluating model performance through carefully controlled generation processes for both covariates and response variables. The covariate generation is meticulously designed to simulate real-world data scenarios, where the explanatory variable X 1 is sampled from the standard normal distribution N( 0,1 ) with 1000 samples, and X 2 is drawn from the normal distribution N( 1,0.5 ) with the same sample size. To ensure experimental reproducibility, we set the random seed to 1 and include an intercept column vector 1 , ultimately constructing the design matrix X=[ 1, X 1 , X 2 ] . This design maintains the stochastic characteristics of continuous variables while guaranteeing result reproducibility through fixed random seeds.

The response variable generation achieves non-i.i.d. properties through a parameter heterogeneity mechanism. Each response variable Y i corresponds to a unique rate parameter λ i =exp( X i β ) , where the regression coefficients are set as β= ( 0.3,0.5,0.2 ) . Since the CMP distribution lacks an analytical sampling method, we employ Metropolis-Hastings algorithm for MCMC sampling: initial values y ( 0 ) are sampled from Poisson (2) distribution, followed by 2000 burn-in iterations. During this process, we set shared dispersion parameters to make all response variables exhibit either under-dispersion or over-dispersion characteristics.

We compared two methods: the standard Bayesian inference and our proposed DSFD-Bayes method, and the method for calibrating the weight w in DSFD-Bayes is described in Section 3.3.

For the underdispersed case, the prior distribution for the regression coefficients was specified as a multivariate normal distribution N( 0,I ) , and the prior for the dispersion parameter v was set as a log-normal distribution with log-mean 0 and log-standard deviation 0.5. For the overdispersed case, the prior for β was also specified as N( 0,I ) , while the prior for v was chosen as a log-normal distribution with log-mean −0.5 and log-standard deviation 0.5.

We used the No-U-Turn Sampler (NUTS) to sample from all posterior distributions. The specific settings were as follows: 5000 post-warmup samples were retained after discarding 5000 warm-up iterations, and 10 independent chains were run. In all cases, the R ^ statistics for the four parameters were below 1.01, indicating good convergence.

Table 1 and Table 2 present comparative evaluations of DSFD-Bayes versus standard Bayes under under-dispersed ( ν=1.5 ) and over-dispersed ( ν=0.8 ) conditions, respectively. The evaluation results suggest that the DSFD-Bayes generally outperforms the standard Bayes. DSFD-Bayes consistently yields smaller absolute biases and exhibits a reasonable decreasing trend in standard deviations as the sample size increases. Its 95% credible intervals tend to cover the true parameter values in most cases, with more appropriate interval widths. In contrast, the credible intervals produced by the standard Bayesian method are often too narrow and deviate from the true values.

The evaluation results suggest that the DSFD-Bayes method generally outperforms the standard Bayesian approach. DSFD-Bayes consistently yields smaller absolute biases and exhibits a reasonable decreasing trend in standard deviations as the sample size increases. Its 95% credible intervals tend to cover the true parameter values in most cases, with more appropriate interval widths. In contrast, the credible intervals produced by the standard Bayesian method are often too narrow and deviate from the true values.

In terms of applicability, DSFD-Bayes demonstrates greater robustness, performing well across a wide range of sample sizes. The standard Bayesian method, on the other hand, tends to fail in small-sample settings due to the overwhelming influence of the prior, and is only potentially reliable in large-sample cases. These findings indicate that DSFD-Bayes offers a better balance between prior information and data, making it a more reliable choice for Bayesian inference.

Table 1. Comparison of performance metrics for parameter estimation using DSFD-Bayes and Bayes methods for different sample sizes (v = 1.5).

DSFD-Bayes

Bayes

Parameter

Bias

SD

95% CI

Bias

SD

95% CI

n=10

β 0

−0.1656

0.2016

(−0.0619, 0.7429)

14.8625

0.7317

(14.0575, 16.5038)

β 1

0.8602

0.0887

(−0.5151, −0.1643)

0.5365

0.0002

(−0.6639, −0.6633)

β 2

−0.2978

0.1911

(0.1290, 0.8749)

3.6892

0.0001

(4.4889, 4.4893)

ν

−0.7817

0.1216

(0.4900, 0.9620)

6.7955

0.4314

(7.5355, 8.9768)

n=50

β 0

−0.8410

0.0267

(−0.3934, −0.2891)

1.5275

0.0024

(2.0236, 2.0317)

β 1

0.8565

0.0123

(−0.3677, −0.3196)

0.8541

0.0044

(−0.3540, −0.3394)

β 2

−0.5653

0.0210

(0.1935, 0.2765)

−2.4191

0.0266

(−1.6665, −1.5775)

ν

−1.3282

0.0119

(0.1487, 0.1953)

−1.1706

0.0016

(0.3269, 0.3323)

n=100

β 0

−2.8805

0.7161

(−3.7940, −0.9872)

1.3351

0.0004

(1.8343, 1.8358)

β 1

1.1286

0.8506

(−1.9555, 1.4455)

0.8484

0.0005

(−0.3524, −0.3508)

β 2

−2.2036

0.7836

(−2.9597, 0.1096)

−2.6799

0.0027

(−1.8849, −1.8754)

ν

0.2217

0.9582

(0.5191, 4.2530)

−0.9369

0.0003

(0.5626, 0.5637)

n=150

β 0

−0.2400

0.2973

(−0.3199, 0.8321)

1.3845

0.0003

(1.8840, 1.8850)

β 1

0.8605

0.1507

(−0.6516, −0.0523)

0.8918

0.0007

(−0.3093, −0.3072)

β 2

−0.5587

0.2507

(−0.2422, 0.7509)

−2.6532

0.0008

(−1.8547, −1.8519)

ν

−0.9300

0.1550

(0.3002, 0.9041)

−1.0164

0.0002

(0.4833, 0.4840)

n=200

β 0

−0.2508

0.5806

(−0.8783, 1.3975)

1.4505

0.0007

(1.9495, 1.9515)

β 1

0.7173

0.4130

(−1.3580, 0.2573)

0.8979

0.0003

(−0.3028, −0.3017)

β 2

−0.4258

0.5417

(−0.6746, 1.4571)

−2.5694

0.0020

(−1.7727, −1.7660)

ν

−0.7506

0.2855

(0.3182, 1.4075)

−0.9134

0.0003

(0.5862, 0.5871)

Note: SD = Standard Deviation of the Estimate; CI = Credible Interval.

Table 2. Comparison of performance metrics for parameter estimation using DSFD-Bayes and Bayes methods for different sample sizes (v = 0.8).

DSFD-Bayes

Bayes

Parameter

Bias

SD

95% CI

Bias

SD

95% CI

n=10

β 0

−0.0396

0.4317

(−0.3697, 1.3089)

1.4959

0.0463

(1.9087, 2.0707)

β 1

−0.0690

0.2873

(−1.8351, −0.7181)

1.3508

0.0236

(0.1087, 0.1900)

β 2

−0.0300

0.4924

(−0.1492, 1.7416)

−1.6168

0.1284

(−1.0286, −0.5840)

ν

−0.1300

0.1154

(0.4758, 0.9304)

−0.5851

0.0017

(0.2120, 0.2169)

n=50

β 0

0.0460

0.0218

(0.5027, 0.5884)

1.2085

0.0001

(1.7083, 1.7087)

β 1

−0.0869

0.0093

(−1.3055, −1.2687)

0.8769

0.0002

(−0.3234, −0.3228)

β 2

−0.0411

0.0178

(0.7242, 0.7937)

−2.8052

0.0007

(−2.0063, −2.0041)

ν

0.1360

0.0025

(0.9314, 0.9412)

−0.2629

0.0000

(0.2371, 0.2372)

n=100

β 0

0.0412

0.0399

(0.4631, 0.6202)

1.1342

0.0071

(1.6201, 1.6483)

β 1

−0.0934

0.0195

(−1.3320, −1.2551)

0.8874

0.0021

(−0.3167, −0.3084)

β 2

−0.0139

0.0315

(0.7254, 0.8470)

−2.8845

0.0028

(−2.0897, −2.0803)

ν

0.0055

0.0088

(0.7896, 0.8244)

−0.2736

0.0016

(0.5232, 0.5295)

n=150

β 0

−0.2278

0.0028

(0.2669, 0.2779)

1.1623

0.0035

(1.6560, 1.6684)

β 1

0.0009

0.0012

(−1.2014, −1.1968)

0.8980

0.0016

(−0.3052, −0.2989)

β 2

0.0861

0.0022

(0.8817, 0.8904)

−2.8611

0.0017

(−2.0641, −2.0581)

ν

−0.1277

0.0008

(0.6707, 0.6739)

−0.3707

0.0012

(0.4268, 0.4319)

n=200

β 0

−0.0375

0.0246

(0.4144, 0.5090)

1.1607

0.0029

(1.6589, 1.6660)

β 1

−0.0353

0.0110

(−1.2565, −1.2134)

0.8976

0.0013

(−0.3043, −0.2994)

β 2

−0.0179

0.0216

(0.7406, 0.8249)

−2.8601

0.0013

(−2.0613, −2.0579)

ν

0.0048

0.0047

(0.7962, 0.8146)

−0.3705

0.0010

(0.4275, 0.4311)

Note: SD = Standard Deviation of the Estimate; CI = Credible Interval.

As shown in Figure 1, although DSFD-Bayes incurs slightly higher computational cost than Standard Bayes in small-sample scenarios, its computational time becomes significantly lower as the sample size increases, with a marked difference observed between the two methods. While DSFD-Bayes demonstrates superior computational efficiency in larger datasets, its relatively higher costs in small-sample settings may constrain its practical adoption in resource-constrained environments.

(a) Underdispersed ( v=1.5 )

(b) Overdispersed ( v=0.8 )

Figure 1. Comparison of Computational Time between DSFD-Bayes and Standard Bayes.

4.2. Case Studies

We use the airfreight breakage dataset from [17], which contains 10 batches of air cargo transportation records. The explanatory variable ( X i ) is the number of transfers each batch experienced during transportation (a discrete count), and the response variable ( Y i ) is the number of damaged ampules out of 1000 in each batch (also a discrete count). This dataset exhibits underdispersion.

We model the real dataset using a Conway-Maxwell-Poisson regression model. The regression coefficients β=( β 0 , β 1 ) are assigned a multivariate normal prior N( 0,4I ) , and the dispersion parameter v follows a Gamma( 7,1 ) prior.

Figure 2 shows the estimated dispersion parameter v>1 , consistent with the under-dispersion property of the data. A comparison of the forest plots reveals that DSFD-Bayes yields more concentrated parameter estimates with narrower intervals, indicating greater stability and reliability.

(a) DSFD-Bayes

(b) Bayes

Figure 2. Comparison of posterior parameter estimates between DSFD-Bayes and Standard Bayes (95% Credible Intervals).

To compare the predictive performance of the standard Bayes and the DSFD-Bayes, we conduct a posterior predictive check. Figure 3 and Figure 4 present the posterior predictive distributions of the two methods for the number of defects. As shown in Figure 4 (standard Bayes), while the posterior mean generally follows the trend of the observed data, there is considerable variability, especially in regions with higher observed values. The widened credible intervals in those areas suggest that the standard Bayesian model struggles to capture heteroskedasticity and local structure, resulting in low confidence in predicting extreme values. In contrast, Figure 3 demonstrates that DSFD-Bayes significantly improves predictive performance. Its posterior mean closely aligns with the observed data, and the prediction intervals are narrower and more adaptive across different regions. Particularly in the low-to-moderate response range, the credible intervals remain compact while still covering the true observations, indicating better calibration.

Figure 3. Comparison between observed data and dsfd-bayes posterior predictive estimates.

Figure 4. Comparison between observed data and bayes posterior predictive estimates.

5. Conclusions and Suggestions

This paper proposes a novel generalized Bayesian inference method called DSFD-Bayes for discrete regression models with intractable normalizing constants. The method overcomes the limitations of existing KSD-Bayes and DFD-Bayes approaches, and is applicable to independent but not necessarily identically distributed (INID) discrete data, filling a gap in generalized Bayesian inference for regression modeling. Both simulation studies and empirical analysis demonstrate that the DSFD-Bayes performs better in model adaptability and robustness. However, a limitation of this study is that the proposed method is currently restricted to regression-type models, and the lack of exploration into more diverse model classes—such as time series or hierarchical models—restricts the general applicability of the proposed method. This limitation suggests a promising direction for future research.

Appendix

Proof of Theorem 1. Recall that the transformation function t( f 1 ( y )/ f 2 ( y ) ) returns zero when f 2 ( y )=0 , and returns one when f 1 ( y )=0 . In other words, the transformation t( f 1 ( y )/ f 2 ( y ) ) yields a constant value whenever either f 1 ( y ) or f 2 ( y ) is zero. Therefore, for simplicity, we assume that both q( y| x i ) and p( y| x i ,θ ) are non-zero for all yD . Note that we have

After expanding the squares, reorganizing the summation terms, and separating out the constant term C , we obtain:

D DSF ( q i , p i )= yD q( y i | x i ) j=1 d { t ( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) 2 +t ( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) 2 } 2 yD q( y i | x i ) j=1 d { t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) )t( q( y i ( j+ ) | x i ) q( y i | x i ) ) + t( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) )t( q( y i | x i ) q( y i ( j ) | x i ) ) }+C

where C does not depend on θ . Firstly examining the second term:

yD q( y i | x i ) j=1 d { t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) )t( q( y i ( j+ ) | x i ) q( y i | x i ) ) + t( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) )t( q( y i | x i ) q( y i ( j ) | x i ) ) }

According to the definition of the transformation function t( u )= 1 1+u , we have:

t( q( y i ( j+ ) | x i ) q( y i | x i ) )= q( y i | x i ) q( y i | x i )+q( y i ( j+ ) | x i )

t( q( y i | x i ) q( y i ( j ) | x i ) )= q( y i ( j ) | x i ) q( y i ( j ) | x i )+q( y i | x i )

Substituting into the second term’s summation:

yD q( y i | x i ) j=1 d { t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) q( y i | x i ) q( y i | x i )+q( y i ( j+ ) | x i ) + t( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) q( y i ( j ) | x i ) q( y i ( j ) | x i )+q( y i | x i ) }

This part can be split into two independent summation terms:

yD q( y i | x i ) j=1 d t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) q( y i | x i ) q( y i | x i )+q( y i ( j+ ) | x i ) + yD q( y i ( j+ ) | x i ) j=1 d t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) q( y i | x i ) q( y i | x i )+q( y i ( j+ ) | x i ) = yD q( y i | x i ) j=1 d t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) )

Therefore, returning to the original expression:

D DSF ( q i , p i )= yD q( y i | x i ) j=1 d { t ( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) 2 +t ( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) 2 } 2 yD q( y i | x i ) j=1 d t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) )+C = yD q( y i | x i ) j=1 d { t ( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) 2 +t ( p( y i | x i ,θ ) p( y i ( j ) | x i ,θ ) ) 2

2t( p( y i ( j+ ) | x i ,θ ) p( y i | x i ,θ ) ) }+C

This completes the proof.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Jin, I.H. and Liang, F. (2014) Use of SAMC for Bayesian Analysis of Statistical Models with Intractable Normalizing Constants. Computational Statistics & Data Analysis, 71, 402-416.
https://doi.org/10.1016/j.csda.2012.07.005
[2] Park, J. and Haran, M. (2018) Bayesian Inference in the Presence of Intractable Normalizing Functions. Journal of the American Statistical Association, 113, 1372-1390.
https://doi.org/10.1080/01621459.2018.1448824
[3] Murray, I., Ghahramani, Z. and Mackay, D. (2006) MCMC for Doubly-Intractable Distributions. Proceedings of the 22nd Conference on Uncertainty in Artificial Intelligence, Cambridge, 13-16 July 2006, 359-366.
[4] Møller, J., Pettitt, A.N., Reeves, R. and Berthelsen, K.K. (2006) An Efficient Markov Chain Monte Carlo Method for Distributions with Intractable Normalising Constants. Biometrika, 93, 451-458.
https://doi.org/10.1093/biomet/93.2.451
[5] Marin, J., Pudlo, P., Robert, C.P. and Ryder, R.J. (2011) Approximate Bayesian Computational Methods. Statistics and Computing, 22, 1167-1180.
https://doi.org/10.1007/s11222-011-9288-2
[6] Gutmann, M. and Hyvärinen, A. (2010) Noise-Contrastive Estimation: A New Estimation Principle for Unnormalized Statistical Models. Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, Sardinia, 13-15 May 2010, 297-304.
[7] Matsubara, T., Knoblauch, J., Briol, F. and Oates, C.J. (2022) Robust Generalised Bayesian Inference for Intractable Likelihoods. Journal of the Royal Statistical Society Series B: Statistical Methodology, 84, 997-1022.
https://doi.org/10.1111/rssb.12500
[8] Matsubara, T., Knoblauch, J., Briol, F. and Oates, C.J. (2023) Generalized Bayesian Inference for Discrete Intractable Likelihood. Journal of the American Statistical Association, 119, 2345-2355.
https://doi.org/10.1080/01621459.2023.2257891
[9] Key, J.T., Pericchi, L.R. and Smith, A.F.M. (1999) Bayesian Model Choice: What and Why? In: Bernardo, J.M., et al., Eds., Bayesian Statistics, Vol. 6, Oxford University Press, 343-370.
https://doi.org/10.1093/oso/9780198504856.003.0015
[10] Bissiri, P.G., Holmes, C.C. and Walker, S.G. (2016) A General Framework for Updating Belief Distributions. Journal of the Royal Statistical Society Series B: Statistical Methodology, 78, 1103-1130.
https://doi.org/10.1111/rssb.12158
[11] Jewson, J., Smith, J.Q. and Holmes, C. (2018) Principles of Bayesian Inference Using General Divergence Criteria. Entropy, 20, Article No. 442.
https://doi.org/10.3390/e20060442
[12] Sellers, K.F. and Shmueli, G. (2010) A Flexible Regression Model for Count Data. The Annals of Applied Statistics, 4, 943-961.
[13] Hyvrinen, A. and Dayan, P. (2005) Estimation of Non-Normalized Statistical Models by Score Matching. Journal of Machine Learning Research, 6, 695-709.
[14] Lyu, S. (2012) Interpretation and Generalization of Score Matching. Proceedings of the 25th Conference on Uncertainty in Artificial Intelligence, Montreal, 18-21 June 2009, 359-366.
[15] Xu, J., Scealy, J.L., Wood, A.T. and Zou, T. (2022) Generalized Score Matching for Regression.
[16] Hyvärinen, A. (2007) Some Extensions of Score Matching. Computational Statistics & Data Analysis, 51, 2499-2512.
https://doi.org/10.1016/j.csda.2006.09.003
[17] Kutner, M.H., Nachtsheim, C.J. and Neter, J. (1984) Applied Linear Regression Models. Technometrics, 26, 415-416.
https://doi.org/10.1080/00401706.1984.10487998

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.