Making inferences in incomplete Bayesian networks: A Dempster-Shafer belief function approach
View/ Open
Issue Date
2023-09Author
Shenoy, Prakash P.
Publisher
Elsevier
Type
Article
Article Version
Scholarly/refereed, publisher version
Rights
© 2023 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license.
Metadata
Show full item recordAbstract
How do you make inferences from a Bayesian network (BN) model with missing information? For example, we may not have priors for some variables or may not have conditionals for some states of the parent variables. It is well-known that the Dempster-Shafer (D-S) belief function theory is a generalization of probability theory. So, a solution is to embed an incomplete BN model in a D-S belief function model, omit the missing data, and then make inferences from the belief function model. We will demonstrate this using an implementation of a local computation algorithm for D-S belief function models called the “Belief function machine.” One advantage of this approach is that we get interval estimates of the probabilities of interest. Using Laplacian (equally likely) or maximum entropy priors or conditionals for missing data in a BN may lead to point estimates for the probabilities of interest, masking the uncertainty in these estimates. Bayesian reasoning cannot reason from an incomplete model. A Bayesian sensitivity analysis of the missing parameters is not a substitute for a belief-function analysis.
Collections
Citation
Shenoy, P.P., (2023), Making inferences in incomplete Bayesian networks: A Dempster-Shafer belief function approach, International Journal of Approximate Reasoning, vol. 160, 108967, https://doi.org/10.1016/j.ijar.2023.108967
Items in KU ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
We want to hear from you! Please share your stories about how Open Access to this item benefits YOU.