Results 11 to 20 of about 151,404 (250)
Improved Local Search with Momentum for Bayesian Networks Structure Learning
Bayesian Networks structure learning (BNSL) is a troublesome problem that aims to search for an optimal structure. An exact search tends to sacrifice a significant amount of time and memory to promote accuracy, while the local search can tackle complex ...
Xiaohan Liu +3 more
doaj +1 more source
Bayesian probabilistic modeling is supported by powerful computational tools like probabilistic programming and efficient Markov Chain Monte Carlo (MCMC) sampling. However, the results of Bayesian inference are challenging for users to interpret in tasks
Evdoxia Taka +2 more
doaj +1 more source
Probabilistic Graphical Models [PDF]
In this chapter, we will briefly summarize the basic concepts of probability as well as graph theory.We start with important terms and definitions from graph theory and emphasize the relation to our hierarchical models. Directed and undirected graphs will be introduced and compared with each other. Then, we will give an overview of the random variables
+5 more sources
Extending Stan for Deep Probabilistic Programming [PDF]
Stan is a popular declarative probabilistic programming language with a high-level syntax for expressing graphical models and beyond. Stan differs by nature from generative probabilistic programming languages like Church, Anglican, or Pyro.
Baudart, Guillaume +5 more
core +2 more sources
Score-based algorithms that learn Bayesian Network (BN) structures provide solutions ranging from different levels of approximate learning to exact learning.
Zhigao Guo, Anthony C. Constantinou
doaj +1 more source
Using Value-Based Potentials for Making Approximate Inference on Probabilistic Graphical Models
The computerization of many everyday tasks generates vast amounts of data, and this has lead to the development of machine-learning methods which are capable of extracting useful information from the data so that the data can be used in future decision ...
Pedro Bonilla-Nadal +4 more
doaj +1 more source
Computation of Kullback–Leibler Divergence in Bayesian Networks
Kullback–Leibler divergence KL(p,q) is the standard measure of error when we have a true probability distribution p which is approximate with probability distribution q.
Serafín Moral +2 more
doaj +1 more source
Bayesian Test of Significance for Conditional Independence: The Multinomial Model
Conditional independence tests have received special attention lately in machine learning and computational intelligence related literature as an important indicator of the relationship among the variables used by their models.
Pablo de Morais Andrade +2 more
doaj +1 more source
Inference Attacks on Genomic Data Based on Probabilistic Graphical Models
The rapid progress and plummeting costs of human-genome sequencing enable the availability of large amount of personal biomedical information, leading to one of the most important concerns — genomic data privacy. Since personal biomedical data are highly
Zaobo He, Junxiu Zhou
doaj +1 more source
SIMLR: Machine Learning inside the SIR Model for COVID-19 Forecasting
Accurate forecasts of the number of newly infected people during an epidemic are critical for making effective timely decisions. This paper addresses this challenge using the SIMLR model, which incorporates machine learning (ML) into the epidemiological ...
Roberto Vega +2 more
doaj +1 more source

