Sato, Taisuke ;
Kameya, Yoshitaka ;
Kurihara, Kenichi
Variational Bayes via Propositionalization
Abstract
We propose a unified approach to VB (variational Bayes) in
symbolicstatistical modeling via propositionalization.
By propositionalization we mean, broadly, expressing and
computing probabilistic models such as BNs (Bayesian
networks) and PCFGs (probabilistic context free grammars)
in terms of propositional logic that considers
propositional variables as binary random variables.
Our proposal is motivated by three observations. The
first one is that PPC (propostionalized probability
computation), i.e. probability computation formalized in
a propositional setting, has turned out to be general and
efficient when variable values are sparsely
interdependent. Examples include (discrete) BNs, PCFGs
and more generally PRISM which is a Turing complete logic
programming language with EM learning ability we have been
developing, and computes probabilities using graphically
represented AND/OR boolean formulas. Efficiency of PPC is
classically testified by the InsideOutside algorithm in
the case of PCFGs and by recent PPC approaches in the case
of BNs such as the one by Darwiche et al. that exploits
$0$ probability and CSI (context specific independence).
Dechter et al. also revealed that PPC is a general
computation scheme for BNs by their formulation of AND/OR
search spaces.
Second of all, while VB has been around for sometime as a
practically effective approach to Bayesian modeling, it's
use is still somewhat restricted to simple models such as
BNs and HMMs (hidden Markov models) though its usefulness
is established through a variety of applications from
model selection to prediction. On the other hand it is
already proved that VB can be extended to PCFGs and is
efficiently implementable using dynamic programming. Note
that PCFGs are just one class of PPC and much more general
PPC is realized by PRISM. Accordingly if VB is extened to
PRISM's PPC, we will obtain VB for general probabilistic
models, far wider than BNs and PCFGs.
The last observation is that once VB becomes available in
PRISM, it saves us a lot of time and energy. First we do
not have to derive a new VB algorithm from scratch for
each model and implement it. All we have to do is just to
write a probabilistic model at predicate level. The rest
of work will be carried out automatically in a unified
manner by the PRISM system as it happens in the case of EM
learning. Deriving and implementing a VB algorithm is a
tedious errorprone process, and ensuring its correctness
would be difficult beyond PCFGs without formal semantics.
PRISM augmented with VB will completely eliminate such
needs and make it easy to explore and test new Bayesian
models by helping the user cope with data sparseness and
avoid overfitting.
BibTeX  Entry
@InProceedings{sato_et_al:DSP:2008:1386,
author = {Taisuke Sato and Yoshitaka Kameya and Kenichi Kurihara},
title = {Variational Bayes via Propositionalization},
booktitle = {Probabilistic, Logical and Relational Learning  A Further Synthesis},
year = {2008},
editor = {Luc de Raedt and Thomas Dietterich and Lise Getoor and Kristian Kersting and Stephen H. Muggleton},
number = {07161},
series = {Dagstuhl Seminar Proceedings},
ISSN = {18624405},
publisher = {Internationales Begegnungs und Forschungszentrum f{\"u}r Informatik (IBFI), Schloss Dagstuhl, Germany},
address = {Dagstuhl, Germany},
URL = {http://drops.dagstuhl.de/opus/volltexte/2008/1386},
annote = {Keywords: Variational Bayes, propositionalized probability computation, PRISM}
}
2008
Keywords: 

Variational Bayes, propositionalized probability computation, PRISM 
Seminar: 

07161  Probabilistic, Logical and Relational Learning  A Further Synthesis

Related Scholarly Article: 


Issue date: 

2008 
Date of publication: 

2008 