Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilme s Page 1 University of Washington Department of Electrical Engineering EE512 Spring, 2006 Graphical Models Jeff A. Bilmes <[email protected]> Jeff A. Bilmes <[email protected]> Lecture 1 Slides March 28 th , 2006
51
Embed
Lec 1: March 28th, 2006EE512 - Graphical Models - J. BilmesPage 1 Jeff A. Bilmes University of Washington Department of Electrical Engineering EE512 Spring,
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 1
University of WashingtonDepartment of Electrical Engineering
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 6
• A graphical model is a visual, abstract, and mathematically formal description of properties of families of probability distributions (densities, mass functions)
• There are many different types of Graphical model, ex:
– Bayesian Networks
– Markov Random Fields
– Factor Graph
– Chain Graph
Graphical Models
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 7
Graphical Models
Chain GraphsCausal Models
DGMsUGMs
Bayesian Networks
MRFs
Gibbs/Boltzman Distributions
DBNs Mixture Models
Decision Trees
Simple Models
PCA
LDAHMM
Factorial HMM/Mixed Memory Markov Models
BMMs
Kalman
Other Semantics
FST
Dependency Networks
Segment Models
AR
ZMs
Factor Graphs
GMs cover many well-known methods
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 8
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 9
GMs give us:
I. Structure: A method to explore the structure of “natural” phenomena (causal vs. correlated relations, properties of natural signals and scenes)
II. Algorithms: A set of algorithms that provide “efficient” probabilistic inference and statistical decision making
III. Language: A mathematically formal, abstract, visual language with which to efficiently discuss families of probabilistic models and their properties.
Graphical Models Provide
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 10
GMs give us (cont):
IV. Approximation: Methods to explore systems of approximation and their implications. E.g., what are the consequences of a (perhaps known to be) wrong assumption?
V. Data-base: Provide a probabilistic “data-base” and corresponding “search algorithms” for making queries about properties in such model families.
GMs Provide
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 11
GMs
• There are many different types of GM.• Each GM has its semantics• A GM (under the current semantics) is really a set of
constraints. The GM represents all probability distributions that obey these constraints, including those that obey additional constraints (but not including those that obey fewer constraints).
• Most often, the constraints are some form of factorization property, e.g., f() factorizes (is composed of a product of factors of subsets of arguments).
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 12
Types of Queries
• Several types of queries we may be interested in:– Compute: p(one subset of vars)– Compute: p(one subset of vars| another subset of vars)– Find the N most probable configurations of one subset of variables
given assignments of values to some other sets– Q: Is one subset independent of another subset?– Q: Is one subset independent of another given a third?
• How efficiently can we do this? Can this question be answered? What if it is too costly, can we approximate, and if so, how well? These are questions we will answer this term.
• GMs are like a probabilistic data-base (or data structure), a system that can be queried to provide answers to these sorts of questions.
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 13
Example
• Typical goal of pattern recognition:– training (say, EM or gradient descent), need query of form:
In this form, we need to compute p(o,h) efficiently.– Bayes decision rule, need to find best class for a given unknown
pattern:
– but this is yet another query on a probability distribution.– We can train, and perform Bayes decision theory quickly if we can
compute with probabilities quickly. Graphical models provide a way to reason about, and understand when this is possible, and if not, how to reasonably approximate.
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 14
Some Notation
• Random variables , , , (scalar or vector)
• Distributions:
• Subsets:
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 15
Main types of Graphical Models
• Markov Random Fields – a form of undirected graphical model– relatively simple to understand their semantics– also, log-linear models, Gibbs distributions, Boltzman distributions,
many “exponential models”, conditional random fields (CRFs), etc.
• Bayesian networks– a form of directed graphical model– originally developed to represent a form of causality, but not ideal for
that (they still represent factorization)– Semantics more interesting (but trickier) than MRFs
• Factor Graphs– pure, the assembly language models for factorization properties– came out of coding theory community (LDPC, Turbo codes)
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 16
Main types of Graphical Models
• Chain graphs: – Hybrid between Bayesian networks and MRFs– A set of clusters of undirected nodes connected as directed links– Not as widely used, but very powerful.
• Ancestral graphs– we probably won’t cover these.
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 17
C
X
Mixturemodels
( ) ( | )ii
p x c p x i
Q1 Q2 Q3 Q4
Markov Chains
)|()|( 11:1 tttt qqpqqp
Q1 Q2 Q3 Q4
),|()|( 211:1 ttttt qqqpqqp
Bayesian Network Examples
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 18
X
Y
Other generalizations possible E.g., Q = gen. diagonal, or capture using general A since
),( RNAXY
(0, )X N Q
Q
R
PCA: Q = , R 0, A = ortho
FA: Q = I, R = diagonal
XY
Y X u
( , )TY N AQA R
GMs: PCA and Factor Analysis
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 19
X1 X2 X3 X4
I1 I21 2I I
The data X1:4 is explained by the two (marginally) independent causes.
Independent Component Analysis
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 20
C
X
( ) ( )( | )
( ) ( )i
kk
f X p iP C i X
f X p k
( ) ( , )j jf X N
• Class conditional data has diff. mean but common covariance matrix.
• Fisher’s formulation: project onto space spanned by the means.
Linear Discriminant Analysis
Lec 1: March 28th, 2006 EE512 - Graphical Models - J. Bilmes Page 21