Scenarios
- Completely observed GMs
- Partially or Unobserved GMs
Estimation principles
- MLE.
- Bayesian estimation.
- Maximum conditional likelihood.
- Maximum “Margin”.
- Maximum entropy.
ℹ️
Learning $=$ estimating the parameters + the topology of the network.
Parameter Learning
-
Assume $\mathcal{G}$ is known and fixed,
- from expert design.
- from an intermediate outcome of iterative structure learning.
-
$D = \{x_1, \dots ,x_N\}$: $N$ independent, identically distributed (iid)
training cases.
- $x_n = (x_{n, 1}, \dots, x_{n, M})$ is a vector of $M$ values
- the model is observed, every $x_n$ is known
- or, paritially observed.
-
Goal: Estimate from $D$.
Consider learning params for a BN given the structure and is completely observable. The likelihood of the data
$$
I(\theta | D) = \log p(D | \theta) = \log \left( \prod_{i} \prod_{j} p(x_{i,j} | \theta) \right) = \sum_{i} \sum_{j} \log p(x_{i,j} | \theta)
$$