summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorunknown <Brano@Toshibicka.(none)>2012-03-05 11:55:59 -0800
committerunknown <Brano@Toshibicka.(none)>2012-03-05 11:55:59 -0800
commit4f4d6896872f30bf36bba8975aca61ddf3cda156 (patch)
treece668c6c575ac98198f87183dcccfc873f7ec373
parent48df7f24114a224c4e8266813f5002cc791704f2 (diff)
downloadkinect-4f4d6896872f30bf36bba8975aca61ddf3cda156.tar.gz
Algorithms
-rw-r--r--algorithm.tex2
1 files changed, 1 insertions, 1 deletions
diff --git a/algorithm.tex b/algorithm.tex
index 1747227..30ce77c 100644
--- a/algorithm.tex
+++ b/algorithm.tex
@@ -20,7 +20,7 @@ distribution of the model is given by:
P(\bx, y) = \cN(\bx | \bar{\bx}_y, \Sigma) P(y),
\label{eq:mixture of Gaussians}
\end{align}
-where $\bx$ denotes an observation, $y$ is a class, $P(y)$ is the probability of the class, and $\cN(\bx | \bar{\bx}_y, \Sigma)$ is the conditional probability of $\bx$ given $y$. The conditional is a multivariate normal distribution. The mean of the distribution is $\bar{\bx}_y$ and the variance of $\bx$ is captured by the covariance matrix $\Sigma$. The decision boundary between any two classes is linear when all conditionals $\cN(\bx | \bar{\bx}_y, \Sigma)$ have the same covariance matrix \cite{bishop06pattern}. In this setting, the mixture of Gaussians model can be viewed as a probabilistic variant of the nearest-neighbor (NN) classifier in Section~\ref{sec:uniqueness}.
+where $\bx$ denotes an observation, $y$ is a class, $P(y)$ is the probability of the class, and $\cN(\bx | \bar{\bx}_y, \Sigma)$ is the conditional probability of $\bx$ given $y$. The conditional is a multivariate normal distribution. Its mean is $\bar{\bx}_y$ and the variance of $\bx$ is captured by the covariance matrix $\Sigma$. The decision boundary between any two classes $y$ is linear when all conditionals $\cN(\bx | \bar{\bx}_y, \Sigma)$ have the same covariance matrix $\Sigma$ \cite{bishop06pattern}. In this scenario, the mixture of Gaussians model can be viewed as a probabilistic variant of the nearest-neighbor (NN) classifier in Section~\ref{sec:uniqueness}.
The mixture of Gaussians model has many advantages. First, the model can be
easily learned using maximum-likelihood (ML) estimation \cite{bishop06pattern}.