diff options
Diffstat (limited to 'finale/sections/experiments.tex')
| -rw-r--r-- | finale/sections/experiments.tex | 48 |
1 files changed, 38 insertions, 10 deletions
diff --git a/finale/sections/experiments.tex b/finale/sections/experiments.tex index 534ea7b..00950aa 100644 --- a/finale/sections/experiments.tex +++ b/finale/sections/experiments.tex @@ -26,7 +26,10 @@ one for the infected nodes and one for the susceptible nodes, the variational inference objective can be written as a simple computation graph on matrices that Theano compiles to machine code optimized for GPU execution. -\paragraph{Synthetic graph.} Since intuitively if nodes are exchangeable +The code for our Variational Inference as well as MCMC implementations can be +found at \url{http://thibaut.horel.org/cs281_project.zip}. + +\paragraph{Synthetic graph.} Since intuitively, if nodes are exchangeable (\emph{i.e} their topological properties like degree are identical) in our graph, the active learning policy will make little difference compared to the uniform-source policy, we decided to test our algorithms on an unbalanced graph @@ -42,19 +45,44 @@ A = \left( \begin{array}{cccccc} \right) \end{equation*} In other words, graph $\mathcal{G}_A$ is a wheel graph where every node, except -for the center node, points to its (clock-wise) neighbor. +for the center node, points to its (clock-wise) neighbor. Even though the wheel +graph might not seem realistic, it is known that social networks (as well as +graphs found in many other real-life applications) exhibit a power-law degree +distribution which is also highly unbalanced. + +\paragraph{Baselines.} Our experiments involved comparing the active learning +described in Section 4 to two baselines. For easier comparison, both baselines +are also online learning methods: cascades are generated on the fly and only +used once by the SGD algorithm to update the approximate posterior parameters. +The two baselines are: +\begin{itemize} + \item \emph{uniform source:} for each cascade the source node is chosen + uniformly at random among all vertices. Intuitively, this method should + behaves similarly to offline learning. + \item \emph{out-degree:} the source node is selected with + probability proportional to the estimated out-degree (\emph{i.e} the + sum of the estimated means on the outgoing edges). +\end{itemize} -\paragraph{Baseline.} In order for the baseline to be fair, we choose to create -cascades starting from the source node on the fly both in the case of the -uniform source and for the active learning policy. Each cascade is therefore -`observed' only once. We plot the RMSE of the graph i.e. $RMSE^2 -= \frac{1}{n^2} \|\hat{\mathbf{\Theta}} - \mathbf{\Theta}\|^2_2$. The results -are presented in +Our active learning method, henceforth called \emph{out-variance} is then +compared to the two baselines by plotting the RMSE of the estimated graph, +\emph{i.e} $\text{RMSE}^2 = \frac{1}{n^2} \|\hat{\mathbf{\Theta}} +- \mathbf{\Theta}\|^2_2$, as a function of the observed number of cascades. The +results are presented in \cref{fig:results}. As we can see, both +\emph{out-degree} and \emph{out-variance} which are active in nature vastly +outperform \emph{uniform source} which is passive. Surprisingly, +\emph{out-degree} which is seemingly very crude since it does not take into +account the current uncertainty performs comparably to \emph{out-variance}. We +believe that this is due to the specific nature of the wheel graph: for this +graph \emph{out-degree} repeatedly selects the central node which is adjacent +to 50\% of the graph edges. \begin{figure} \centering -\label{fig:graphical} \includegraphics[scale=.4]{finale.png} -\caption{} +\caption{RMSE of the proposed active learning heuristic compared to the +baselines. The $x$ axis shows the number of batches: each batch contains 100 +time steps of the cascade process.} +\label{fig:results} \end{figure} |
