In this section, we apply the framework from Section~\ref{sec:bayes} and~\ref{sec:active} on synthetic graphs and cascades to validate the Bayesian approach as well as the effectiveness of the Active Learning heuristics. We started by using the PyMC library \cite{pymc} which implements MCMC algorithms to sample from the posterior distribution directly. We found this method to scale poorly with the number of nodes in the graph, such that graphs of size $\geq 100$ could not be reasonably be learned in the amount of time we had. In the appendix, we show the progressive convergence of the posterior around the true values of the edge weights of the graph for a graph of size $4$. \paragraph{VI implementation.} In order to show the effect of the active learning policies, we needed to scale the experiments to graphs of size $\geq 1000$. A graph of size $1000$ has $1M$ parameters to be learned ($2M$ in the product-prior of Eq.~\ref{eq:gaussianprior}). The maximum-likelihood estimator converges to an $\ell_\infty$-error of $.05$ for most graphs after having observed at least $100M$ distinct cascade-steps. Given the number of parameters to be learned and the number of cascades to process, we had to resort to Variational Inference. We therefore used the Blocks~\cite{blocks} framework, written on top of Theano for automatic differentiation and highly efficient SGD routines with minimal memory overhead. By encoding a cascade step as a tuple of two binary vectors, one for the infected nodes and one for the susceptible nodes, the variational inference objective can be written as a simple computation graph on matrices that Theano compiles to machine code optimized for GPU execution. The code for our Variational Inference as well as MCMC implementations can be found at \url{http://thibaut.horel.org/cs281_project.zip}. \paragraph{Synthetic graph.} Since intuitively, if nodes are exchangeable (\emph{i.e} their topological properties like degree are identical) in our graph, the active learning policy will make little difference compared to the uniform-source policy, we decided to test our algorithms on an unbalanced graph $\mathcal{G}_A$ whose adjacency matrix $A$ is as follows: \begin{equation*} A = \left( \begin{array}{cccccc} 0 & 1 & 1 & 1 & \dots & 1 \\ 0 & 0 & 1 & 0 & \dots & 0 \\ 0 & 0 & 0 & 1 & \dots & 0 \\ \vdots & \vdots & \vdots & \vdots & \vdots & \vdots \\ 0 & 1 & 0 & 0 & \dots & 0 \end{array} \right) \end{equation*} In other words, graph $\mathcal{G}_A$ is a wheel graph where every node, except for the center node, points to its (clock-wise) neighbor. Even though the wheel graph might not seem realistic, it is known that social networks (as well as graphs found in many other real-life applications) exhibit a power-law degree distribution which is also highly unbalanced. \paragraph{Baselines.} Our experiments involved comparing the active learning described in Section 4 to two baselines. For easier comparison, both baselines are also online learning methods: cascades are generated on the fly and only used once by the SGD algorithm to update the approximate posterior parameters. The two baselines are: \begin{itemize} \item \emph{uniform source:} for each cascade the source node is chosen uniformly at random among all vertices. Intuitively, this method should behaves similarly to offline learning. \item \emph{out-degree:} the source node is selected with probability proportional to the estimated out-degree (\emph{i.e} the sum of the estimated means on the outgoing edges). \end{itemize} Our active learning method, henceforth called \emph{out-variance} is then compared to the two baselines by plotting the RMSE of the estimated graph, \emph{i.e} $\text{RMSE}^2 = \frac{1}{n^2} \|\hat{\mathbf{\Theta}} - \mathbf{\Theta}\|^2_2$, as a function of the observed number of cascades. The results are presented in \cref{fig:results}. As we can see, both \emph{out-degree} and \emph{out-variance} which are active in nature vastly outperform \emph{uniform source} which is passive. Surprisingly, \emph{out-degree} which is seemingly very crude since it does not take into account the current uncertainty performs comparably to \emph{out-variance}. We believe that this is due to the specific nature of the wheel graph: for this graph \emph{out-degree} repeatedly selects the central node which is adjacent to 50\% of the graph edges. \begin{figure} \centering \includegraphics[scale=.4]{finale.png} \caption{RMSE of the proposed active learning heuristic compared to the baselines. The $x$ axis shows the number of batches: each batch contains 100 time steps of the cascade process.} \label{fig:results} \end{figure}