aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--brainstorm/results.tex47
1 files changed, 46 insertions, 1 deletions
diff --git a/brainstorm/results.tex b/brainstorm/results.tex
index 4f2a70d..0453d95 100644
--- a/brainstorm/results.tex
+++ b/brainstorm/results.tex
@@ -562,7 +562,11 @@ will be correct with high probability is $S$ is drawn from the same
distribution as above.
\paragraph{Claim 2:} $g$ is not optimizable under budget $\sqrt{n}$ because you
-never learn anything about $X$.
+never learn anything about $X$. The set you output will necessarily be random
+with respect to $X$: $\sqrt{n}|X \cap S| \approx
+n^{\frac{1}{3}}(\frac{\sqrt{n}}{n})\sqrt{n}$ and therefore $g(S) \approx
+\sqrt{n}$ with high probability. However if we had been able to output$X$, we
+would have $g(X) = n^{\frac{5}{6}}$.
\paragraph{Open Question:} Cook a similar example where $g$ is submodular and
where you are observing sets of the same size as your budget.
@@ -573,6 +577,12 @@ guarantees are sufficient for optimization. First look at learning weights in
a cover function. Maybe facility location? Sums of concave over modular are
probably too hard because of the connection to neural networks.
+\paragraph{Notes}
+
+Are you sure the example shouldn't hav $n^{1/3}$ rather than $\sqrt{n}$?
+Otherwise a random set $S$ will be such that $\sqrt{n}|X \cap S|
+\sqrt{n}\frac{\sqrt{n}}{2} = \frac{n}{2}$, which is roughly equal to $|S|$.
+
\section{Which learning guarantees imply optimization?}
Here, we consider the following question: which learning guarantees imply
@@ -610,6 +620,41 @@ S}_i^c$, $v(S\cup\{j\}) \geq \alpha_i - 4\epsilon$. It follows that
$\mathbb{E}_{x \sim {\cal U}}[|v(x)|] \geq \frac{1}{4}(\alpha_j - 4\epsilon)$
and therefore we have $\|w - \hat w\|_{\infty} \leq 8 \epsilon$.
+\section{Tricky coverage function example}
+
+Let $U = \{x_1, x_2\}$ be a universe of size $2$ and $w : x \in U \mapsto
+\mathbb{R}$ a weight function on $U$. Consider a family ${\cal A}$ of $n$ sets
+$A_1, A_2, \dots A_n$ such that $A_1 \equiv \{x_1\}$, $A_2 \equiv \{x_2\}$ and
+$\forall i \neq 1,2, A_i = \{x_1, x_2\}$. Let $k \in \mathbb{R}^+$ be any
+positive constant and consider the family of coverage functions defined on $U$
+and the family of sets ${\cal A}$:
+$${\cal C} \equiv \{c : {\cal A} \mapsto \mathbb{R}^+ : \forall S \subset {\cal
+A}, c(S) = w(\cup_{S_i \in S} S_i) \text{ and } w(U) = k \}$$
+
+Note that any coverage function from this family of coverage functions is
+constant almost everywhere, equal to $w(x_1)+w(x_2) = k$. $\forall S \neq A_1,
+A_2, c(S) = k$ except for two singleton sets of ${\cal A}$: $c(\{A_1\}) =
+w(x_1)$, and $c(\{A_2\}) = w(x_2)$.
+
+It can be shown that this class of functions is $PAC$-learnable, and
+$PMAC$-learnable. However, given any distribution ${\cal D}$ on $2^U$ such that
+$\mathbb{P}(\{A_1\}) + \mathbb{P}(\{A_2\}) = o(n^{-c})$ for any constant $c$,
+then for $m = \Omega(n^c)$ for a constant $c \in \mathbb{R}^+$ and for $\hat S$
+returned by any one-shot optimization algorithm ${\cal B}$:
+$$\mathbb{P}_{S_1, \dots S_m \sim {\cal D}}\left(\mathbb{P}_{\hat S}
+\left(f(\hat S) \geq \frac{OPT}{n}\right) \geq 1- \epsilon\right) \leq \delta$$
+
+This example is not fully satisfying for two reasons:
+\begin{itemize}
+ \item the coverage function is not monotone.
+ \item the sets we observe most do not satisfy the constraint.
+\end{itemize}
+Note that even if we observe sets which \emph{almost} satisfy the cardinality
+constraint, such as all sets of size 2, we would still have no hope to optimize
+the function within better than a $\frac{1}{n}$ approximation factor.
+
+This example can be modified to address the second issue.
+
\bibliography{optimize} \bibliographystyle{apalike}