diff options
Diffstat (limited to 'brainstorm')
| -rw-r--r-- | brainstorm/results.tex | 47 |
1 files changed, 46 insertions, 1 deletions
diff --git a/brainstorm/results.tex b/brainstorm/results.tex index 4f2a70d..0453d95 100644 --- a/brainstorm/results.tex +++ b/brainstorm/results.tex @@ -562,7 +562,11 @@ will be correct with high probability is $S$ is drawn from the same distribution as above. \paragraph{Claim 2:} $g$ is not optimizable under budget $\sqrt{n}$ because you -never learn anything about $X$. +never learn anything about $X$. The set you output will necessarily be random +with respect to $X$: $\sqrt{n}|X \cap S| \approx +n^{\frac{1}{3}}(\frac{\sqrt{n}}{n})\sqrt{n}$ and therefore $g(S) \approx +\sqrt{n}$ with high probability. However if we had been able to output$X$, we +would have $g(X) = n^{\frac{5}{6}}$. \paragraph{Open Question:} Cook a similar example where $g$ is submodular and where you are observing sets of the same size as your budget. @@ -573,6 +577,12 @@ guarantees are sufficient for optimization. First look at learning weights in a cover function. Maybe facility location? Sums of concave over modular are probably too hard because of the connection to neural networks. +\paragraph{Notes} + +Are you sure the example shouldn't hav $n^{1/3}$ rather than $\sqrt{n}$? +Otherwise a random set $S$ will be such that $\sqrt{n}|X \cap S| +\sqrt{n}\frac{\sqrt{n}}{2} = \frac{n}{2}$, which is roughly equal to $|S|$. + \section{Which learning guarantees imply optimization?} Here, we consider the following question: which learning guarantees imply @@ -610,6 +620,41 @@ S}_i^c$, $v(S\cup\{j\}) \geq \alpha_i - 4\epsilon$. It follows that $\mathbb{E}_{x \sim {\cal U}}[|v(x)|] \geq \frac{1}{4}(\alpha_j - 4\epsilon)$ and therefore we have $\|w - \hat w\|_{\infty} \leq 8 \epsilon$. +\section{Tricky coverage function example} + +Let $U = \{x_1, x_2\}$ be a universe of size $2$ and $w : x \in U \mapsto +\mathbb{R}$ a weight function on $U$. Consider a family ${\cal A}$ of $n$ sets +$A_1, A_2, \dots A_n$ such that $A_1 \equiv \{x_1\}$, $A_2 \equiv \{x_2\}$ and +$\forall i \neq 1,2, A_i = \{x_1, x_2\}$. Let $k \in \mathbb{R}^+$ be any +positive constant and consider the family of coverage functions defined on $U$ +and the family of sets ${\cal A}$: +$${\cal C} \equiv \{c : {\cal A} \mapsto \mathbb{R}^+ : \forall S \subset {\cal +A}, c(S) = w(\cup_{S_i \in S} S_i) \text{ and } w(U) = k \}$$ + +Note that any coverage function from this family of coverage functions is +constant almost everywhere, equal to $w(x_1)+w(x_2) = k$. $\forall S \neq A_1, +A_2, c(S) = k$ except for two singleton sets of ${\cal A}$: $c(\{A_1\}) = +w(x_1)$, and $c(\{A_2\}) = w(x_2)$. + +It can be shown that this class of functions is $PAC$-learnable, and +$PMAC$-learnable. However, given any distribution ${\cal D}$ on $2^U$ such that +$\mathbb{P}(\{A_1\}) + \mathbb{P}(\{A_2\}) = o(n^{-c})$ for any constant $c$, +then for $m = \Omega(n^c)$ for a constant $c \in \mathbb{R}^+$ and for $\hat S$ +returned by any one-shot optimization algorithm ${\cal B}$: +$$\mathbb{P}_{S_1, \dots S_m \sim {\cal D}}\left(\mathbb{P}_{\hat S} +\left(f(\hat S) \geq \frac{OPT}{n}\right) \geq 1- \epsilon\right) \leq \delta$$ + +This example is not fully satisfying for two reasons: +\begin{itemize} + \item the coverage function is not monotone. + \item the sets we observe most do not satisfy the constraint. +\end{itemize} +Note that even if we observe sets which \emph{almost} satisfy the cardinality +constraint, such as all sets of size 2, we would still have no hope to optimize +the function within better than a $\frac{1}{n}$ approximation factor. + +This example can be modified to address the second issue. + \bibliography{optimize} \bibliographystyle{apalike} |
