From 2435d66feaaccea1b1cdf289b9e8305692d14dd0 Mon Sep 17 00:00:00 2001 From: Thibaut Horel Date: Sun, 10 Feb 2013 16:40:30 -0800 Subject: Fix citations in beginning of section 4 --- main.tex | 24 +++++++++++------------- 1 file changed, 11 insertions(+), 13 deletions(-) (limited to 'main.tex') diff --git a/main.tex b/main.tex index 1f92091..8c57916 100644 --- a/main.tex +++ b/main.tex @@ -45,7 +45,7 @@ approximation ratio \paragraph{Strategic case} We could design the allocation function of our mechanism by following the full information approach: allocating to agent $i^*$ when $V(i^*)\geq V(S_G)$ and to -$S_G$ otherwise. However, Singer~\cite{singer-influence} notes that this +$S_G$ otherwise. However, \citeN{singer-influence} notes that this allocation function is not monotonous, and thus breaks incentive compatibility by Myerson's theorem (Theorem~\ref{thm:myerson}). @@ -55,14 +55,13 @@ full-information case after removing $i^*$ from the set $\mathcal{N}$: OPT_{-i^*} \defeq \max_{S\subseteq\mathcal{N}\setminus\{i^*\}} \Big\{V(S) \;\Big| \; \sum_{i\in S}c_i\leq B\Big\} \end{equation} -Singer~\cite{singer-mechanisms} and Chen \emph{et al.}~\cite{chen} prove that -allocating to $i^*$ when $V(i^*)\geq C.OPT_{-i^*}$ (for some constant $C$) and -to $S_G$ otherwise yields a 8.34-approximation mechanism. However, -$OPT_{-i^*}$, the solution of the full-information case, cannot be computed in -poly-time when the underlying problem is NP-hard (unless P=NP), as is the case -for \EDP{}. - -Instead, Singer~\cite{singer-mechanisms} and Chen \emph{et al.}~\cite{chen} +\citeN{singer-mechanisms} and \citeN{chen} prove that allocating to $i^*$ when +$V(i^*)\geq C.OPT_{-i^*}$ (for some constant $C$) and to $S_G$ otherwise yields +a 8.34-approximation mechanism. However, $OPT_{-i^*}$, the solution of the +full-information case, cannot be computed in poly-time when the underlying +problem is NP-hard (unless P=NP), as is the case for \EDP{}. + +Instead, \citeN{singer-mechanisms} and Chen \emph{et al.}~\cite{chen} suggest to replace $OPT_{-i^*}$ by a quantity $L^*$ satisfying the following properties: \begin{itemize} @@ -94,11 +93,10 @@ Section~\ref{sec:relaxation}) yields an constant-approximation mechanism for In \cite{singer-influence}, the optimization program \eqref{relax} cannot be solved efficiently when $R$ is the multilinear extension of $V$. Consequently, Singer introduces a second relaxation which he relates to the -multilinear extension through the \emph{pipage rounding} framework of Ageev and -Sviridenko~\cite{pipage}. +multilinear extension through the \emph{pipage rounding} framework of \citeN{pipage}. \paragraph{Our approach} -Following Chen \emph{et al.}~\cite{chen} and Singer~\cite{singer-influence}, +Following Chen \citeN{chen} and \citeN{singer-influence}, we in turn introduce a relaxation $L$ specifically tailored to the value function of \EDP: \begin{displaymath} @@ -114,7 +112,7 @@ $O(\log\log\varepsilon^{-1})$ iterations. Being the solution to a maximization problem, $L^*$ satisfies the required monotonicity property. The main challenge will be to prove that $OPT'_{-i^*}$, for the relaxation $R=L$, is close to $OPT$. As in~\cite{singer-influence} this will be done by using the -\emph{pipage rounding} framework of~\cite{pipage} and forms the technical bulk +\emph{pipage rounding} framework of~\citeN{pipage} and forms the technical bulk of the paper. The resulting mechanism for \EDP{} is composed of -- cgit v1.2.3-70-g09d2