This repository has been archived by the owner on Feb 28, 2019. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 6
/
Lecture07-LowerBoundConstructions.tex
690 lines (611 loc) · 34 KB
/
Lecture07-LowerBoundConstructions.tex
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
\input{template}
\input{macros}
\usepackage{dsfont}
\theoremstyle{plain}
\theorembodyfont{\rmfamily}
\newtheorem*{claim}{Claim:}
\renewcommand{\P}[1]{\mathds{P}\left[{#1}\right]}
\newcommand{\E}[1]{\mathds{E}\left[{#1}\right]}
\newcommand{\1}[1]{\mathds{1}[{#1}]}
\begin{document}
\lecture{7}{Communication Lower Bounds}{Clay Thomas}
\disclaimer
\section{Logistics}
First homework will be posted soon.
\section{Outline}
This class will just present a list of constructions proving communication lower
bounds in various settings. Many of them are simple, few of them are advanced
and may skip details. Just in case some people haven't seen communication
complexity before, I'll quickly go through the basic arguments to start.
\begin{definition} A two-party communication problem consists of a boolean
function $f:\{0,1\}^a \times \{0,1\}^b \rightarrow \{0,1\}$ (for this entire
lecture, let $a = b = k$). Alice is given some input $A \in \{0,1\}^a$ (but
doesn't know anything about $B$), and Bob is given some input $B \in \{0,1\}^b$
(but doesn't know anything about $A$). Their goal is to compute $f(A, B)$.
\end{definition} \begin{definition}
A \emph{deterministic communication
protocol} for $f$ specifies for Alice, as a function of her input $A$ and all
messages sent so far $a_1, b_1,\ldots, a_\ell, b_\ell$, what message
$a_{\ell+1}$ to send next. It also specifies for Bob, as a function of his
input $B$ and all messages sent so far what message $b_{\ell+1}$ to send next.
We'll be interested in the \emph{communication complexity} of a protocol: the
maximum number of bits ever sent in total between Alice and Bob.
\end{definition}
You can take the final output of a protocol as
a function of the entire transaction
history $a_1, b_1, \ldots, a_\ell, b_\ell$.
Alternatively, you can stop communication once either party knows the
answer (or communicate that answer back with one extra bit).
The exact details on this point don't matter.
\begin{Exa}[Equality]
Let $f(A, B) = 1$ if and only if $A = B$. One protocol to
solve this is the following: Alice and Bob both output their entire string in
the first round, and declare $1$ if and only if their outputs match. A slightly
more clever protocol might have Alice and Bob each output the first bit of their
string and see if they match. If so, continue. If not, output $0$. If they make
it all the way to the end without stopping, output $1$.
\end{Exa}
\begin{Exa}[Disjointness] For this problem, it will make more sense to think of
$A$ as a subset of $[k]$ (the coordinates which are $1$), and $B$ as a subset of
$[k]$. Then consider $f(A, B) = 1$ if and only if $A \cap B = \emptyset$. One
protocol to solve this is the following: Alice and Bob both output their entire
set, and declare $1$ if and only if their outputs match.\end{Exa}
Both examples are core problems for communication complexity. Interestingly, the
natural procedures we posed seem to basically communicate the entire input in
the worst case. It turns out that this is because the protocols are (almost)
optimal (among deterministic protocols). We'll prove this using a generic type
of argument, called a \emph{fooling set argument}. The main idea is the
following: Imagine for instance that the complete transcript (that is, all
messages sent by Alice and all messages sent by Bob, in order) are identical for
inputs $(A_1, B_1)$ and $(A_2, B_2)$. Then we claim that the transcripts must
also be identical for $(A_1, B_2)$ and $(A_2, B_1)$.
\begin{observation}[Rectangles]
Let $T(A, B)$ denote the complete transcript
(all messages sent, in order, and their final answers) produced by Alice and Bob
using a communication protocol on input $(A, B)$.
Suppose $T(A_1, B_1) = T(A_2, B_2) = T$.
Then $T(A_2, B_1) = T = T(A_1, B_2)$ as well,
and thus the protocol gives the same answer on all four input pairs.
\end{observation}
\begin{proof}
Assume for contradiction that this were not the
case, and w.l.o.g. say that the different transcript is $T(A_1, B_2) \neq
T(A_2, B_2)$. Because the protocol is deterministic, and Bob has the same
input in both instances, Bob will continue outputting the same message every
round until Alice sends a different message. Therefore, in order for
$T(A_1, B_2)$ to possibly not equal $T(A_2, B_2)$,
it must be the case that Alice sends the first different message.
Let $T$ be the transcript prior to the first different message.
Because $T(A_2, B_2) = T(A_1, B_1)$, and Alice's next message can
depend only on $T$ and her own input,
Alice must send the same next message if her input is $A_1$ or $A_2$.
This is a contradiction, so the transcripts must indeed be equal.
\end{proof}
The above is called a ``rectangle argument'' because if we were to draw a matrix
where the rows are Alice's input and columns are Bob's, it means that entries
that share the same transcript must form a rectangle (okay, not exactly, but for
every transcript $T$, there exists a relabeling of the rows/columns so that the
pairs that induce this transcript are a rectangle). It seems quite simple but is
surprisingly powerful because if protocols use limited communication, it must be
the case that multiple inputs have the same transcript
(by the pigeonhole principle), and the rectangle
argument lets us conclude that other inputs necessarily use the same transcript
as well.
\begin{definition} A \emph{fooling set} for the function $f$ is a set $F
\subseteq \{0,1\}^k \times \{0,1\}^k$ such that: \begin{itemize} \item $f(A, B)
= 1$ for all $(A, B) \in F$. \item For every $(A_1, B_1), (A_2, B_2) \in
F$, either $f(A_1, B_2) \neq 1$ or $f(A_2, B_1) \neq 1$ (or both).
\end{itemize} \end{definition}
\begin{proposition}
Let $F$ be a fooling set for $f$. Then the deterministic
communication complexity of $f$ is at least $\log_2(|F|)$.
\end{proposition}
\begin{proof}
Assume for contradiction that there exists a deterministic
protocol with communication $< \log_2|F|$.
There are less than $2^{\log_2|F|}=|F|$ possible transactions such a
protocol can use.
Thus, by the pigeonhole principle,
there exists some $(A_1, B_1), (A_2, B_2) \in F$ such that
$T(A_1, B_1) = T(A_2, B_2)$.
By the rectangle argument, this necessarily implies $T(A_1, B_2)
= T(A_2, B_1) = T(A_1, B_1)$ as well.
Because $f$ is completely determined by the transaction history,
we must have $f(A_1, B_2) = f(A_2, B_1) = 1$ as well.
However, this directly contradicts the second property of fooling sets.
\end{proof}
So if a function admits a large fooling set, it is provably impossible to have a
short deterministic communication protocol. The remaining work is just to show
that both Equality and Disjointness admit large fooling sets.
\begin{proposition} Equality admits a fooling set of size $2^k$.
\end{proposition} \begin{proof}
Let $F = \{(X, X)\ |\ X \in \{0,1\}^k \}$.
Then $f(A, B) = 1$ for all $(A, B) \in F$. But $F(A_1, B_2) = 0$ for
all $A_1 \neq B_2$. So the second property of the fooling set is satisfied.
\end{proof}
Notice that after unraveling all the definitions, the proof is essentially
saying the following: if you had a deterministic protocol that used $< k$ bits,
there are two inputs of the form $(X, X)$, $(Y, Y)$ that are treated exactly the
same. This protocol must therefore also treat $(X, Y)$ and $(Y, X)$ the same as
$(X, X)$, even though the difference affects the value of $f$.
That is, the protocol is ``fooled''
into thinking the input is $(X, X)$ or $(Y, Y)$ when it really could be $(X, Y)$
or $(Y, X)$, and this affects the output of $f$.
\begin{proposition}
Disjointness admits a fooling set of size $2^k$.
\end{proposition}
\begin{proof}
Let $F = \{ (X, \bar{X})\ |\ X \subseteq [k] \}$.
Then $f(A, B) = 1$ for all $(A, B) \in F$. But now consider any $(X,
\bar{X}), (Y, \bar{Y}) \in F$. It cannot be the case that $X \cap \bar{Y} =
\emptyset$ \textbf{and} $Y \cap \bar{X} = \emptyset$ unless $Y = X$. The
former implies that $X \subseteq Y$ and the latter implies that $Y \subseteq
X$. Therefore, either $f(X, \bar{Y}) = 0$ or $f(Y, \bar{X} ) = 0$.
\end{proof}
So both protocols necessarily require communication $k$ to solve
deterministically. It turns out that both problems are also hard to solve
\emph{nondeterministically}. Disjointness is also hard to solve with a
randomized protocol, but Equality can be solved with high probability with very
low randomized communication.
\subsection{Multi-party communication}
We'll also talk about multi-party
communication instead of just two parties - this will be relevant when figuring
out how good combinatorial auctions we should possibly hope for. Here, there's
again a boolean function $f$ but it takes as input $X_1, \ldots, X_n$. Various
input models are studied, but we'll only talk about the \emph{number-in-hand}
model, where player $i$ knows only $X_i$ and nothing else.
We'll also want a canonical hard problem to start with, multi-disjointness. It
also introduces a new concept, called a \emph{promise problem}.
\begin{Exa}[Multi-Disjointness]
Let $f(X_1,\ldots, X_n) = 1$ iff $X_i \cap X_j = \emptyset$
for all $i \neq j$. Moreover, we'll be interested in the
\emph{promise problem}, where the protocol is promised that one of the
following is true about the input:
\begin{itemize}
\item $X_i \cap X_j =
\emptyset$ for all $i \neq j$. \item $\cap_i X_i \neq \emptyset$ (there
exists an $\ell$ in all $X_i$).
\end{itemize}
More clearly, we say that a
communication protocol solves Multi-Disjointness if it outputs $1$ whenever $X_i
\cap X_j = \emptyset$ for all $i \neq j$,
outputs $0$ whenever $\cap_i X_i \neq \emptyset$,
and is allowed to behave arbitrarily if neither condition holds.
\end{Exa}
\begin{theorem}[\cite{NisanPackCover02}, theorem 2]
Every deterministic protocol for
multi-disjointness has communication complexity $\Omega(k/n)$.
\end{theorem}
\begin{proof}
First, observe that an analog of the rectangles argument holds in
multi-party communication as well: if $T(X_1,\ldots, X_n) = T(Y_1,\ldots,
Y_n)$, then $T(Z_1,\ldots, Z_n) = T(X_1,\ldots, X_n)$ whenever $Z_i \in \{X_i,
Y_i\}$. The proof is the same: if not, there must be some player who is the
first to send a different message. But their input matches their input in one
of the two cases $(X_1,\ldots, X_n), (Y_1,\ldots, Y_n)$, so this is a
contradiction.
The fooling set argument also extends, but we'll use the same concepts slightly
differently below.
% Namely, we count the number of transcripts needed
First, we count the total number $1$ instances
(i.e. cases where the input sets are disjoint).
Observe that each such instance can be selected by choosing, for
each item, a player to receive that item, or no player at all.
There are $n+1$ choices, independently for all items,
giving $(n+1)^k$ possible instances.
Second, we'll first show that no protocol can use the same
transcript for more than $n^k$ inputs which evaluate to $1$.
Let $S$ be a set of inputs, say
$\{(X^{(j)}_1,\ldots,X^{(j)}_n)\}_{j=1}^{|S|}$ that
share the same transcript and are all $1$ instances.
\begin{claim}
For each $\ell \in [k]$, there exists a player $i$
such that player $i$ never gets $\ell$ in $S$
(i.e. for every input $(X^{(j)}_1,\ldots,X^{(j)}_n)\in S$,
we have $\ell \notin X^{(j)}_i$)
\end{claim}
\begin{proof}
The negation of the above statement is exactly the following:
There exists an $\ell\in[k]$ such that for all players $i$,
$i$ get $\ell$ in some input from $S$, say
$\ell \in X^{j(\ell)}_i$.
But by the rectangle argument, the input
$ (X^{j(\ell)}_1, \ldots, X^{j(\ell)}_n) $ has the same transcript
as all the elements of $S$, and thus should be a $1$ instance.
However, $\ell \in X^{j(\ell)}_1 \cap \ldots \cap X^{j(\ell)}_n$,
a contradiction.
\end{proof}
With this claim in mind, we bound the total possible number of inputs
in $S$. Consider all the ways to put an item $\ell\in[k]$
into an input tuple: it may go to no bidder, or it may go
to exactly one bidder (because the sets of an input tuple are disjoint).
But, by the above claim, each item has one bidder which it
\emph{cannot} go to.
This gives each item exactly $n$ choices,
for a total size of $S$ of at most $n^k$.
Finally, observe that the above two results mean that there are at least
$(n+1)^k/n^k = (1 + 1/n)^k$ different transcripts are needed simply to handle the
$1$-instances of the problem.
This gives communication complexity at least
$k\log_2(1+1/n) = \Theta(k/n)$.
\end{proof}
\section{Monotone Valuations}
First, we'll present a lower bound for monotone
valuations, by reduction from MultiDisjointness.
\begin{theorem}[\cite{NisanPackCover02} Theorem 3]
The communication required to guarantee strictly
better than an $n$-approximation for welfare maximization with $m$ items on all
monotone valuations is $2^{\Omega(m/n^2)}$. Note that this implies that
exponential communication among the bidders is necessary in order to guarantee
that an allocation is within a factor of $n$ of optimal when $n =
m^{1/2-\varepsilon}$.
\end{theorem}
\begin{proof}
Think of $m \gg n$ and let $k$ be an integer we'll tune later.
First, assume that we have
a $k$-size family of length-$n$ partitions of items $M$, say
$\{(T_1^\ell,\ldots, T_n^\ell)\}_{\ell \in [k]}$,
where for all $\ell$, $T_i^\ell \cap T_j^\ell = \emptyset$ (note\footnote{
Incidentally, when we actually construct this collection,
we'll get \emph{actual} partitions, i.e. $\bigcup_i T_i^\ell = M$
for each $\ell$.
However, this is unimportant for our reduction.
}).
Additionally, we require that for
all $\ell \neq \ell'$, $T_j^\ell \cap T_i^{\ell'} \neq \emptyset$;
in other words the disjointness property does not ever mix and match
across different partitions.
We will associate bidder $i$ to the (some sub collection of) the sets
$\{T^\ell_i\}$.
% The condition that $(T_1^\ell,\ldots, T_n^\ell)$ are disjoint will
% correspond to situations where every bidder can be satisfied,
% while the condition that
% That is, we
% can assign each bidder $T_i^\ell$ for the same $\ell$, but not mix and match
% across different $\ell$s.
Now, consider the following class of valuations, called \emph{multi-minded}, or
\emph{coverage} valuations: bidder $i$ has some subset $X_i \subseteq [k]$, and
$v_i(S) = 1$ if and only if $T_i^\ell \subseteq S$ for some $\ell \in X_i$ (and
$v_i(S) = 0$ otherwise). That is, bidder $i$ ``likes'' all the sets $T_i^\ell$,
for $\ell \in X_i$, and gets value $1$ as long as they get some set they like.
Observe that if there exists a single $\ell \in \cap_i X_i$, then it is possible
to achieve welfare $n$: simply give each bidder set $T_i^\ell$ for the $\ell \in
\cap_i X_i$. Also, observe that if it is possible for both bidder $i$ and bidder
$j$ to get value $1$, then the sets that they receive and are interested in must
be disjoint. For desired sets $T_i^\ell$ and $T_j^{\ell'}$, this can occur
only if $\ell = \ell'$.
So if for all $i, j$, $X_i \cap X_j = \emptyset$,
then there is no way for both bidder $i$ and bidder $j$
to get non-zero welfare, and the maximum possible welfare is $1$. So by a direct
reduction to MultiDisjointness, the communication complexity of determining
whether the welfare is $\geq n$ or $\leq 1$ is at least $\Omega(k/n)$.
Now, we just need to figure out how big of a $k$ we can have while still
guaranteeing a ``collection of partitions''
construction of the desired form. Consider $k$ random
partitions of $[m]$. That is, each item is independently and uniformly awarded
to a bidder, and $T_i^\ell$ is the set of items awarded to bidder $i$ in
partition $\ell$. Clearly, this satisfies the first property: for all $\ell$,
the $T_i^\ell$s indeed form a partition, by definition. We just need to make
sure we don't take too many random partitions so that we have bad luck and some
$T_i^\ell \cap T_j^{\ell'}= \emptyset$.
So consider a fixed $T_i^\ell, T_j^{\ell'}$ for $\ell\ne\ell'$.
What is the probability that they don't intersect?
This is the probability that for each item $x$, it is either
not put into $T_i^\ell$ or not put into $T_j^{\ell'}$.
Because partitions are selected independently, we get
\begin{align*}
\P{x\notin T_i^\ell \vee x\notin T_j^{\ell'}}
= 1 - \P{x\in T_i^\ell \wedge x\in T_j^{\ell'}}
= 1 - 1/n^2
\end{align*}
And because items are allocated independently, we get
\begin{align*}
\P{ T_i^\ell \cap T_j^{\ell'} = \emptyset}
= \prod_{x\in M} \P{x\notin T_i^\ell \vee x\notin T_j^{\ell'}}
= (1 - 1/n^2)^m
\approx e^{-m/n^2}
\end{align*}
% So the probability that this happens for every item is
% exactly $(1-1/n^2)^m$.
Now just take a union bound over all $kn\cdot(k-1)n\le (kn)^2$
pairs of sets which we need to have disjoint.
We get that the probability that any two sets happen to be
disjoint by bad luck is $\approx k^2n^2 e^{-m/n^2}$. Therefore, we can take $k =
2^{\Omega(m/n^2)}$ and get a non-zero probability of success (even if we can't
find this instance efficiently, it exists, so we can use that for the
construction and proof).
So to wrap up: we need a construction of the form given
at the beginning of the proof
for large $k$. The argument above provides a construction for $k =
2^{\Omega(m/n^2)}$, which completes the proof.
\end{proof}
Recall that we have protocols matching this: a $\sqrt{m}$-approximation due
to~\cite{LaviS05, DobzinskiNS05}, or the trivial $n$-approximation which just
allocates randomly (both of the previous results would also guarantee an
$n$-approximation).
\section{Subadditive Valuations}
\begin{theorem}[\cite{DobzinskiNS05}]The
communication required to guarantee strictly better than a
$(1/2+1/(2n))$-approximation for welfare maximization with $m$ items on all
subadditive valuations is $2^{\Omega(m/n^2)}$.
\end{theorem}
\begin{proof}
Again, assume that we have the following construction of partitions:
$\{(T_1^\ell,\ldots, T_n^\ell)\}_{\ell \in [k]}$, where for all $\ell$,
$T_i^\ell \cap T_j^\ell = \emptyset$, but for all $\ell \neq \ell'$,
$T_j^\ell \cap T_i^{\ell'} \neq \emptyset$.
% That is, we can assign each
% bidder $T_i^\ell$ for the same $\ell$, but not mix and match across
% different $\ell$s.
Observe that the previous construction is not subadditive. In particular,
player's have value $1$ for some sets, while having value $0$ for every proper
subset. The only change we'll make is to bump up the players' values for every
non-empty subset by $1$.
Specifically, bidder $i$ has some subset $X_i \subseteq [k]$,
and $v_i(S) = 2$ if
$T_i^\ell \subseteq S$ for some $\ell \in X_i$ (and $v_i(S) = 1$
otherwise). This function is now clearly subadditive: any two non-trivial sets
have value one, and no set has value more than two.
The only difference is in the welfare guarantee we can claim. Now, if we're in
the disjoint case, we can certainly get welfare $2n$. But even in the
non-disjoint case, we can get welfare $n+1$ (one player can get a set they like,
and everyone else gets a single item). So our lower bound is now only $2n/(n+1)$
instead of $n$.
\end{proof}
Recall that we have a 2-approximation (that we didn't see) due
to~\cite{Feige06}. This is asymptotically tight as $n\rightarrow \infty$, but
not for small $n$ (actually, the algorithm is tight for small $n$, and the lower
bound can be improved, see later section).
\section{XOS Valuations}
\begin{theorem}[\cite{DobzinskiNS05}]
The communication
required to guarantee strictly better than a $1-(1-1/n)^n$-approximation for
welfare maximization with $m$ items on all XOS valuations is
$2^{\Omega(m/n^2)}$.
\end{theorem}
\begin{proof} This time, we'll need a slightly stronger construction (actually,
the same construction works, we just need to appeal to a stronger property).
This time, assume that we have the following construction of partitions:
$\{T_1^\ell,\ldots, T_n^\ell\}_{\ell \in [k]}$, where for all $\ell$,
$T_i^\ell \cap T_j^\ell = \emptyset$, but for all $\ell_1,\ldots, \ell_n$ (all
distinct) we have $|\cup_i T_i^{\ell_i}| \leq m\cdot (1-(1-1/n)^n)$.
Now, consider the following class of valuations. First, for all $\ell$ define an
additive valuation $v_i^\ell(S) := |S \cap T_i^\ell|$ (that is, value one per
item in $T_i^\ell$). Next, give each bidder some set $X_i$ of indices, and
define $v_i(S) := \max_{\ell \in X_i} v_i^\ell(S)$. This is clearly XOS (even
Binary XOS).
Observe again that if there exists a single $\ell \in \cap_i X_i$, then it is
possible to achieve welfare $m$: simply give each bidder set $T_i^\ell$ for the
$\ell \in \cap_i X_i$. But now the welfare might be better in the ``no'' case,
because the bidders will still get value $1$ per item even if they barely
overlap with their sets. Specifically, if each $X_i$ is disjoint,
in any allocation there will be at most
$|\cup_i T_i^{\ell_i}|$ items which are valued at $1$
(for some $\ell_1,\ldots, \ell_n$ all distinct).
Thus, the welfare can only be as large as
$m \cdot (1-(1-1/n)^n)$.
Again, we just need to figure out how big of a $k$ we can have while still
guaranteeing a construction of the desired form. Again consider $k$ random
partitions of $[m]$. That is, each item is independently and uniformly awarded
to a bidder, and $T_i^\ell$ is the set of items awarded to bidder $i$ in
partition $\ell$. Clearly, this satisfies the first property: for all $\ell$,
the $T_i^\ell$s indeed form a partition, by definition. We just need to make
sure we don't take too many random partitions so that we have bad luck and some
$T_i^\ell \cap T_j^{\ell'}= \emptyset$.
% CHECK THIIIIS
Consider a fixed set of distinct indices $\ell_1,\ldots, \ell_n$.
We have:
\begin{align*}
\E{\ \left|\bigcup_iT_i^{\ell_i}\right|\ }
& = \sum_{j\in M} \P{j\in T_i^{\ell_i}\text{ for some $i$}} \\
& = m\left(1 - \P{j\ne T_i^{\ell_i}\text{ for all $i$}}\right) \\
& = m\left(1 - (1-1/n)^n \right) \\
\end{align*}
% (that is, for any item to miss the union,
% it needs to be missed independently $n$ times).
Because the indicator random variables for the events
``$j\in T_i^{\ell_i}$ for some $i$'' are independent for different $j$,
we can apply a Chernoff bound.
Namely, the probability that the size of the union deviates from
its expectation by more than
$\varepsilon \cdot m$ is upper bounded by $e^{-\varepsilon^2 m/2}$. So taking a
union bound over all $\leq k^n$ distinct indices, we get that this holds for all
indices with probability
$k^ne^{-\varepsilon^2 m/2} = e^{-\varepsilon^2 m/2 + n\log k}$. Setting $n \log k =
\varepsilon^2 m/8$ results in $k \leq e^{O(m\varepsilon^2/n)}$. We could, for
instance, let $\varepsilon = 1/\sqrt{n}$, giving the bound in the theorem
statement (but there are other interesting choices of $\varepsilon$ too).
\end{proof}
\section{XOS Simultaneous Lower Bound}
Consider the following construction:
\begin{enumerate}
\item $S$ is drawn uniformly at random from sets of size $m/2$.
\item Let $a_1,\ldots, a_k$ be uniformly random sets of size $m/2$,
conditioned on $|S \cap a_i| = m/3$.
\item Alice's initial BXOS function has $A(X) = \max_i \{|a_i \cap X|\}$.
\item $T$ is drawn uniformly at random from sets of size $m/2$,
conditioned on $|S \cap T| = m/3$.
\item Let $b_1,\ldots, b_k$ be uniformly random sets of size $m/2$,
conditioned on $|T \cap a_i| = m/3$.
\item Bob's initial BXOS function has $B(X) = \max_i \{|b_i \cap X|\}$.
\begin{itemize}
\item Note: if we used BXOS evaluations $A$ and $B$,
the maximum achievable welfare would be at most $(3/4 - 1/108)m$
\end{itemize}
\item Let $Y$ be a uniformly random set
satisfying $|Y \cap S| = |\bar{Y} \cap T| = m/3$. \\
Let $Z$ be a uniformly
random set satisfying $|Z \cap S| = |Z \cap T| = m/3$.
\begin{enumerate}
\item With probability
$1/2$, set $v^A(X) = \max\{A(X),|Y \cap X|\}$ \\
and $v^B(X) = \max\{B(X), |\bar{Y} \cap X|\}$ \\
(in this case, the welfare shoots up to $m$).
\item With probability $1/2$ set $v^A(X) = \max\{A(X), |Z \cap X|\}$ \\
and $v^B(X) = \max\{B(X), |Z\cap X|\}$ \\
(in this case, the welfare remains approximately the same).
\end{enumerate}
\end{enumerate}
The following example sets give some intuition for why such sets $S,T,Y,Z$
should exist:
\begin{center}
\begin{tabular}{ c | c c c c c c }
& a & b & c & d & e & f \\
\hline
S: & 1 & 1 & 1 & & \\
T: & & 1 & 1 & 1 & \\
Y: & 1 & 1 & & & 1 \\
Z: & & 1 & 1 & & 1
\end{tabular}
\end{center}
\begin{theorem}[\cite{BravermanMW18}]
No simultaneous protocol with
subexponential in $m$ communication can beat a $(3/4-1/108)$ approximation on
the above construction with probability $> 2/3$ for decision.
\end{theorem}
We'll present the following steps, without proof.
\begin{enumerate}
\item For sufficiently small $k$, but $k = \exp(m)$,
the maximum welfare between $A$ and $B$ $(3/4-1/108)m$.
This is basically because the union of any two
sets constructed as in steps 1 thru 6 is $(3/4-1/108)m$.
\item If we do case (a) in the last step,
the maximum achievable welfare between $v^A$ and $v^B$ is $m$,
by giving $Y$ to Alice and $\bar{Y}$ to Bob.
\item In case (b), the maximum achievable welfare
between $v^A$ and $v^B$ is $(3/4-1/108)m$. This is essentially because the
additional sets are no better than another random set.
\item In either case, it is impossible for Alice to
distinguish which of her sets is this
special one (i.e. $Y$ in case (a) or $Z$ in case (b)).
This is because the conditioning related to $S$ is the same as for
each of the sets $a_i$, and only the conditioning related to $T$ is different.
Ditto for Bob.
\item Therefore, in order to possibly solve the
decision problem, Alice or Bob would have to convey some information about
every single one of their sets, which is exponential. (This can be made
into a formal statement, but we won't prove that in this class).
\end{enumerate}
% I'M NOT SURE IF THIS IS ACCURATE::
Observe that with two rounds, Alice could first send Bob the set $S$
(which is painfully obvious by her construction -- each $a_i$ intersects
$S$ way more than it should).
Bob could then determine which set he has that is ``special.''
(in case (a), his set $\bar{Y}$ should intersect $S$ way less than average,
where in case (b), his set $Z$ should intersect $S$ way more than average).
Also observe that for search, Alice will just pick a random $i$ and keep the
items in $a_i$ and give the rest to Bob. This will get the right welfare
approximation in both cases (it will just get different absolute welfare in
both).
\section{Subadditive Lower Bound for Two Players}
Consider the following
construction, which is well-defined for any collection $S$ of subsets: for a
given set $T$, say that $T$ is covered by collection
$\mathcal Z$ of sets if $T\subseteq \cup_{z \in\mathcal Z} z$.
Let $f_{S}(T)$ be the minimum number of sets in
$S$ that cover $T$.
\begin{lemma}
$f_S(\cdot)$ is subadditive for all $S$.
\end{lemma}
\begin{proof}
Let $Z$ cover $T$ and $Y$ cover $U$. Then clearly $Z \cup Y$ covers $T \cup
U$. Therefore, if $Z$ and $Y$ are the minimum covers of $T$ and $U$
respectively, we have that $f_S(T \cup U) \leq |Z|+|Y| = f_S(T) + f_S(U)$.
\end{proof}
Now, we wish to further modify $f_S(\cdot)$ into $f_S^\ell(\cdot)$ as follows:
\begin{itemize} \item If $f_S(X) < \ell/2$, $f_S^\ell(X) = f_S(X)$. \item If
$f_S(\bar{X}) < \ell/2$, $f_S^\ell(X) = \ell - f_S(\bar{X})$. \item If
neither, then $f_S^\ell(X) = \ell/2$. \end{itemize}
Observe that $f_S^\ell(\cdot)$ might not be well-defined (the first two bullets
might conflict). But for $S$ with the right properties, it is always
well-defined. Also observe that $f^\ell_S(X) + f^\ell_S(\bar{X}) = \ell$
for any $S$ and $X$.
\begin{definition}[$\ell$-sparse] We say that $\mathcal{S}$ is
\emph{$\ell$-sparse} if for all $T_1,\ldots, T_{\ell-1} \in \mathcal{S}$,
$\cup_j T_j \neq M$. \end{definition}
That is, $\mathcal{S}$ is $\ell$-sparse if there
does not exist $\ell-1$ elements of $\mathcal{S}$ whose union is the entire
ground set $M$.
\begin{lemma} If $\mathcal{S}$ is $\ell$-sparse, then
$f^\ell_\mathcal{S}(\cdot)$ is well-defined. \end{lemma}
\begin{proof} It is clear that $f^\ell_\mathcal{S}(X)$ is always defined at
least once. The only way in which $f^\ell_\mathcal{S}(X)$ could be defined
multiple times is if $f_\mathcal{S}(X) < \frac{\ell}{2}$ (in which case
$f^\ell_\mathcal{S}(X) = f_\mathcal{S}(X)$) and
$f_\mathcal{S}(\overline{X}) < \frac{\ell}{2}$ (in which case
$f^\ell_\mathcal{S}(X) = \ell - f_\mathcal{S}(\overline{X})$).
So assume for contradiction that both events hold, and let $X \subseteq
\cup_{i=1}^{\ell/2-1} T_i$, and $\overline{X} \subseteq \cup_{i=1}^{\ell/2-1}
U_i$, where each $T_i, U_i \in \mathcal{S}$
(such sets exist by the definition of $f_S$ and the assumption that
$f_\mathcal{S}(X), f_\mathcal{S}(\overline{X}) < \ell/2$).
But now consider that we can write $M = X \cup \overline{X}$ as a union
of $\leq \ell -2$ elements of $\mathcal{S}$, contradicting that $\mathcal{S}$ is
$\ell$-sparse.
\end{proof}
\begin{proposition}
If $\mathcal{S}$ is $\ell$-sparse, then
$f^\ell_\mathcal{S}(\cdot)$ is monotone and subadditive.
\end{proposition}
(Proof omitted). So now we have a class of subadditive functions. We now wish to
come up with a hard instance. To this effect, we need one further definition.
\begin{definition}[\cite{KleitmanS73}]
A collection $\mathcal{S} = \{S_1,\ldots,
S_k\}$ is \emph{$\ell$-independent} if
$\{T_1,\ldots, T_k\}$ is $\ell$-sparse
whenever $T_i \in \{S_i, \overline{S_i}\}$.
\end{definition}
\begin{proposition}\label{prop:lindependent}Let $\mathcal{S}$ be an
$\ell$-independent collection with $|\mathcal{S}| = k$. Then any deterministic
communication protocol that guarantees a $(\frac{1}{2} +
\frac{1}{2\ell-3})$-approximation to the optimal welfare for two monotone
subadditive bidders requires communication at least $k$. \end{proposition}
\begin{proof} Let $\mathcal{S} = \{S_1,\ldots, S_k\}$ be $\ell$-independent. For
each $i$, define $S_i^1:= S_i$, and $S_i^0 := \overline{S_i}$. Now, consider
an instance of EQUALITY where Alice is given $a$ and Bob is given $b$. Alice
will create the valuation function $f^\ell_\mathcal{A}$, where
$\mathcal{A}:=\{S_1^{a_1},\ldots, S_k^{a_k}\}$ (i.e. Alice builds
$\mathcal{A}$ by taking either $S_i^1$ or $S_i^0$, depending on $a_i$). Bob
will create the valuation function $f^\ell_\mathcal{B}$, where
$\mathcal{B}:=\{S_1^{b_1},\ldots, S_k^{b_k}\}$. Observe first that
$f^\ell_\mathcal{A}(\cdot)$ and $f^\ell_\mathcal{B}(\cdot)$ are indeed
well-defined, monotone, and subadditive as $\mathcal{S}$ is $\ell$-independent
(and therefore $\mathcal{A}$ and $\mathcal{B}$ are both $\ell$-sparse).
Observe that if $a = b$, then $\mathcal{A} = \mathcal{B}$ and moreover
$f^\ell_\mathcal{A}(\cdot) = f^\ell_\mathcal{B}(\cdot)$. Observe that for any
$S, \ell$ we have $f_S^\ell(X) + f_S^\ell(\bar{X}) =\ell$, so the maximum
achievable welfare is $\ell$.
On the other hand, if there exists an $i$ such that $a_i \neq b_i$ (without loss
of generality say that $a_i = 1$ and $b_i = 0$), we claim that welfare $2\ell-2$
is achievable. To see this, consider the allocation which awards
$\overline{S_i}$ to Alice and $S_i$ to Bob. Indeed,
$f_\mathcal{A}(S_i) = 1$ (as $S_i \in \mathcal{A}$),
so $f^\ell_\mathcal{A}(\overline{S_i}) = \ell - f_\mathcal A(S_i) = \ell - 1$.
Similarly, $f_B(\overline{S_i}) = 1$, so
$f_\mathcal B(S_i) = \ell - f_\mathcal B (\overline{S_i}) = \ell - 1$,
achieving total welfare $2(\ell-1)$.
(note\footnote{
As an aside, note that welfare exceeding $2\ell-2$ is not possible.
Proof: for all nonempty $X$, $f_S(X)$ is at
least $1$, and therefore $f^\ell_S(X)$ is at most $\ell-1$
by the second bullet point in the definition of $f^\ell_S$.
Therefore, the only way Alice or Bob could have value exceeding
$\ell-1$ is to get all of $M$, meaning that the other player receives value
$0$.
})
So assume for contradiction that a deterministic $\frac{1}{2} +
\frac{1}{2\ell-3} > \frac{\ell}{2\ell-2}$-approximation exists to the optimal
welfare for two monotone subadditive bidders with communication $< k$. Then such
a protocol would solve EQUALITY with communication $<k$ by the reduction above,
a contradiction. \end{proof}
\begin{lemma} \label{lem:l-indep-exist} For all $m$, $x > 1$, and $\ell =
\log_2(m) - \log_2(x)$, there exists a $\ell$-independent collection of subsets
of $[m]$ of size $k = e^{x/\ell}$. \end{lemma}
\section{Possible Project Topics}
\begin{itemize} \item See if the presented
subadditive construction ``separates'' XOS from subadditive in other cases
where good algorithms for subadditive aren't known but good ones for XOS
are (ask me for examples).
\end{itemize}
\bibliographystyle{alpha} \bibliography{../MasterBib}
\end{document}