Multiparty communication complexity
In the 2–party communication game, introduced in [1] in 1979, two players, P1 and P2 attempt to compute a Boolean function
Player P1 knows the value of x2, P2 knows the value of x1, but Pi does not know the value of xi, for i=1,2.
In other words, the players know the other's variables, but not their own. The minimum number of bits that must be communicated by the players to compute f is the communication complexity of f, denoted by κ(f).
The multiparty communication game, defined in 1983 [2], it is a powerful generalization of the 2–party case: Here the players know all the others' input, except their own. Because of this property, sometimes this model is called "numbers on the foreheadˇ model, since if the players were seated around a roundtable, each wearing their own input on the forehead, then every player would see all the others' input, except their own.
The formal definition is as follows: k players: P1,P2,...,Pk intend to compute a Boolean function
On set S={x1,x2,...,xn} of variables there is a fixed partition A of k classes A1,A2,...,Ak, and player P1 knows every variable, except those in Ai, for i=1,2,...,k. The players have unlimited computational power, and they communicate with the help of a blackboard, viewed by all players.
The aim is to compute f(x1,x2,...,xn), such that at the end of the computation, every player knows this value. The cost of the computation is the number of bits written onto the blackboard for the given input x=(x1,x2,...,xn) and partition A=(A1,A2,...,Ak). The cost of a multiparty protocol is the maximum number of bits communicated for any x from the set {0,1}n and the given partition A. The k-party communication complexity, C(k)A(f), of a function f, with respect to partition A, is the minimum of costs of those k-party protocols which compute f. The k-party symmetric communication complexity of f is defined as
where the maximum is taken over all k-partitions of set x=(x1,x2,...,xn).
Upper and lower bounds
For a general upper bound both for two and more players, let us suppose that A1 is one of the smallest classes of of the partition A1,A2,...,Ak. Then P1 can compute any Boolean function of S with |A1|+1 bits of communication: P2 writes down the |A1| bits of A1 on the blackboard, P1 reads it, and computes and announces the value f(x). So, we can write:
The Generalized Inner Product function (GIP)[3] is defined as follows: Let y1,y2,...,yk be n-bit vectors, and let Y be the n times k matrix, with k columns as the y1,y2,...,yk vectors. Then GIP(y1,y2,...,yk) is the number of the all-1 rows of matrix Y, taken modulo 2. In other words, if the vectors y1,y2,...,yk correspond to the characteristic vectors of k subsets of an n element base-set, then GIP corresponds to the parity of the intersection of these k subsets.
It was shown[3] that
with a constant c>0.
An upper bound to the multiparty communication complexity of GIP shows[4] that
with a constant c>0.
For a general function f, one can bound the multiparty communication complexity of f by using its L1 norm[5] as follows[6]
Multiparty communication complexity and pseudorandom generators
A construction of a pseudorandom number generator was based on the BNS lower bound for the GIP function[3].
Circuits and multiparty communication complexity
Open problems
- ^ A. C. YAO, Some complexity questions related to distributed computing, in Proc. 11th Ann. ACM Symp. Theor. Comput., 1979, pp. 209–213.
- ^ A. K. CHANDRA, M. L. FURST, AND R. J. LIPTON, Multi-party protocols, in Proc. 15th Ann. ACM Symp. Theor. Comput., 1983, pp. 94–99.
- ^ a b c L. BABAI, N. NISAN, AND M. SZEGEDY, Multiparty protocols, pseudorandom generators for logspace, and time-space trade-offs, J. Comput. System Sci., 45 (1992), pp. 204–232.
- ^ V. GROLMUSZ: The BNS lower bound for multi-party protocols is nearly optimal, Inform. and Comput., 112 (1994), pp. 51–54.
- ^ J. BRUCK AND R. SMOLENSKY, Polynomial threshold functions, AC0 functions and spectral norms, in Proc. 32nd Ann. IEEE Symp. Found. Comput. Sci., 1991, pp. 632–641
- ^ V. GROLMUSZ: Harmonic Analysis, Real Approximation, and the Communication Complexity of Boolean Functions, Algorithmica Vol. 23. (1999) pp. 341-353