par Fontes, Lila;Jain, Rahul;Kerenidis, Iordanis;Laplante, Sophie;Lauri`ere, Mathieu;Roland, Jérémie
Référence 42nd International Colloquium on Automata, Languages and Programming (ICALP'15), page (506-516)
Publication Publié, 2015
Publication dans des actes
Résumé : Does the information complexity of a function equal its communication complexity? We examine whether any currently known techniques might be used to show a separation between the two notions. Ganor et al. recently provided such a separation in the distributional case for a specific input distribution. We show that in the non-distributional setting, the relative discrepancy bound is smaller than the information complexity, hence it cannot separate information and communication complexity. In addition, in the distributional case, we provide a linear program formulation for relative discrepancy and relate it to variants of the partition bound, resolving also an open question regarding the relation of the partition bound and information complexity. Last, we prove the equivalence between the adaptive relative discrepancy and the public-coin partition, implying that the logarithm of the adaptive relative discrepancy bound is quadratically tight with respect to communication.