Communication-Constrained Hypothesis Testing: Optimality, Robustness, and Reverse Data Processing Inequalities

IEEE TRANSACTIONS ON INFORMATION THEORY(2024)

引用 0|浏览30
暂无评分
摘要
We study hypothesis testing under communication constraints, where each sample is quantized before being revealed to a statistician. Without communication constraints, it is well known that the sample complexity of simple binary hypothesis testing is characterized by the Hellinger distance between the distributions. We show that the sample complexity of simple binary hypothesis testing under communication constraints is at most a logarithmic factor larger than in the unconstrained setting and this bound is tight. We develop a polynomial-time algorithm that achieves the aforementioned sample complexity. Our framework extends to robust hypothesis testing, where the distributions are corrupted in the total variation distance. Our proofs rely on a new reverse data processing inequality and a reverse Markov inequality, which may be of independent interest. For simple M-ary hypothesis testing, the sample complexity in the absence of communication constraints has a logarithmic dependence on M. We show that communication constraints can cause an exponential blow-up leading to Omega(M) sample complexity even for adaptive algorithms.
更多
查看译文
关键词
Hypothesis testing,Hellinger distance,data processing inequality,Markov inequality,quantization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要