Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu

APA

(2020). Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu. The Simons Institute for the Theory of Computing. https://simons.berkeley.edu/talks/tbd-258

MLA

Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu. The Simons Institute for the Theory of Computing, Dec. 17, 2020, https://simons.berkeley.edu/talks/tbd-258

BibTex

          @misc{ scivideos_16883,
            doi = {},
            url = {https://simons.berkeley.edu/talks/tbd-258},
            author = {},
            keywords = {},
            language = {en},
            title = {Near-Optimal Learning of Tree-Structured Distributions by Chow-Liu},
            publisher = {The Simons Institute for the Theory of Computing},
            year = {2020},
            month = {dec},
            note = {16883 see, \url{https://scivideos.org/Simons-Institute/16883}}
          }
          
Eric Price (University of Texas, Austin)
Source Repository Simons Institute

Abstract

The classical Chow-Liu algorithm estimates a tree-structured graphical model of a distribution using the maximum spanning tree on the pairwise mutual information graph.  We show that, if the true distribution P is tree-structured over a discrete domain Σ^n, then the Chow-Liu algorithm returns a distribution Q with D(P||Q) < ε after O~(Σ^3 n / ε) samples, which is nearly optimal in n and ε. Our analysis of Chow-Liu is based on a new result in conditional independence testing: we prove that for three random variables X,Y,Z each over Σ, testing if I(X;Y∣Z) is 0 or ≥ε is possible with O˜(Σ^3/ε) samples using the empirical mutual information.