Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Polynomial-time derivation of optimal k-tree topology from Markov networks (2404.05991v1)

Published 9 Apr 2024 in cs.DS and stat.ML

Abstract: Characterization of joint probability distribution for large networks of random variables remains a challenging task in data science. Probabilistic graph approximation with simple topologies has practically been resorted to; typically the tree topology makes joint probability computation much simpler and can be effective for statistical inference on insufficient data. However, to characterize network components where multiple variables cooperate closely to influence others, model topologies beyond a tree are needed, which unfortunately are infeasible to acquire. In particular, our previous work has related optimal approximation of Markov networks of tree-width k >=2 closely to the graph-theoretic problem of finding maximum spanning k-tree (MSkT), which is a provably intractable task. This paper investigates optimal approximation of Markov networks with k-tree topology that retains some designated underlying subgraph. Such a subgraph may encode certain background information that arises in scientific applications, for example, about a known significant pathway in gene networks or the indispensable backbone connectivity in the residue interaction graphs for a biomolecule 3D structure. In particular, it is proved that the \beta-retaining MSkT problem, for a number of classes \beta of graphs, admit O(n{k+1})-time algorithms for every fixed k>= 1. These \beta-retaining MSkT algorithms offer efficient solutions for approximation of Markov networks with k-tree topology in the situation where certain persistent information needs to be retained.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. F. R. Kschischang, B. J. Frey, and H. A. Loeliger, “Factor graphs and the sum-product algorithm,” IEEE Transactions on Information Theory, vol. 47, no. 2, pp. 498–519, 2001.
  2. D. M. Blei, A. Kucukelbir, and J. D. McAuliffe, “Variational inference: A review for statisticians,” Journal of the American Statistical Association, vol. 112, no. 518, pp. 859–877, 2017.
  3. A. Gionis, P. Indyk, and R. Motwani, “Similarity search in high dimensions via hashing,” in Proceedings of Conference on Vvery Large Databases, 1999, pp. 518—529.
  4. T. Mohseni Ahooyi, J. E. Arbogast, and M. Soroush, “Rolling pin method: efficient general method of joint probability modeling,” Industrial & Engineering Chemistry Research, vol. 53, no. 52, pp. 20 191–20 203, 2014.
  5. M. Melucci, “A brief survey on probability distribution approximation,” Computer Science Review, vol. 33, pp. 91–97, 2019.
  6. J. Jiao, Y. Han, and T. Weissman, “Beyond maximum likelihood: Boosting the chow-liu algorithm for large alphabets,” in The 50th Asilomar Conference on Signals, Systems and Computers, 2016, pp. 321–325.
  7. H. Chitsaz, R. Backofen, and S. C. Sahinalp, “biRNA: Fast RNA-RNA binding sites prediction,” in Proceedings of 9th International Workshop in Algorithms in Bioinformatics.   Springer, 2009, pp. 25–36.
  8. C. Chow and C. Liu, “Approximating discrete probability distributions with dependence trees,” IEEE transactions on Information Theory, vol. 14, no. 3, pp. 462–467, 1968.
  9. E. Kovács and T. Szántai, “On the approximation of a discrete multivariate probability distribution using the new concept of t-cherry junction tree,” in Coping with Uncertainty: Robust Solutions.   Springer, 2009, pp. 39–56.
  10. J. Suzuki, “A novel Chow–Liu algorithm and its application to gene differential analysis,” International Journal of Approximate Reasoning, vol. 80, pp. 1–18, 2017.
  11. L. V. Montiel and J. E. Bickel, “Approximating joint probability distributions given partial information,” Decision Analysis, vol. 10, no. 1, pp. 26–41, 2013.
  12. N. Hammami and M. Bedda, “Improved tree model for Arabic speech recognition,” in The 3rd International Conference on Computer Science and Information Technology, vol. 5, 2010, pp. 521–526.
  13. K. Huang, I. King, and M. Lyu, “Constructing a large node Chow-Liu tree based on frequent itemsets,” in Proceedings of the 9th International Conference on Neural Information Processing,, vol. 1, 2002, pp. 498–502 vol.1.
  14. B. M. King and B. Tidor, “Mist: Maximum information spanning trees for dimension reduction of biological data sets,” Bioinformatics, vol. 25, no. 9, pp. 1165–1172, 2009.
  15. S. Kullback and R. Leibler, “On information and sufficiency,” Annals of Mathematical Statistics, vol. 22, no. 1, pp. 79–86, 1951.
  16. E. Boix-Adserà, G. Bresler, and F. Koehler, “Chow-Liu++: Optimal prediction-centric learning of tree ising models,” in IEEE 62nd Annual Symposium on Foundations of Computer Science, 2022, pp. 417–426.
  17. V. Y. Tan, S. Sanghavi, J. W. Fisher, and A. S. Willsky, “Learning graphical models for hypothesis testing and classification,” IEEE Transactions on Signal Processing, vol. 58, no. 11, pp. 5481–5495, 2010.
  18. X. Zhang, D. Chang, W. Qi, and Z. Zhan, “Tree-like dimensionality reduction for cancer-informatics,” in IOP Conference Series: Materials Science and Engineering, vol. 490, no. 4.   IOP Publishing, 2019, p. 042028.
  19. J. Jiao, Y. Han, and T. Weissman, “Beyond maximum likelihood: Boosting the chow-liu algorithm for large alphabets,” in 2016 50th Asilomar Conference on Signals, Systems and Computers.   IEEE, 2016, pp. 321–325.
  20. C. Chan and T. Liu, “Clustering by multivariate mutual information under chow-liu tree approximation,” in 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton), 2015, pp. 993–999.
  21. A. Steimer, F. Zubler, and K. Schindler, “Chow–Liu trees are sufficient predictive models for reproducing key features of functional networks of periictal eeg time-series,” NeuroImage, vol. 118, pp. 520–537, 2015.
  22. M. J. Wainwright, M. I. Jordan et al., “Graphical models, exponential families, and variational inference,” Foundations and Trends in Machine Learning, vol. 1, no. 1–2, pp. 1–305, 2008.
  23. D. R. Karger and N. Srebro, “Learning markov networks: maximum bounded tree-width graphs.” in SODA, vol. 1, 2001, pp. 392–401.
  24. G. Bresler, “Efficiently learning ising models on arbitrary graphs,” in Proceedings of the forty-seventh annual ACM symposium on Theory of computing, 2015, pp. 771–782.
  25. A. Singh and M. D. Humphries, “Finding communities in sparse networks,” Nature Scientific Reports, vol. 5, no. 1, p. 8828, 2015.
  26. C. Gunduz, B. Yener, and S. H. Gultekin, “The cell graphs of cancer,” Bioinformatics, vol. 20, no. suppl_1, pp. i145–i151, 2004.
  27. M. D. Humphries and K. Gurney, “Network ‘small-world-ness’: a quantitative method for determining canonical network equivalence,” PloS one, vol. 3, no. 4, p. e0002051, 2008.
  28. M. E. Newman, “The structure and function of complex networks,” SIAM review, vol. 45, no. 2, pp. 167–256, 2003.
  29. H. Patil, “On the structure of k-trees,” Journal of Combinatorics, Information and System Sciences, vol. 11, no. 2-4, pp. 57–64, 1986.
  30. N. Robertson and P. D. Seymour, “Graph minors. i. excluding a forest,” Journal of Combinatorial Theory, Series B, vol. 35, no. 1, pp. 39–61, 1983.
  31. H. L. Bodlaender, “A partial k-arboretum of graphs with bounded treewidth,” Theoretical Computer Science, vol. 209, no. 1, pp. 1–45, 1998.
  32. N. Srebro, “Maximum likelihood bounded tree-width markov networks,” Artificial Intelligence, vol. 143, pp. 123–138, 2003.
  33. T. Szántai and K. E., “Hypergraphs as a mean of discovering the dependence structure of a discrete multivariate probability distribution,” Annals of Operations Research, vol. 193, pp. 71–90, 2012.
  34. D. Chang, L. Ding, R. Malmberg, D. Robinson, M. Wicker, H. Yan, A. Martinez, and L. Cai, “Optimal learning of markov k-tree topology,” Journal of Computational Mathematics and Data Science, vol. 4, p. 100046, 2022.
  35. L. Cai and F. Maffray, “On the sprnning k-tree problem,” Discrete Applied Mathematics, vol. 44, no. 1-3, pp. 139–156, 1993.
  36. L. Lee, “Fast context-free grammar parsing requires fast boolean matrix multiplication,” Journal of ACM, vol. 49, no. 1, pp. 1–15, 2002.
  37. A. Abboud, A. Backurs, and V. V. Williams, “If the current clique algorithms are optimal, so is valiant’s parser,” SIAM Journal of Computing, vol. 47, no. 6, 2018.
  38. G. R. Grimmett, “A theorem about random fields,” Bulletin of the London Mathematical Society, vol. 5, no. 1, pp. 81–84, 1973.
  39. L. Ding, “Maximum spanning k-trees: models and applications,” PhD Dissertation, University of Georgia, 2016.
  40. P. Chatterjee and N. R. Pal, “Discovery of synergistic genetic network: A minimum spanning tree-based approach,” Journal of Bioinformatics and Computational Biology, vol. 14, no. 1, 2016.

Summary

We haven't generated a summary for this paper yet.