Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Principled and Efficient Motif Finding for Structure Learning of Lifted Graphical Models (2302.04599v3)

Published 9 Feb 2023 in cs.AI

Abstract: Structure learning is a core problem in AI central to the fields of neuro-symbolic AI and statistical relational learning. It consists in automatically learning a logical theory from data. The basis for structure learning is mining repeating patterns in the data, known as structural motifs. Finding these patterns reduces the exponential search space and therefore guides the learning of formulas. Despite the importance of motif learning, it is still not well understood. We present the first principled approach for mining structural motifs in lifted graphical models, languages that blend first-order logic with probabilistic models, which uses a stochastic process to measure the similarity of entities in the data. Our first contribution is an algorithm, which depends on two intuitive hyperparameters: one controlling the uncertainty in the entity similarity measure, and one controlling the softness of the resulting rules. Our second contribution is a preprocessing step where we perform hierarchical clustering on the data to reduce the search space to the most relevant data. Our third contribution is to introduce an O(n ln n) (in the size of the entities in the data) algorithm for clustering structurally-related data. We evaluate our approach using standard benchmarks and show that we outperform state-of-the-art structure learning approaches by up to 6% in terms of accuracy and up to 80% in terms of runtime.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (45)
  1. Hinge-loss markov random fields and probabilistic soft logic. Journal of Machine Learning Research.
  2. Belevitch, V. 1959. On the statistical laws of linguistic distributions. Annales de la Société Scientifique de Bruxelles, 73: 310–326.
  3. Cheeger’s cut, maxcut and the spectral theory of 1-Laplacian on graphs. Science China Mathematics, 60(11): 1963–1980.
  4. Knowledge expansion over probabilistic knowledge bases. In Proceedings of the 2014 ACM SIGMOD International Conference on Management of Data, 649–660. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-4503-2376-5.
  5. TensorLog: A Probabilistic Database Implemented Using Deep-Learning Infrastructure. J. Artif. Intell. Res., 67: 285–325.
  6. Investigating markov logic networks for collective classification. Proceedings of the 4th International Conference on Agents and Artificial Intelligence, 1: 5–15.
  7. Go for a Walk and Arrive at the Answer: Reasoning Over Paths in Knowledge Bases using Reinforcement Learning. CoRR, abs/1711.05851.
  8. Davies, R. B. 1973. Numerical inversion of a characteristic function. Biometrika, 60(2): 415–417.
  9. Learning Explanatory Rules from Noisy Data. J. Artif. Intell. Res., 61: 1–64.
  10. Forgy, E. 1965. Cluster analysis of multivariate data : efficiency versus interpretability of classifications. Biometrics, 21: 768–780.
  11. Friedman, J. H. 2000. Greedy Function Approximation: A Gradient Boosting Machine. Annals of Statistics, 29: 1189–1232.
  12. Introduction to Statistical Relational Learning (Adaptive Computation and Machine Learning). The MIT Press.
  13. Traversing Knowledge Graphs in Vector Space. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, 318–327. Lisbon, Portugal: Association for Computational Linguistics.
  14. Ha; Rowe; Mott; and Lester. 2011. Goal Recognition with Markov Logic Networks for Player-Adaptive Games. Proceedings of the Seventh AAAI Conference on Artificial Intelligence and Interactive Digital Entertainment (AIIDE-11).
  15. Harnessing Deep Neural Networks with Logic Rules. In ACL, 2410–2420.
  16. Johnson, N. L. 1960. An Approximation to the Multinomial Distribution: Some Properties and Applications. Biometrika, 47(1): 93–102.
  17. Gradient-based boosting for statistical relational learning: the Markov logic network and missing data cases. Machine Learning, 100(1): 75–100.
  18. Learning the structure of Markov logic networks. In Proceedings of the 22nd international conference on Machine learning - ICML ’05, 441–448. Bonn, Germany: ACM Press. ISBN 978-1-59593-180-1.
  19. Learning Markov Logic Networks Using Structural Motifs. AAAIWS: Proceedings of the 6th AAAI Conference on Statistical Relational Artificial Intelligence.
  20. The alchemy system for statistical relational AI: User Manual.
  21. Prediction of protein β𝛽\betaitalic_β-residue contacts by Markov logic networks with grounding-specific weights. Bioinformatics, 25(18): 2326–2333.
  22. DeepProbLog: Neural Probabilistic Logic Programming. In NeurIPS, 3749–3759.
  23. Quadratic forms in random variables: theory and applications. Number v. 126 in Statistics, textbooks and monographs. M. Dekker. ISBN 978-0-8247-8691-5.
  24. Bottom-up learning of Markov logic network structure. In Proceedings of the 24th international conference on Machine learning - ICML ’07, 625–632. Corvalis, Oregon: ACM Press. ISBN 978-1-59593-793-3.
  25. Towards Neural Theorem Proving at Scale. CoRR, abs/1807.08204.
  26. The distribution function of a linear combination of chi-squares. Computers & Mathematics with Applications, 10(4): 383–386.
  27. Muggleton, S. H. 1995. Inverse Entailment and Progol. New Gener. Comput., 13(3&4): 245–286.
  28. Popoviciu, T. 1935. Sur les équations algébriques ayant toutes leurs racines réelles. Mathematica, 9(129): 20.
  29. Quinlan, J. R. 1990. Learning Logical Definitions from Relations. Mach. Learn., 5(3): 239–266.
  30. Markov logic networks. Machine Learning, 62(1-2): 107–136.
  31. Collective semantic role labelling with Markov logic. In Proceedings of the Twelfth Conference on Computational Natural Language Learning, CoNLL ’08, 193–197. USA: Association for Computational Linguistics. ISBN 978-1-905593-48-4.
  32. End-to-end Differentiable Proving. In Guyon, I.; Luxburg, U. V.; Bengio, S.; Wallach, H.; Fergus, R.; Vishwanathan, S.; and Garnett, R., eds., Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  33. Russell, S. 2015. Unifying logic and probability. Communications of the ACM, 58(7): 88–97.
  34. DRUM: End-To-End Differentiable Rule Mining On Knowledge Graphs. In Wallach, H. M.; Larochelle, H.; Beygelzimer, A.; d’Alché-Buc, F.; Fox, E. B.; and Garnett, R., eds., Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, 15321–15331.
  35. Fast Incremental Proximity Search in Large Graphs. Proceedings of the 25th International Conference on Machine Learning, 896–903.
  36. Satterthwaite, F. E. 1946. An Approximate Distribution of Estimates of Variance Components. Biometrics Bulletin, 2(6): 110–114.
  37. Statistics for Engineers: An Introduction with Examples from Practice. Wiesbaden Heidelberg: Springer, 1st ed. 2021 edition edition. ISBN 978-3-658-32396-7.
  38. Best-effort inductive logic programming via fine-grained cost-based hypothesis generation - The inspire system at the inductive logic programming competition. Mach. Learn., 107(7): 1141–1169.
  39. Entity Resolution with Markov Logic. In Proceedings of the Sixth International Conference on Data Mining, ICDM ’06, 572–582. USA: IEEE Computer Society. ISBN 978-0-7695-2701-7.
  40. Deep Probabilistic Logic: A Unifying Framework for Indirect Supervision. In EMNLP, 1891–1902.
  41. Welch, B. L. 1938. The Significance of the Difference Between Two Means when the Population Variances are Unequal. Biometrika, 29(3): 350–362.
  42. Automatically refining the wikipedia infobox ontology. In Proceedings of the 17th international conference on World Wide Web, WWW ’08, 635–644. New York, NY, USA: Association for Computing Machinery. ISBN 978-1-60558-085-2.
  43. Differentiable Learning of Logical Rules for Knowledge Base Reasoning. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, 2316–2325. Red Hook, NY, USA: Curran Associates Inc. ISBN 9781510860964.
  44. NeurASP: Embracing Neural Networks into Answer Set Programming. In IJCAI, 1755–1762.
  45. BIRCH: an efficient data clustering method for very large databases. ACM SIGMOD Record, 25(2): 103–114.
Citations (3)

Summary

We haven't generated a summary for this paper yet.