Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
GPT-4o
Gemini 2.5 Pro Pro
o3 Pro
GPT-4.1 Pro
DeepSeek R1 via Azure Pro
2000 character limit reached

Binder: Hierarchical Concept Representation through Order Embedding of Binary Vectors (2404.10924v1)

Published 16 Apr 2024 in cs.CL and cs.AI

Abstract: For natural language understanding and generation, embedding concepts using an order-based representation is an essential task. Unlike traditional point vector based representation, an order-based representation imposes geometric constraints on the representation vectors for explicitly capturing various semantic relationships that may exist between a pair of concepts. In existing literature, several approaches on order-based embedding have been proposed, mostly focusing on capturing hierarchical relationships; examples include vectors in Euclidean space, complex, Hyperbolic, order, and Box Embedding. Box embedding creates region-based rich representation of concepts, but along the process it sacrifices simplicity, requiring a custom-made optimization scheme for learning the representation. Hyperbolic embedding improves embedding quality by exploiting the ever-expanding property of Hyperbolic space, but it also suffers from the same fate as box embedding as gradient descent like optimization is not simple in the Hyperbolic space. In this work, we propose Binder, a novel approach for order-based representation. Binder uses binary vectors for embedding, so the embedding vectors are compact with an order of magnitude smaller footprint than other methods. Binder uses a simple and efficient optimization scheme for learning representation vectors with a linear time complexity. Our comprehensive experimental results show that Binder is very accurate, yielding competitive results on the representation task. But Binder stands out from its competitors on the transitive closure link prediction task as it can learn concept embeddings just from the direct edges, whereas all existing order-based approaches rely on the indirect edges.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. DBpedia: A Nucleus for a Web of Open Data. In The Semantic Web (Lecture Notes in Computer Science), Karl Aberer, Key-Sun Choi, Natasha Noy, Dean Allemang, Kyung-Il Lee, Lyndon Nixon, Jennifer Golbeck, Peter Mika, Diana Maynard, Riichiro Mizoguchi, Guus Schreiber, and Philippe Cudré-Mauroux (Eds.). Springer, Berlin, Heidelberg, 722–735. https://doi.org/10.1007/978-3-540-76298-0_52
  2. Combining Constraint Programming and Local Search for Job-Shop Scheduling. INFORMS Journal on Computing 23, 1 (Feb. 2011), 1–14. https://doi.org/10.1287/ijoc.1100.0388
  3. Robi Bhattacharjee and Sanjoy Dasgupta. 2023. What relations are reliably embeddable in Euclidean space? https://doi.org/10.48550/arXiv.1903.05347 arXiv:1903.05347 [cs, stat].
  4. Capacity and Bias of Learned Geometric Embeddings for Directed Graphs. In Advances in Neural Information Processing Systems, Vol. 34. Curran Associates, Inc., 16423–16436. https://proceedings.neurips.cc/paper/2021/hash/88d25099b103efd638163ecb40a55589-Abstract.html
  5. Translating embeddings for modeling multi-relational data. In Proceedings of the 26th International Conference on Neural Information Processing Systems - Volume 2 (NIPS’13). Curran Associates Inc., Red Hook, NY, USA, 2787–2795.
  6. Introduction to Algorithms, fourth edition. MIT Press. Google-Books-ID: RSMuEAAAQBAJ.
  7. CFO: Conditional Focused Neural Question Answering with Large-scale Knowledge Bases. https://doi.org/10.48550/arXiv.1606.01994 arXiv:1606.01994 [cs].
  8. Improving Local Identifiability in Probabilistic Box Embeddings. NeurIPS 2020 (Virtual) (Oct. 2020). https://doi.org/10.48550/arXiv.2010.04831
  9. Box-To-Box Transformations for Modeling Joint Hierarchies. In Proceedings of the 6th Workshop on Representation Learning for NLP (RepL4NLP-2021). Association for Computational Linguistics, Online, 277–288. https://doi.org/10.18653/v1/2021.repl4nlp-1.28
  10. FCA2VEC: Embedding Techniques for Formal Concept Analysis. https://doi.org/10.48550/arXiv.1911.11496 arXiv:1911.11496 [cs, stat].
  11. Christiane Fellbaum. 2012. WordNet. In The Encyclopedia of Applied Linguistics, Carol Chapelle (Ed.). John Wiley & Sons, Inc., Hoboken, NJ, USA, wbeal1285. https://doi.org/10.1002/9781405198431.wbeal1285
  12. AMIE: association rule mining under incomplete evidence in ontological knowledge bases. In Proceedings of the 22nd international conference on World Wide Web (WWW ’13). Association for Computing Machinery, New York, NY, USA, 413–422. https://doi.org/10.1145/2488388.2488425
  13. Hyperbolic Entailment Cones for Learning Hierarchical Embeddings. International Conference on Machine Learning (ICML) Stockholm, Sweden. (April 2018). https://doi.org/10.48550/arXiv.1804.01882
  14. Bernhard Ganter and Rudolf Wille. 1999. Formal Concept Analysis. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-59830-2
  15. Aditya Grover and Jure Leskovec. 2016. node2vec: Scalable Feature Learning for Networks. (July 2016). https://doi.org/10.48550/arXiv.1607.00653
  16. Learning to Exploit Long-term Relational Dependencies in Knowledge Graphs. https://doi.org/10.48550/arXiv.1905.04914 arXiv:1905.04914 [cs].
  17. Mohammad Al Hasan and Mohammed J. Zaki. 2011. A Survey of Link Prediction in Social Networks. In Social Network Data Analytics, Charu C. Aggarwal (Ed.). Springer US, Boston, MA, 243–275. https://doi.org/10.1007/978-1-4419-8462-3_9
  18. A Greedy Bit-flip Training Algorithm for Binarized Knowledge Graph Embeddings. In Findings of the Association for Computational Linguistics: EMNLP 2020, Trevor Cohn, Yulan He, and Yang Liu (Eds.). Association for Computational Linguistics, Online, 109–114. https://doi.org/10.18653/v1/2020.findings-emnlp.10
  19. Learning to Represent Knowledge Graphs with Gaussian Embedding. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management (CIKM ’15). Association for Computing Machinery, New York, NY, USA, 623–632. https://doi.org/10.1145/2806416.2806502
  20. Andrej Karpathy and Li Fei-Fei. 2015. Deep Visual-Semantic Alignments for Generating Image Descriptions. https://doi.org/10.48550/arXiv.1412.2306 arXiv:1412.2306 [cs].
  21. Alice Lai and Julia Hockenmaier. 2017. Learning to Predict Denotational Probabilities For Modeling Entailment. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, Mirella Lapata, Phil Blunsom, and Alexander Koller (Eds.). Association for Computational Linguistics, Valencia, Spain, 721–730. https://aclanthology.org/E17-1068
  22. Random Walk Inference and Learning in A Large Scale Knowledge Base. In Proceedings of the 2011 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, Edinburgh, Scotland, UK., 529–539. https://aclanthology.org/D11-1049
  23. Lorentzian Distance Learning for Hyperbolic Representations. In Proceedings of the 36th International Conference on Machine Learning. PMLR, 3672–3681. https://proceedings.mlr.press/v97/law19a.html
  24. Smoothing the Geometry of Probabilistic Box Embeddings. https://openreview.net/forum?id=H1xSNiRcF7
  25. Discrete Knowledge Graph Embedding based on Discrete Optimization. https://doi.org/10.48550/arXiv.2101.04817 arXiv:2101.04817 [cs].
  26. Learning Entity and Relation Embeddings for Knowledge Graph Completion. Proceedings of the AAAI Conference on Artificial Intelligence 29, 1 (Feb. 2015). https://doi.org/10.1609/aaai.v29i1.9491
  27. PATTY: a taxonomy of relational patterns with semantic types. In Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL ’12). Association for Computational Linguistics, USA, 1135–1145.
  28. Compositional Vector Space Models for Knowledge Base Completion. https://doi.org/10.48550/arXiv.1504.06662 arXiv:1504.06662 [cs, stat].
  29. A Novel Embedding Model for Knowledge Base Completion Based on Convolutional Neural Network. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers). 327–333. https://doi.org/10.18653/v1/N18-2053 arXiv:1712.02121 [cs].
  30. Maximillian Nickel and Douwe Kiela. 2017. Poincaré Embeddings for Learning Hierarchical Representations. In Advances in Neural Information Processing Systems, Vol. 30. Curran Associates, Inc. https://papers.nips.cc/paper/2017/hash/59dfa2df42d9e3d41f5b02bfc32229dd-Abstract.html
  31. Maximilian Nickel and Douwe Kiela. 2018. Learning Continuous Hierarchies in the Lorentz Model of Hyperbolic Geometry. https://doi.org/10.48550/arXiv.1806.03417 arXiv:1806.03417 [cs, stat].
  32. A three-way model for collective learning on multi-relational data. In Proceedings of the 28th International Conference on International Conference on Machine Learning (ICML’11). Omnipress, Madison, WI, USA, 809–816.
  33. DeepWalk: Online Learning of Social Representations. (March 2014). https://doi.org/10.1145/2623330.2623732
  34. Query2box: Reasoning over Knowledge Graphs in Vector Space using Box Embeddings. https://doi.org/10.48550/arXiv.2002.05969 arXiv:2002.05969 [cs, stat].
  35. Sebastian Rudolph. 2007. Using FCA for Encoding Closure Operators into Neural Networks. In Proceedings of the 15th international conference on Conceptual Structures: Knowledge Architectures for Smart Applications (ICCS ’07). Springer-Verlag, Berlin, Heidelberg, 321–332. https://doi.org/10.1007/978-3-540-73681-3_24
  36. Modeling Relational Data with Graph Convolutional Networks. https://doi.org/10.48550/arXiv.1703.06103 arXiv:1703.06103 [cs, stat].
  37. End-to-end Structure-Aware Convolutional Networks for Knowledge Base Completion. https://doi.org/10.48550/arXiv.1811.04441 arXiv:1811.04441 [cs].
  38. Improving Hypernymy Detection with an Integrated Path-based and Distributional Method. https://doi.org/10.48550/arXiv.1603.06076 arXiv:1603.06076 [cs].
  39. Yago: a core of semantic knowledge. In Proceedings of the 16th international conference on World Wide Web (WWW ’07). Association for Computing Machinery, New York, NY, USA, 697–706. https://doi.org/10.1145/1242572.1242667
  40. RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space. https://doi.org/10.48550/arXiv.1902.10197 arXiv:1902.10197 [cs, stat].
  41. LINE: Large-scale Information Network Embedding. (March 2015). https://doi.org/10.1145/2736277.2741093
  42. Complex Embeddings for Simple Link Prediction. https://doi.org/10.48550/arXiv.1606.06357 arXiv:1606.06357 [cs, stat].
  43. Order-Embeddings of Images and Language. (Nov. 2015). https://doi.org/10.48550/arXiv.1511.06361
  44. Probabilistic Embedding of Knowledge Graphs with Box Lattice Measures. http://arxiv.org/abs/1805.06627 arXiv:1805.06627 [cs, stat].
  45. Show and Tell: A Neural Image Caption Generator. https://doi.org/10.48550/arXiv.1411.4555 arXiv:1411.4555 [cs].
  46. Denny Vrandecic. 2012. Wikidata: a new platform for collaborative data collection. In Proceedings of the 21st World Wide Web Conference, WWW 2012, Lyon, France, April 16-20, 2012 (Companion Volume), Alain Mille, Fabien Gandon, Jacques Misselis, Michael Rabinovich, and Steffen Staab (Eds.). ACM, 1063–1064. https://doi.org/10.1145/2187980.2188242
  47. Learning to Hash for Efficient Search Over Incomplete Knowledge Graphs. In 2019 IEEE International Conference on Data Mining (ICDM). IEEE, Beijing, China, 1360–1365. https://doi.org/10.1109/ICDM.2019.00174
  48. Knowledge Graph Embedding by Translating on Hyperplanes. Proceedings of the AAAI Conference on Artificial Intelligence 28, 1 (June 2014). https://doi.org/10.1609/aaai.v28i1.8870
  49. Probase: a probabilistic taxonomy for text understanding. In Proceedings of the 2012 ACM SIGMOD International Conference on Management of Data (SIGMOD ’12). Association for Computing Machinery, New York, NY, USA, 481–492. https://doi.org/10.1145/2213836.2213891
  50. KG-BERT: BERT for Knowledge Graph Completion. https://doi.org/10.48550/arXiv.1909.03193 arXiv:1909.03193 [cs].
  51. Modeling Transitivity and Cyclicity in Directed Graphs via Binary Code Box Embeddings. Advances in Neural Information Processing Systems 35 (Dec. 2022), 10587–10599. https://papers.nips.cc/paper_files/paper/2022/hash/44a1f18afd6d5cc34d7e5c3d8a80f63b-Abstract-Conference.html
  52. Jinjie Zhang and Rayan Saab. 2021. Faster Binary Embeddings for Preserving Euclidean Distances. https://doi.org/10.48550/arXiv.2010.00712 arXiv:2010.00712 [cs, math, stat].
Citations (1)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.