Weakly Supervised Reasoning by Neuro-Symbolic Approaches (2309.13072v1)
Abstract: Deep learning has largely improved the performance of various NLP tasks. However, most deep learning models are black-box machinery, and lack explicit interpretation. In this chapter, we will introduce our recent progress on neuro-symbolic approaches to NLP, which combines different schools of AI, namely, symbolism and connectionism. Generally, we will design a neural system with symbolic latent structures for an NLP task, and apply reinforcement learning or its relaxation to perform weakly supervised reasoning in the downstream task. Our framework has been successfully applied to various tasks, including table query reasoning, syntactic structure reasoning, information extraction reasoning, and rule reasoning. For each application, we will introduce the background, our approach, and experimental results.
- Towards string-to-tree neural machine translation. In ACL, pages 132–140, 2017.
- The Description Logic Handbook: Theory, Implementation and Applications. Cambridge University Press, 2003.
- Neural machine translation by jointly learning to align and translate. In ICLR, 2015.
- Measuring abstract reasoning in neural networks. In ICML, pages 511–520, 2018.
- Learning to paraphrase: An unsupervised approach using multiple-sequence alignment. In ACL, pages 16–23, 2003.
- Semantic parsing on freebase from question-answer pairs. In EMNLP, pages 1533–1544, 2013.
- Semantic parsing via paraphrasing. In ACL, pages 1415–1425, 2014.
- John A Bernard. Use of a rule-based system for process control. IEEE Control Systems Magazine, 8(5):3–13, 1988.
- Robust incremental neural semantic graph parsing. In NAACL, pages 1215–1226, 2017.
- Connectionism and information processing abstractions. AI Magazine, 9(4):24–24, 1988.
- Learning to compose task-specific tree structures. In AAAI, pages 5094–5101, 2018.
- What does BERT look at? An analysis of BERT attention. In BlackboxNLP, pages 276–286, 2019.
- Unsupervised chunking as syntactic structure induction with a knowledge-transfer approach. In Findings of EMNLP, pages 3626–3634, 2021.
- Li Dong and Mirella Lapata. Language to logical form with neural attention. In ACL, pages 33–43, 2016.
- Unsupervised latent tree induction with deep inside-outside recursive autoencoders. In NAACL-HLT, pages 1129–1141, 2019.
- Template filling with generative Transformers. In NAACL, pages 909–914, 2021.
- Herbert B Enderton. A Mathematical Introduction to Logic. Elsevier, 2001.
- Paraphrase generation with latent bag of words. In NeurIPS, pages 13645–13656, 2019.
- Ashok Goel. Looking back, looking ahead: Symbolic versus connectionist AI. AI Magazine, 42(4):83–85, 2022.
- Neural Turing machines. arXiv preprint arXiv:1410.5401, 2014.
- A deep generative framework for paraphrase generation. In AAAI, pages 5149–5156, 2018.
- DeBERTa: Decoding-enhanced BERT with disentangled attention. In ICLR, 2021.
- Factorising meaning and form for intent-preserving paraphrasing. In ACL-IJCNLP, pages 1405–1418, 2021.
- Grammar induction with neural language models: An unusual replication. In EMNLP, pages 4998–5003, 2018.
- Attention is not explanation. In NAACL-HLT, pages 3543–3556, 2019.
- Categorical reparameterization with Gumbel-softmax. In ICLR, 2017.
- Dynamic entity representations in neural language models. In EMNLP, pages 1830–1839, 2017.
- Yoon Kim. Convolutional neural networks for sentence classification. In EMNLP, pages 1746–1751, 2014.
- Corpus-based induction of syntactic structure: Models of dependency and constituency. In ACL, pages 479–486, 2014.
- Attention is not only a weight: Analyzing Transformers with vector norms. In EMNLP, pages 7057–7075, 2020.
- Building machines that learn and think like people. Behavioral and Brain Sciences, 40:E253, 2017.
- Deep learning. Nature, 521(7553):436–444, 2015.
- An imitation learning approach to unsupervised parsing. In ACL, pages 3485–3492, 2019.
- Neural symbolic machines: Learning semantic parsers on Freebase with weak supervision. In ACL, pages 23–33, 2017.
- Abstract rule learning for paraphrase generation. In IJCAI, pages 4273–4279, 2022.
- Jumper: Learning when to make classification decision in reading. In IJCAI, pages 4237–4243, 2018.
- Finding decision jumps in text classification. Neurocomputing, 371:177–187, 2020.
- Unsupervised paraphrasing by simulated annealing. In ACL, pages 302–312, 2019.
- Object-oriented neural programming (OONP) for document understanding. In ACL, pages 2717–2726, 2018.
- Jointly learning sentence embeddings and syntax with unsupervised Tree-LSTMs. arXiv preprint arXiv:1705.09189, 2017.
- Gary Marcus. Deep learning: A critical appraisal. arXiv preprint arXiv:1801.00631, 2018.
- Rule learning by seven-month-old infants. Science, 283(5398):77–80, 1999.
- John McCarthy. Artificial intelligence, logic and formalizing common sense. In Philosophical Logic and Artificial Intelligence, pages 161–190, 1989.
- Distributed representations of words and phrases and their compositionality. In NeurIPS, 2013.
- Coupling distributed and symbolic execution for natural language queries. In ICML, pages 2518–2526, 2017.
- Hybrid simplification using deep semantics and machine translation. In ACL, pages 435–445, 2014.
- Neural programmer: Inducing latent programs with gradient descent. In ICLR, 2016.
- Thumbs up? Sentiment classification using machine learning techniques. In EMNLP, pages 79–86, 2002.
- Mathematical Methods in Linguistics. Springer Science & Business Media, 2012.
- Neural paraphrase generation with stacked residual LSTM networks. In COLING, pages 2923––2934, 2016.
- Monolingual machine translation for paraphrase generation. In EMNLP, pages 142–149, 2004.
- Bidirectional attention flow for machine comprehension. In ICLR, 2017.
- Neural language modeling by jointly learning syntax and lexicon. In ICLR, 2018.
- An expert system for oncology protocol management. Rule-Based Expert Systems, pages 653–65, 1984.
- Axiomatic attribution for deep networks. In ICML, pages 3319–3328, 2017.
- Reinforcement Learning: An Introduction. MIT Press, 2018.
- Attention is all you need. In NeurIPS, pages 5998–6008, 2017.
- Multilingual relation extraction using compositional universal schema. In NAACL-HLT, pages 886–896, 2016.
- Baselines and bigrams: Simple, good sentiment and topic classification. In ACL, pages 90–94, 2012.
- A task in a suit and a tie: Paraphrase generation with semantic augmentation. In AAAI, pages 7176–7183, 2019.
- Do latent tree learning models identify meaningful structure in sentences? TACL, 6:253–267, 2018.
- Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8:229–256, 1992.
- Weakly supervised explainable phrasal reasoning with neural fuzzy logic. In ICLR, 2023.
- Differentiable learning of logical rules for knowledge base reasoning. In NeurIPS, 2017.
- Neural enquirer: Learning to query tables with natural language. In IJCAI, pages 2308–2314, 2016.
- Learning to compose words into sentences with reinforcement learning. In ICLR, 2017.
- Lotfi A Zadeh. Fuzzy logic. Computer, 21(4):83–93, 1988.
- Kernel methods for relation extraction. Journal of Machine Learning Research, 3:1083–1106, 2003.
- Relation classification via convolutional deep neural network. In COLING, pages 2335–2344, 2014.
- Learning robust rule representations for abstract reasoning via internal inferences. In NeurIPS, 2022.
- Character-level convolutional networks for text classification. In NeurIPS, pages 649–657, 2015.
- Application-driven statistical paraphrase generation. In ACL, pages 834–842, 2009.
- Attention-based LSTM network for cross-lingual sentiment classification. In EMNLP, pages 247–256, 2016.