Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MatSciRE: Leveraging Pointer Networks to Automate Entity and Relation Extraction for Material Science Knowledge-base Construction (2401.09839v1)

Published 18 Jan 2024 in cs.CL, cs.CE, and cs.IR

Abstract: Material science literature is a rich source of factual information about various categories of entities (like materials and compositions) and various relations between these entities, such as conductivity, voltage, etc. Automatically extracting this information to generate a material science knowledge base is a challenging task. In this paper, we propose MatSciRE (Material Science Relation Extractor), a Pointer Network-based encoder-decoder framework, to jointly extract entities and relations from material science articles as a triplet ($entity1, relation, entity2$). Specifically, we target the battery materials and identify five relations to work on - conductivity, coulombic efficiency, capacity, voltage, and energy. Our proposed approach achieved a much better F1-score (0.771) than a previous attempt using ChemDataExtractor (0.716). The overall graphical framework of MatSciRE is shown in Fig 1. The material information is extracted from material science literature in the form of entity-relation triplets using MatSciRE.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (63)
  1. Li-ion battery materials: present and future, Materials today Energy (Mater. Today Energy) 18 (2015) 252–264.
  2. S. Huang, J. M. Cole, A database of battery materials auto-generated using chemdataextractor, Scientific Data (Sci. Data) 7 (2020) 1–13.
  3. T. Nayak, H. T. Ng, Effective modeling of encoder-decoder architecture for joint entity and relation extraction, in: Proceedings of the AAAI conference on artificial intelligence, volume 34, 2020, pp. 8528–8535.
  4. The impact of domain-specific pre-training on named entity recognition tasks in materials science, Available at SSRN 3950755 (2021) 1–43.
  5. Unsupervised word embeddings capture latent knowledge from materials science literature, Nature 571 (2019) 95–98.
  6. Efficient estimation of word representations in vector space, arXiv preprint arXiv:1301.3781 (2013).
  7. Dbpedia: A nucleus for a web of open data, The Semantic Web (2007) 722–735.
  8. Knowledge extraction and application for metal materials based on dbpedia, 2014 10th International Conference on Semantics, Knowledge and Grids (2014) 150–153.
  9. Mmkg: An approach to generate metallic materials knowledge graph based on dbpedia and wikipedia, Computer Physics Communications ( Comput. Phys. Commun.) 211 (2017) 98–112.
  10. Named entity recognition and normalization applied to large-scale information extraction from the materials science literature, Journal of chemical information and modeling (J. Chem. Inf. Model.) 59 (2019) 3692–3702.
  11. Matscie: An automated tool for the generation of databases of methods and parameters used in the computational materials science literature, Computational Materials Science (Comput. Mater. Sci.) 192 (2021) 110325.
  12. Using sentence-level classification helps entity extraction from material science literature, in: Proceedings of the Thirteenth Language Resources and Evaluation Conference, 2022, pp. 4540–4545.
  13. Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction, arXiv preprint arXiv:1808.09602 (2018).
  14. Scibert: A pretrained language model for scientific text, arXiv preprint arXiv:1903.10676 (2019).
  15. Matscibert: A materials domain language model for text mining and information extraction, npj Computational Materials 8 (2022) 1–11.
  16. S. Huang, J. M. Cole, Batterybert: A pretrained language model for battery database enhancement, Journal of Chemical Information and Modeling (2022).
  17. Distant supervision for relation extraction without labeled data, in: Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, 2009, pp. 1003–1011.
  18. Freebase: a collaboratively created graph database for structuring human knowledge, in: Proceedings of the 2008 ACM SIGMOD international conference on Management of data, 2008, pp. 1247–1250.
  19. Modeling relations and their mentions without labeled text, in: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Springer, 2010, pp. 148–163.
  20. Knowledge-based weak supervision for information extraction of overlapping relations, in: Proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies, 2011, pp. 541–550.
  21. Distant supervision for relation extraction via piecewise convolutional neural networks, in: Proceedings of the 2015 conference on empirical methods in natural language processing, 2015, pp. 1753–1762.
  22. Y. Shen, X.-J. Huang, Attention-based convolutional neural network for semantic relation extraction, in: Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, 2016, pp. 2526–2536.
  23. Improving distantly supervised relation extraction using word and entity based attention, arXiv preprint arXiv:1804.06987 (2018).
  24. Reside: Improving distantly-supervised neural relation extraction using side information, arXiv preprint arXiv:1812.04361 (2018).
  25. Z.-X. Ye, Z.-H. Ling, Distant supervision relation extraction with intra-bag and inter-bag attentions, arXiv preprint arXiv:1904.00143 (2019).
  26. Attention guided graph convolutional networks for relation extraction, arXiv preprint arXiv:1906.07510 (2019).
  27. A framework to generate high-quality datapoints for multiple novel intent detection, in: Findings of the Association for Computational Linguistics: NAACL 2022, 2022, pp. 282–292.
  28. Intent identification and entity extraction for healthcare queries in indic languages, in: Findings of the Association for Computational Linguistics: EACL 2023, 2023, pp. 1825–1836.
  29. Fine-grained intent classification in the legal domain, arXiv preprint arXiv:2205.03509 (2022).
  30. A. Mullick, Novel intent detection and active learning based classification (student abstract), arXiv e-prints (2023a) arXiv–2304.
  31. A. Mullick, Exploring multilingual intent dynamics and applications, IJCAI Doctoral Consortium (2023b).
  32. A generic opinion-fact classifier with application in understanding opinionatedness in various news section, in: Proceedings of the 26th International Conference on World Wide Web Companion, 2017, pp. 827–828.
  33. A graphical framework to detect and categorize diverse opinions from online news, in: Proceedings of the Workshop on Computational Modeling of People’s Opinions, Personality, and Emotions in Social Media (PEOPLES), 2016, pp. 40–49.
  34. Identifying opinion and fact subcategories from the social web, in: Proceedings of the 2018 ACM International Conference on Supporting Group Work, 2018a, pp. 145–149.
  35. Harnessing twitter for answering opinion list queries, IEEE Transactions on Computational Social Systems 5 (2018b) 1083–1095.
  36. D-fj: Deep neural network based factuality judgment, Technology 50 (2019) 173.
  37. Extracting social lists from twitter, in: Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2017, 2017, pp. 391–394.
  38. Cotype: Joint extraction of typed entities and relations with knowledge bases, in: Proceedings of the 26th International Conference on World Wide Web, 2017, pp. 1015–1024.
  39. M. Miwa, M. Bansal, End-to-end relation extraction using lstms on sequences and tree structures, arXiv preprint arXiv:1601.00770 (2016).
  40. Joint entity recognition and relation extraction as a multi-head selection problem, Expert Systems with Applications 114 (2018) 34–45.
  41. Joint extraction of entities and relations based on a novel tagging scheme, arXiv preprint arXiv:1706.05075 (2017).
  42. D. Q. Nguyen, K. Verspoor, End-to-end neural relation extraction using deep biaffine attention, in: European Conference on Information Retrieval, Springer, 2019, pp. 729–738.
  43. A. Katiyar, C. Cardie, Investigating lstms for joint extraction of opinion entities and relations, in: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2016, pp. 919–929.
  44. Gtr-lstm: A triple encoder for sentence generation from rdf data, in: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2018, pp. 1627–1637.
  45. D. Marcheggiani, L. Perez-Beltrachini, Deep graph convolutional encoders for structured data to text generation, arXiv preprint arXiv:1810.09995 (2018).
  46. Mt/ie: Cross-lingual open information extraction with neural sequence-to-sequence models, in: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, 2017, pp. 64–70.
  47. Neural open information extraction, arXiv preprint arXiv:1805.04270 (2018).
  48. Neural machine translation by jointly learning to align and translate, arXiv preprint arXiv:1409.0473 (2014).
  49. Effective approaches to attention-based neural machine translation, arXiv preprint arXiv:1508.04025 (2015).
  50. Pointer networks, Advances in neural information processing systems 28 (2015).
  51. S. Kundu, H. T. Ng, A question-focused multi-factor attention network for question answering, in: Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018, pp. 5828–5835.
  52. Atomic layer deposited tio 2 on a nitrogen-doped graphene/sulfur electrode for high performance lithium–sulfur batteries, Energy & Environmental Science (Energy Environ. Sci.) 9 (2016) 1495–1503.
  53. Rte: A tool for annotating relation triplets from text, arXiv preprint arXiv:2108.08184 (2021).
  54. J. R. Landis, G. G. Koch, The measurement of observer agreement for categorical data, biometrics (1977) 159–174.
  55. Bert: Pre-training of deep bidirectional transformers for language understanding, in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Association for Computational Linguistics, 2019, pp. 4171–4186.
  56. Aligning books and movies: Towards story-like visual explanations by watching movies and reading books, in: Proceedings of the IEEE international conference on computer vision, 2015, pp. 19–27.
  57. One billion word benchmark for measuring progress in statistical language modeling, arXiv preprint arXiv:1312.3005 (2013).
  58. Roberta: A robustly optimized bert pretraining approach, ArXiv abs/1907.11692 (2019).
  59. A. Gokaslan, V. Cohen, Openwebtext corpus. http://skylion007.github.io/openwebtextcorpus (2019).
  60. T. H. Trinh, Q. V. Le, A simple method for commonsense reasoning, arXiv preprint arXiv:1806.02847 (2018).
  61. Cu0. 02ti0. 94nb2. 04o7: an advanced anode material for lithium-ion batteries of electric vehicles, Journal of Power Sources (J. Power Sources) 328 (2016) 336–344.
  62. Entity, relation, and event extraction with contextualized span representations, arXiv preprint arXiv:1909.03546 (2019).
  63. Matkg: The largest knowledge graph in materials science–entities, relations, and link prediction through graph representation learning, arXiv preprint arXiv:2210.17340 (2022).
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Ankan Mullick (18 papers)
  2. Akash Ghosh (14 papers)
  3. G Sai Chaitanya (2 papers)
  4. Samir Ghui (2 papers)
  5. Tapas Nayak (17 papers)
  6. Seung-Cheol Lee (49 papers)
  7. Satadeep Bhattacharjee (60 papers)
  8. Pawan Goyal (170 papers)
Citations (9)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets