Find in Library
Search millions of books, articles, and more
Indexed Open Access Databases
MatSciBERT: A materials domain language model for text mining and information extraction
oleh: Tanishq Gupta, Mohd Zaki, N. M. Anoop Krishnan, Mausam
Format: | Article |
---|---|
Diterbitkan: | Nature Portfolio 2022-05-01 |
Deskripsi
Abstract A large amount of materials science knowledge is generated and stored as text published in peer-reviewed scientific literature. While recent developments in natural language processing, such as Bidirectional Encoder Representations from Transformers (BERT) models, provide promising information extraction tools, these models may yield suboptimal results when applied on materials domain since they are not trained in materials science specific notations and jargons. Here, we present a materials-aware language model, namely, MatSciBERT, trained on a large corpus of peer-reviewed materials science publications. We show that MatSciBERT outperforms SciBERT, a language model trained on science corpus, and establish state-of-the-art results on three downstream tasks, named entity recognition, relation classification, and abstract classification. We make the pre-trained weights of MatSciBERT publicly accessible for accelerated materials discovery and information extraction from materials science texts.