Named Entity Recognition for Nepali Using BERT Based Models. Pande, B. D., Shakya, A., Panday, S. P., & Joshi, B. In Fujita, H., Wang, Y., Xiao, Y., & Moonis, A., editors, Advances and Trends in Artificial Intelligence. Theory and Applications, volume 13926, of Lecture Notes in Computer Science (LNAI 13926), pages 93–104, Cham, July, 2023. Springer Nature Switzerland.
Named Entity Recognition for Nepali Using BERT Based Models [link]Paper  doi  abstract   bibtex   2 downloads  
Named Entity Recognition (NER) is one of the vital task for many Natural Language Processing (NLP) tasks. In recent times, transformer architecture-based models have become very popular for NLP tasks including NER achieving state-of-the-art results. The Bidirectional Encoder Representations from Transformers (BERT) model especially has been found to be very good for NER tasks. However, in Nepali limited work has been done using these models with existing works mostly using more traditional techniques. In this work, we show that by using a combination of preprocessing techniques and better-initialized BERT models, we can improve the performance of the NER system in Nepali. We show a significant improvement in results using the multilingual RoBERTa model. Using this, we were able to achieve a 6% overall improvement in the f1 score in EverestNER Dataset. In terms of the fields, we have achieved an increase of up to 22% in the f1 score for the Event entity which has the lowest support.
@inproceedings{10.1007/978-3-031-36822-6_8,
  abstract = {Named Entity Recognition (NER) is one of the vital task for many Natural Language Processing (NLP) tasks. In recent times, transformer architecture-based models have become very popular for NLP tasks including NER achieving state-of-the-art results. The Bidirectional Encoder Representations from Transformers (BERT) model especially has been found to be very good for NER tasks. However, in Nepali limited work has been done using these models with existing works mostly using more traditional techniques. In this work, we show that by using a combination of preprocessing techniques and better-initialized BERT models, we can improve the performance of the NER system in Nepali. We show a significant improvement in results using the multilingual RoBERTa model. Using this, we were able to achieve a 6{\%} overall improvement in the f1 score in EverestNER Dataset. In terms of the fields, we have achieved an increase of up to 22{\%} in the f1 score for the Event entity which has the lowest support.},
  added-at = {2023-07-22T17:27:48.000+0200},
  address = {Cham},
  author = {Pande, Bishal Debb and Shakya, Aman and Panday, Sanjeeb Prasad and Joshi, Basanta},
  biburl = {https://www.bibsonomy.org/bibtex/2bfe49328f9fdfb58d1ee91600b4e7c37/amanshakya},
  booktitle = {Advances and Trends in Artificial Intelligence. Theory and Applications},
  doi = {https://doi.org/10.1007},
  editor = {Fujita, Hamido and Wang, Yinglin and Xiao, Yanghua and Moonis, Ali},
  eventtitle = {36th International Conference on Industrial, Engineering and Other Applications of Applied Intelligent Systems (IEA/AIE 2023)},
  interhash = {885b51776b957fa05117fba3c20e1652},
  intrahash = {bfe49328f9fdfb58d1ee91600b4e7c37},
  isbn = {978-3-031-36822-6},
  keywords = {myown},
  month = {July},
  pages = {93--104},
  publisher = {Springer Nature Switzerland},
  series = {Lecture Notes in Computer Science (LNAI 13926)},
  timestamp = {2023-07-22T17:43:00.000+0200},
  title = {Named Entity Recognition for Nepali Using BERT Based Models},
  url = {https://link.springer.com/chapter/10.1007/978-3-031-36822-6_8},
  venue = {Shanghai, China},
  volume = 13926,
  year = 2023
}

Downloads: 2