ViHealthBERT: Pre-trained Language Models for Vietnamese in Health Text Mining

Pre-trained language models have become crucial to achieving competitive results across many Natural Language Processing (NLP) problems. For monolingual pre-trained models in low-resource languages, the quantity has been significantly increased. However, most of them relate to the general domain, and there are limited strong baseline language models for domain-specific. We introduce ViHealthBERT, the first domain-specific pre-trained language model for Vietnamese healthcare. The performance of our model shows strong results while outperforming the general domain language models in all health-related datasets. Moreover, we also present Vietnamese datasets for the healthcare domain for two tasks are Acronym Disambiguation (AD) and Frequently Asked Questions (FAQ) Summarization. We release our ViHealthBERT to facilitate future research and downstream application for Vietnamese NLP in domain-specific.

PDF
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Named Entity Recognition In Vietnamese PhoNER COVID19 ViHealthBERT F1 (%) 96.7 # 3

Methods


No methods listed for this paper. Add relevant methods here