VarMAE: Pre-training of Variational Masked Autoencoder for Domain-adaptive Language Understanding

1 Nov 2022  ·  Dou Hu, Xiaolong Hou, Xiyang Du, Mengyuan Zhou, Lianxin Jiang, Yang Mo, Xiaofeng Shi ·

Pre-trained language models have achieved promising performance on general benchmarks, but underperform when migrated to a specific domain. Recent works perform pre-training from scratch or continual pre-training on domain corpora. However, in many specific domains, the limited corpus can hardly support obtaining precise representations. To address this issue, we propose a novel Transformer-based language model named VarMAE for domain-adaptive language understanding. Under the masked autoencoding objective, we design a context uncertainty learning module to encode the token's context into a smooth latent distribution. The module can produce diverse and well-formed contextual representations. Experiments on science- and finance-domain NLU tasks demonstrate that VarMAE can be efficiently adapted to new domains with limited resources.

PDF Abstract
No code implementations yet. Submit your code now

Datasets


Introduced in the Paper:

OIR IEE MTC PSM

Used in the Paper:

SciCite EBM-NLP JNLPBA ACL ARC
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Citation Intent Classification ACL-ARC VarMAE F1 76.98 # 1
Participant Intervention Comparison Outcome Extraction EBM-NLP VarMAE F1 76.01 # 1

Methods


No methods listed for this paper. Add relevant methods here