Memory-efficient Stochastic methods for Memory-based Transformers

14 Nov 2023  ·  Vishwajit Kumar Vishnu, C. Chandra Sekhar ·

Training Memory-based transformers can require a large amount of memory and can be quite inefficient. We propose a novel two-phase training mechanism and a novel regularization technique to improve the training efficiency of memory-based transformers, which are often used for long-range context problems. For our experiments, we consider transformer-XL as our baseline model which is one of memorybased transformer models. We show that our resultant model, Skip Cross-head TransformerXL, outperforms the baseline on character level language modeling task with similar parameters and outperforms the baseline on word level language modelling task with almost 20% fewer parameters. Our proposed methods do not require any additional memory. We also demonstrate the effectiveness of our regularization mechanism on BERT which shows similar performance with reduction in standard deviation of scores of around 30% on multiple GLUE tasks.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Language Modelling enwik8 Skip Cross-Head Transformer-XL Bit per Character (BPC) 1.033 # 24
Number of params 41M # 27
Paraphrase Identification Quora Question Pairs Dev BERT + SCH attn Val F1 Score 88.436 # 1
Paraphrase Identification Quora Question Pairs Dev BERT + SCH attm Val Accuracy 91.422 # 1
Language Modelling WikiText-103 Skip Cross-Head Transformer-XL Validation perplexity 21.87 # 22
Test perplexity 22.91 # 49
Number of params 122M # 40

Methods