Reusing Weights in Subword-Aware Neural Language Models
Loading...
Date
Journal Title
Journal ISSN
Volume Title
Publisher
Association for Computational Linguistics
Abstract
The authors introduce methods for reusing subword embeddings and other parameters in subword-aware neural language models. Techniques improve syllable- and morpheme-aware models' performance while greatly reducing model size. A practical principle is identified: when reusing embedding layers at the output, they should be tied consecutively from bottom up. The best morpheme-aware model significantly outperforms word-level baselines across languages with 20–87 % fewer parameters.
Description
Citation
Assylbekov Z, Takhanov R (2018). Reusing Weights in Subword‑Aware Neural Language Models. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp 1413–1423. Association for Computational Linguistics. doi=10.18653/v1/N18‑1128