Reusing Weights in Subword-Aware Neural Language Models

Loading...
Thumbnail Image

Journal Title

Journal ISSN

Volume Title

Publisher

Association for Computational Linguistics

Abstract

The authors introduce methods for reusing subword embeddings and other parameters in subword-aware neural language models. Techniques improve syllable- and morpheme-aware models' performance while greatly reducing model size. A practical principle is identified: when reusing embedding layers at the output, they should be tied consecutively from bottom up. The best morpheme-aware model significantly outperforms word-level baselines across languages with 20–87 % fewer parameters.

Description

Citation

Assylbekov Z, Takhanov R (2018). Reusing Weights in Subword‑Aware Neural Language Models. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp 1413–1423. Association for Computational Linguistics. doi=10.18653/v1/N18‑1128

Collections

Endorsement

Review

Supplemented By

Referenced By