Reusing Weights in Subword-Aware Neural Language Models
| dc.contributor.author | Takhanov, Rustem | |
| dc.contributor.author | Assylbekov, Zhenisbek | |
| dc.contributor.author | Assylbekov, Zhenisbek | |
| dc.date.accessioned | 2025-08-19T09:22:40Z | |
| dc.date.available | 2025-08-19T09:22:40Z | |
| dc.date.issued | 2018-01-01 | |
| dc.description.abstract | The authors introduce methods for reusing subword embeddings and other parameters in subword-aware neural language models. Techniques improve syllable- and morpheme-aware models' performance while greatly reducing model size. A practical principle is identified: when reusing embedding layers at the output, they should be tied consecutively from bottom up. The best morpheme-aware model significantly outperforms word-level baselines across languages with 20–87 % fewer parameters. | |
| dc.identifier.citation | Assylbekov Z, Takhanov R (2018). Reusing Weights in Subword‑Aware Neural Language Models. In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp 1413–1423. Association for Computational Linguistics. doi=10.18653/v1/N18‑1128 | |
| dc.identifier.doi | 10.18653/v1/n18-1128 | |
| dc.identifier.other | Filename:10.18653_v1_n18-1128.pdf | |
| dc.identifier.uri | https://doi.org/10.18653/v1/n18-1128 | |
| dc.identifier.uri | https://nur.nu.edu.kz/handle/123456789/9514 | |
| dc.language.iso | en | |
| dc.publisher | Association for Computational Linguistics | |
| dc.relation.ispartof | Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers) | en |
| dc.source | Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), 1413-1423, (2018) | en |
| dc.subject | subword embeddings, weight tying, subword-aware models, morpheme-aware language models, model compression | |
| dc.title | Reusing Weights in Subword-Aware Neural Language Models | en |
| dc.type | Journal Article | en |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- 10.18653_v1_n18-1128.pdf
- Size:
- 351.8 KB
- Format:
- Adobe Portable Document Format