Initial Explorations on Regularizing the SCRN Model
dc.contributor.author | Kabdolov, Olzhas | |
dc.date.accessioned | 2018-05-28T08:56:37Z | |
dc.date.available | 2018-05-28T08:56:37Z | |
dc.date.issued | 2018-05 | |
dc.description.abstract | Recurrent neural networks are very powerful sequence models which are used for language modeling as well. Under correct regularization such as naive dropout these models are able to achieve substantial improvement in their performance. We regularize the Structurally Constrained Recurrent Network (SCRN) model and show that despite its simplicity it can achieve the performance comparable to the ubiquitous LSTM model in language modeling task while being smaller in size and up to 2x faster to train. Further analysis shows that regularizing both context and hidden states of the SCRN is crucial. | en_US |
dc.identifier.uri | http://nur.nu.edu.kz/handle/123456789/3200 | |
dc.language.iso | en | en_US |
dc.publisher | Nazarbayev University School of Science and Technology | en_US |
dc.rights | Attribution-NonCommercial-ShareAlike 3.0 United States | * |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/3.0/us/ | * |
dc.subject | recurrent neural networks | en_US |
dc.subject | words embeddings | en_US |
dc.subject | regularization | en_US |
dc.subject | naive dropout | en_US |
dc.title | Initial Explorations on Regularizing the SCRN Model | en_US |
dc.type | Capstone Project | en_US |
workflow.import.source | science |