Initial Explorations on Regularizing the SCRN Model

dc.contributor.authorKabdolov, Olzhas
dc.date.accessioned2018-05-28T08:56:37Z
dc.date.available2018-05-28T08:56:37Z
dc.date.issued2018-05
dc.description.abstractRecurrent neural networks are very powerful sequence models which are used for language modeling as well. Under correct regularization such as naive dropout these models are able to achieve substantial improvement in their performance. We regularize the Structurally Constrained Recurrent Network (SCRN) model and show that despite its simplicity it can achieve the performance comparable to the ubiquitous LSTM model in language modeling task while being smaller in size and up to 2x faster to train. Further analysis shows that regularizing both context and hidden states of the SCRN is crucial.en_US
dc.identifier.urihttp://nur.nu.edu.kz/handle/123456789/3200
dc.language.isoenen_US
dc.publisherNazarbayev University School of Science and Technologyen_US
dc.rightsAttribution-NonCommercial-ShareAlike 3.0 United States*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-sa/3.0/us/*
dc.subjectrecurrent neural networksen_US
dc.subjectwords embeddingsen_US
dc.subjectregularizationen_US
dc.subjectnaive dropouten_US
dc.titleInitial Explorations on Regularizing the SCRN Modelen_US
dc.typeCapstone Projecten_US
workflow.import.sourcescience

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Olzhas.pdf
Size:
223.34 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
6 KB
Format:
Item-specific license agreed upon to submission
Description: