DSpace Repository

Initial Explorations on Regularizing the SCRN Model

Show simple item record

dc.contributor.author Kabdolov, Olzhas
dc.date.accessioned 2018-05-28T08:56:37Z
dc.date.available 2018-05-28T08:56:37Z
dc.date.issued 2018-05
dc.identifier.uri http://nur.nu.edu.kz/handle/123456789/3200
dc.description.abstract Recurrent neural networks are very powerful sequence models which are used for language modeling as well. Under correct regularization such as naive dropout these models are able to achieve substantial improvement in their performance. We regularize the Structurally Constrained Recurrent Network (SCRN) model and show that despite its simplicity it can achieve the performance comparable to the ubiquitous LSTM model in language modeling task while being smaller in size and up to 2x faster to train. Further analysis shows that regularizing both context and hidden states of the SCRN is crucial. en_US
dc.language.iso en en_US
dc.publisher Nazarbayev University School of Science and Technology en_US
dc.rights Attribution-NonCommercial-ShareAlike 3.0 United States *
dc.rights.uri http://creativecommons.org/licenses/by-nc-sa/3.0/us/ *
dc.subject recurrent neural networks en_US
dc.subject words embeddings en_US
dc.subject regularization en_US
dc.subject naive dropout en_US
dc.title Initial Explorations on Regularizing the SCRN Model en_US
dc.type Capstone Project en_US
workflow.import.source science


Files in this item

The following license files are associated with this item:

This item appears in the following Collection(s)

Show simple item record

Attribution-NonCommercial-ShareAlike 3.0 United States Except where otherwise noted, this item's license is described as Attribution-NonCommercial-ShareAlike 3.0 United States

Video Guide

Submission guideSubmission guide

Submit your materials for publication to

NU Repository Drive

Browse

My Account

Statistics