dc.contributor.author | Zollanvari, Amin | |
dc.date.accessioned | 2017-01-06T08:23:05Z | |
dc.date.available | 2017-01-06T08:23:05Z | |
dc.date.issued | 2016-04-12 | |
dc.identifier.citation | Zollanvari, A. (2016). High-dimensional statistical learning: Roots, justifications, and potential machineries. Cancer Informatics, 15, 109-121. DOI: 10.4137/CIN.S30804 | ru_RU |
dc.identifier.uri | http://nur.nu.edu.kz/handle/123456789/2183 | |
dc.description.abstract | High-dimensional data generally refer to data in which the number of variables is larger than the sample size. Analyzing such datasets poses great challenges for classical statistical learning because the finite-sample performance of methods developed within classical statistical learning does not live up to classical asymptotic premises in which the sample size unboundedly grows for a fixed dimensionality of observations. | ru_RU |
dc.language.iso | en | ru_RU |
dc.publisher | Cancer Informatics | ru_RU |
dc.rights | Attribution-NonCommercial-ShareAlike 3.0 United States | * |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-sa/3.0/us/ | * |
dc.subject | Curse of dimensionality | ru_RU |
dc.subject | Double asymptotics | ru_RU |
dc.subject | G-analysis | ru_RU |
dc.subject | High-dimensional analysis | ru_RU |
dc.subject | Kolmogorov asymptotics | ru_RU |
dc.subject | Random matrix theory | ru_RU |
dc.subject | Ridge estimation | ru_RU |
dc.subject | Shrinkage | ru_RU |
dc.subject | Sparsity | ru_RU |
dc.subject | Research Subject Categories::TECHNOLOGY::Chemical engineering | ru_RU |
dc.title | High-dimensional statistical learning: Roots, justifications, and potential machineries | ru_RU |
dc.type | Article | ru_RU |
The following license files are associated with this item: