Show simple item record

contributor authorMonahan, Adam H.
date accessioned2017-06-09T15:48:37Z
date available2017-06-09T15:48:37Z
date copyright2000/02/01
date issued2000
identifier issn0894-8755
identifier otherams-5405.pdf
identifier urihttp://onlinelibrary.yabesh.ir/handle/yetl/4194012
description abstractA nonlinear generalization of principal component analysis (PCA), denoted nonlinear principal component analysis (NLPCA), is implemented in a variational framework using a five-layer autoassociative feed-forward neural network. The method is tested on a dataset sampled from the Lorenz attractor, and it is shown that the NLPCA approximations to the attractor in one and two dimensions, explaining 76% and 99.5% of the variance, respectively, are superior to the corresponding PCA approximations, which respectively explain 60% (mode 1) and 95% (modes 1 and 2) of the variance. It is found that as noise is added to the Lorenz attractor, the NLPCA approximations remain superior to the PCA approximations until the noise level is so great that the lower-dimensional nonlinear structure of the data is no longer manifest to the eye. Finally, directions for future work are presented, and a cinematographic technique to visualize the results of NLPCA is discussed.
publisherAmerican Meteorological Society
titleNonlinear Principal Component Analysis by Neural Networks: Theory and Application to the Lorenz System
typeJournal Paper
journal volume13
journal issue4
journal titleJournal of Climate
identifier doi10.1175/1520-0442(2000)013<0821:NPCABN>2.0.CO;2
journal fristpage821
journal lastpage835
treeJournal of Climate:;2000:;volume( 013 ):;issue: 004
contenttypeFulltext


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record