Using Artificial Neural Networks and Support Vector Regression to Model the Lyapunov Exponent
-
Upload
cain-gibson -
Category
Documents
-
view
19 -
download
2
description
Transcript of Using Artificial Neural Networks and Support Vector Regression to Model the Lyapunov Exponent
Using Artificial Neural Networks and Support Vector Regression to Model the Lyapunov Exponent
Adam Maus
Nonlinear Prediction
Largest Lyapunov Exponent (LE)– Measure of long term predictability
Models used to reconstruct LE– Artificial Neural Network– Support Vector Regression
Henon Map (HM)– Henon Map’s LE = .42093
kx
Strange Attractor of Hénon Map221 3.04.11 kkk xxx
kx
1kx
Artificial Neural Network
Single Layer Feed-Forward Architecture– 8 neurons and 3 dimensions– Trained using next step prediction– Weights updated using a variant
of simulated annealing– 400 training points– 1 million training epochs
d
jjniji
n
iik xaabx
10
1
tanhˆ
Support Vector Regression
Global Solution to a Convex Programming Problem
Uses only a Subset of Points– Points outside of ɛ-tube thrown out
User-Defined Parameters– C - control flatness of the output function– ɛ - controls size of the tube– kernel function and its parameters– Training dataset size
Many toolboxes available– LibSVM Toolbox
Toolbox Available At: http://tinyurl.com/sxnsePicture: http://tinyurl.com/6pscdr
s = length of w vectorK = kernel function
Kwbxs
iip
1
ˆ
Dynamic and Attractor Reconstruction
Results– Artificial Neural Network
• LE = .38431• Mean Square Error = 4.65 x 10
– Support Vector Regression• LE = .41288 • Weighted Mean Square Error = 4.69 x 10
Strange Attractor of Neural Network
Strange Attractor of Support Vector Regression
-5
c
Xxc
kkk
1
2ˆ
Mean Square Error
c = length of time series
-5
c
xxc
iiii
1
2ˆWeighted Mean Square Error
Conclusions
Hypothesis– Dynamic and Attractor Reconstruction
are correlated• Neural Networks
– Perform proper reconstruction given adequate training
• Support Vector Regression– Performs proper reconstruction given that we
weight the model so that it replicates the first points more due to the chaotic nature of the data.
Actual NN SVR
LM .69526 .63744 .68923
DHM .37038 .35030 .35550
HM .42093 .38431 .41288
Strange Attractor of the Logistic Map (LM)
Strange Attractor of the delayed Hénon map (DHM)