DOC PREVIEW
UW-Madison ECE 539 - Nonlinear Conjugate Gradient Method for Supervised Training of MLP

This preview shows page 1-2 out of 5 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 5 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 5 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 5 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

PowerPoint PresentationIntroductionThe AlgorithmImplementation and ResultsResults (II)Nonlinear Conjugate Gradient Method for Supervised Training of MLPAlexandra RateringECE/CS/ME 539December 14, 2001IntroductionIntroductionBack-Propagation AlgorithmCan oscillate and be caught in local minimaSlow convergence rate (zigzag path to the minimum)Many parameters have to be adjusted be the user –learning rate, momentum constant …Nonlinear Conjugate Gradient MethodSecond order optimization approachFaster convergenceFewer parameters to adjustThe AlgorithmThe AlgorithmDirection vector = conjugate gradient vector–Linear combination of past direction vectors and the current negative gradient vector –Reduces oscillatory behavior in the minimum search –Reinforces weight adjustment in accordance with previous successful path directionsLearning rate–Optimal rate determined for every iteration via line search–Robustness of line search is critical for performance of CG-Algorithm )()(min)( nnEnavpwImplementation and ResultsImplementation and ResultsIn Matlab code with interface similar to bpResults for approximation problem of homework #40 50 100 150051015202530training error (epoch size = 64)epocherror0 20 40 60 80 100 120 140 1600102030405060708090training error EavepocherrorBP CGTraining error 0.0021 6.0807 e-4Testing error 4.9477 e-42.4293 e-4Results (II)Results (II)Results for pattern classification problem –Two equally sized 2D Gaussian distributions (30 samples)–Final training result for both CG and BP: Crate = 88.3% after 500 iterations0 100 200 300 400 500 600678910111213141516training error (epoch size = 60)epocherror0 100 200 300 400 500 600678910111213141516training error


View Full Document

UW-Madison ECE 539 - Nonlinear Conjugate Gradient Method for Supervised Training of MLP

Documents in this Course
Load more
Download Nonlinear Conjugate Gradient Method for Supervised Training of MLP
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Nonlinear Conjugate Gradient Method for Supervised Training of MLP and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Nonlinear Conjugate Gradient Method for Supervised Training of MLP 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?