DOC PREVIEW
Berkeley COMPSCI 252 - Branch Prediction using Advanced Neural Methods

This preview shows page 1-2-3-4-5 out of 14 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 14 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

Slide 1Slide 2Slide 3Slide 4Slide 5Slide 6Slide 7Slide 8Slide 9Slide 10Slide 11Slide 12Slide 13Slide 14Branch Prediction using Branch Prediction using Advanced Neural Advanced Neural MethodsMethodsSunghoon KimSunghoon KimCS252 ProjectCS252 ProjectIntroductionIntroductionDynanmic Branch PredictionDynanmic Branch PredictionNo doubt about its importance in No doubt about its importance in Speculation performanceSpeculation performanceGiven history of branch behaviors, predict Given history of branch behaviors, predict branch behaviors at the next stepbranch behaviors at the next stepCommon solutions: gshare, bimode,hybrid…Common solutions: gshare, bimode,hybrid… Replace saturating counters with neural Replace saturating counters with neural methods?methods?Neural methodsNeural methodsCapable of classification (predicting Capable of classification (predicting into which set of classes a particular into which set of classes a particular instance will fall)instance will fall)Learns correlations between inputs Learns correlations between inputs and output, and generalized learning and output, and generalized learning to other inputsto other inputsPotential to solve problems of most Potential to solve problems of most two-levels predictors two-levels predictorsSimulation Models - GshareSimulation Models - Gshare20-bit Global history shift register20-bit Global history shift registerPer-address history table with 2-bit Per-address history table with 2-bit saturation countersaturation counterPHTBrachPC>>2GHT Predict Updatecounters2-bit saturation counterSimulation Models - Simulation Models - PerceptronPerceptron14-bit Global history shift register14-bit Global history shift registerPer-address history table with 8-bit weights Per-address history table with 8-bit weights and biasand biasIndexed by Gshare or BranchPC aloneIndexed by Gshare or BranchPC alonePHTBrachPC>>2GHT PredictTrainingweights & biasBrachPC>>2ORSimulation Models - Simulation Models - BackpropagationBackpropagation10-bit GHR10-bit GHRSigmoid transfer functionSigmoid transfer functionFloating point computationFloating point computationFloating point weights and biasesFloating point weights and biases20 neurons one hidden layer20 neurons one hidden layerPHTBrachPC>>2GHT PredictTrainingweights & biasBrachPC>>2ORSimulation Models Simulation Models – Radial Basis Networks– Radial Basis NetworksTransfer function for radial basis Transfer function for radial basis neuron exp(-nneuron exp(-n22))Distance function between an input Distance function between an input vector and a weight vectorvector and a weight vectorPHTBrachPC>>2GHT PredictTrainingweights & biasBrachPC>>2ORSimulation Models Simulation Models – Elman Networks– Elman NetworksFeedback from the hidden layer Feedback from the hidden layer outputs to the first layeroutputs to the first layerPHTBrachPC>>2GHT PredictTrainingweights & biasBrachPC>>2ORSimulation Models – Learning Simulation Models – Learning Vector Quantization NetworksVector Quantization NetworksDistance function as Radial but without biasesDistance function as Radial but without biasesCompetitive function gives one only to an winning Competitive function gives one only to an winning input (biggest value) and zeroe to the otherinput (biggest value) and zeroe to the otherPHTBrachPC>>2GHT PredictTrainingweights & biasBrachPC>>2ORSimulation EnvironmentSimulation EnvironmentSimpleScalar ToolSimpleScalar ToolSome of SPEC2000 benchmarksSome of SPEC2000 benchmarksExecute 100,000,000 instructions and Execute 100,000,000 instructions and dump conditional branch historiesdump conditional branch histories5000 branch instructions are used for 5000 branch instructions are used for training training Make all budgets for PHTs the sameMake all budgets for PHTs the sameFloating point is 4 byteFloating point is 4 byteResultsResultsHardware constraintsHardware constraintsPredictors must predict within a (few) Predictors must predict within a (few) cyclecycleGshare : easy to achieve Gshare : easy to achieve Perceptron : Integer adders, possible Perceptron : Integer adders, possible alternative, more accurate if more layersalternative, more accurate if more layersOther advanced neural net : Hard to Other advanced neural net : Hard to implement, Floating point functional implement, Floating point functional units,units,Future WorksFuture WorksReplace floating point weights and biases Replace floating point weights and biases with scaled integer ones?with scaled integer ones?Replace floating point function with Replace floating point function with approximately equivalent integer function, approximately equivalent integer function, using Taylor’s series?using Taylor’s series?Without budget consideration, what will be Without budget consideration, what will be the best performance of advanced neural the best performance of advanced neural network methods?network methods?Look at codes carefully if there are mistakesLook at codes carefully if there are mistakesConclusionsConclusionsThere is not much of benefit using There is not much of benefit using advanced neural networks on the same advanced neural networks on the same budget as Gshare, sometimes worsebudget as Gshare, sometimes worseElman Networks method is the bestElman Networks method is the bestHard to implement in hardware unless Hard to implement in hardware unless floating point computation are easy to floating point computation are easy to dodoNN can be alternative predictors if well NN can be alternative predictors if well


View Full Document

Berkeley COMPSCI 252 - Branch Prediction using Advanced Neural Methods

Documents in this Course
Quiz

Quiz

9 pages

Caches I

Caches I

46 pages

Lecture 6

Lecture 6

36 pages

Lecture 9

Lecture 9

52 pages

Figures

Figures

26 pages

Midterm

Midterm

15 pages

Midterm

Midterm

14 pages

Midterm I

Midterm I

15 pages

ECHO

ECHO

25 pages

Quiz  1

Quiz 1

12 pages

Load more
Download Branch Prediction using Advanced Neural Methods
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Branch Prediction using Advanced Neural Methods and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Branch Prediction using Advanced Neural Methods 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?