Unformatted text preview:

Learning from ObservationsChapter 18, Sections 1–3Chapter 18, Sections 1–3 1Outline♦ Learning agents♦ Inductive learning♦ Decision tree learning♦ Measuring learning performanceChapter 18, Sections 1–3 2LearningLearning is essential for unknown environments,i.e., when designer lacks omniscienceLearning is useful as a system construction method,i.e., expose the agent to reality rather than trying to write it downLearning modifies the agent’s decision mechanisms to improve performanceChapter 18, Sections 1–3 3Learning agentsPerformance standardAgentEnvironmentSensorsEffectorsPerformance elementchangesknowledgelearning goals Problem generator feedback Learning elementCriticexperimentsChapter 18, Sections 1–3 4Learning elementDesign of learning element is dictated by♦ what type of performance element is used♦ which functional component is to be learned♦ how that functional compoent is represented♦ what kind of feedback is availableExample scenarios:Performance elementAlpha−beta searchLogical agentSimple reflex agentComponentEval. fn.Transition modelTransition modelRepresentationWeighted linear functionSuccessor−state axiomsNeural netDynamic Bayes netUtility−based agentPercept−action fnFeedbackOutcomeOutcomeWin/lossCorrect actionSupervised learning: correct answers for each instanceReinforcement learning: occasional rewardsChapter 18, Sections 1–3 5Inductive learning (a.k.a. Science)Simplest form: learn a function from examples (tabula rasa)f is the target functionAn example is a pair x, f(x), e.g.,OO XXX, +1Problem: find a(n) hypothesis hsuch that h ≈ fgiven a training set of examples(This is a highly simplified mo del of real learning:– Ignores prior knowledge– Assumes a deterministic, observable “environment”– Assumes examples are given– Assumes that the agent wants to learn f—why?)Chapter 18, Sections 1–3 6Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Chapter 18, Sections 1–3 7Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Chapter 18, Sections 1–3 8Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Chapter 18, Sections 1–3 9Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Chapter 18, Sections 1–3 10Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Chapter 18, Sections 1–3 11Inductive learning methodConstruct/adjust h to agree with f on training set(h is consistent if it agrees with f on all examples)E.g., curve fitting:xf(x)Ockham’s razor: maximize a combination of consistency and simplicityChapter 18, Sections 1–3 12Attribute-based representationsExamples described by attribute values (Boolean, discrete, continuous, etc.)E.g., situations where I will/won’t wait for a table:ExampleAttributes TargetAlt Bar Fri Hun Pat Price Rain Res Type Est WillWaitX1T F F T Some $$$ F T French 0–10 TX2T F F T Full $ F F Thai 30–60 FX3F T F F Some $ F F Burger 0–10 TX4T F T T Full $ F F Thai 10–30 TX5T F T F Full $$$ F T French >60 FX6F T F T Some $$ T T Italian 0–10 TX7F T F F None $ T F Burger 0–10 FX8F F F T Some $$ T T Thai 0–10 TX9F T T F Full $ T F Burger >60 FX10T T T T Full $$$ F T Italian 10–30 FX11F F F F None $ F F Thai 0–10 FX12T T T T Full $ F F Burger 30–60 TClassification of examples is positive (T) or negative (F)Chapter 18, Sections 1–3 13Decision treesOne possible representation for hypothesesE.g., here is the “true” tree for deciding whether to wait:No YesNo YesNo YesNo YesNo YesNo YesNone Some Full>60 30−60 10−30 0−10No YesAlternate?Hungry?Reservation?Bar? Raining?Alternate?Patrons?Fri/Sat?WaitEstimate?F TF TTTF TTFTTFChapter 18, Sections 1–3 14ExpressivenessDecision trees can express any function of the input attributes.E.g., for Boolean functions, truth table row → path to leaf:FTABF TBA B A xor BFF FF TTT F TTT FFFF T T TTrivially, there is a consistent decision tree for any training setw/ one path to leaf for each example (unless f nondeterministic in x)but it probably won’t generalize to new examplesPrefer to find more compact decision treesChapter 18, Sections 1–3 15Hypothesis spacesHow many distinct decision trees with n Boolean attributes??Chapter 18, Sections 1–3 16Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functionsChapter 18, Sections 1–3 17Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functions= number of distinct truth tables with 2nrowsChapter 18, Sections 1–3 18Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functions= number of distinct truth tables with 2nrows = 22nChapter 18, Sections 1–3 19Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functions= number of distinct truth tables with 2nrows = 22nE.g., with 6 Boolean attributes, there are 18,446,744,073,709,551,616 treesChapter 18, Sections 1–3 20Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functions= number of distinct truth tables with 2nrows = 22nE.g., with 6 Boolean attributes, there are 18,446,744,073,709,551,616 treesHow many purely conjunctive hypotheses (e.g., Hungry ∧¬Rain)??Chapter 18, Sections 1–3 21Hypothesis spacesHow many distinct decision trees with n Boolean attributes??= number of Boolean functions= number of distinct truth tables with 2nrows = 22nE.g., with 6 Boolean attributes, there are 18,446,744,073,709,551,616 treesHow many purely conjunctive hypotheses (e.g., Hungry ∧¬Rain)??Each attribute can be in (positive), in (negative), or out⇒ 3ndistinct conjunctive hypothesesMore expressive hypothesis space– increases chance that target function can be expressed– increases number of hypotheses consistent w/ training set⇒ may get worse predictionsChapter 18, Sections 1–3 22Decision tree learningAim: find a small tree consistent with the training examplesIdea:


View Full Document

U of M CS 5541 - CS 5541 Lecture notes

Download CS 5541 Lecture notes
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view CS 5541 Lecture notes and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view CS 5541 Lecture notes 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?