ICS-171:Notes 1: 1 Welcome to CompSci 171 Fall 2010 Introduction to AI. Instructor: Max Welling, [email protected] Office hours: Wed. 4-5pm in BH 4028 Teaching Assistant: Levi Boyles Book: Artificial Intelligence, A Modern Approach Russell & Norvig Prentice Hall http://www.ics.uci.edu/~welling/teaching/ICS171spring07/ICS171fall09.html Note: 3rd edition! (I allow 2nd edition as well)ICS-171:Notes 1: 2 • Grading: -Homework (10%, mandatory) -Quizzes (about 8 quizzes) (30%) You will need green large scantron files for this! -One project (30%) -Final Exam (30%) Graded Quizzes/Exams -Answers will be available on the class website Grading Disputes: Turn in your work for re-grading at the discussion section to the TA within 1 week. Note: we will re-grade the entire exam: so your new grade could be higher or lower. Course related issues can be addressed in the first 10 minutes of every class.ICS-171:Notes 1: 3 Academic (Dis)Honesty • It is each student’s responsibility to be familiar with UCI’s current policies on academic honesty • Violations can result in getting an F in the class (or worse) • Please take the time to read the UCI academic honesty policy – in the Fall Quarter schedule of classes – or at: http://www.reg.uci.edu/REGISTRAR/SOC/adh.html • Academic dishonesty is defined as: – Cheating – Dishonest conduct – Plagiarism – Collusion Note: we have been instructed to be tougher on cheating. Everything will be reported.ICS-171:Notes 1: 4 Syllabus: Lecture 1. Introduction: Goals, history (Ch.1) Lecture 2. Philosophical Foundations (Ch.26). Lecture 2. Agents (Ch.2) Lecture 3-4. Uninformed Search (Ch.3) Lecture 5-6 Informed Search (Ch.4) Lecture 7-8. Constraint satisfaction (Ch.5). Project Lecture 9-10 Games (Ch.6) Lecture 11-12. Propositional Logic (Ch.7) Lecture 13-14. First Order Logic (Ch.8) Lecture 15-16-17. Inference in logic (Ch.9) Lecture 18 Uncertainty (Ch.13) Lecture 20. AI Present and Future (Ch.27). Final This is a very rough syllabus. It is almost certainly the case that we will deviate from this. Some chapters will be treated only partially.ICS-171:Notes 1: 5 1. No class Oct 12 2. No discussion in first week 3. Quizzes on Thursdays, first 20 mins in class 4. First quiz Oct. 7 5. Homework due next Monday midnight. 6. We will check if you answered all questions. You must do your HW yourself. You can work in a group, but not copy from a friend. Homework questions will come back in quizzes. 7. Remind me to break for 5-10 mins at 4.10. Important NotesICS-171:Notes 1: 6 Project Build a program that will generate hard random mazes. Build a program that can solve mazes. Compete?ICS-171:Notes 1: 7 Philosophical Foundations • Weak AI: machines can act as if they were intelligent • Strong AI: machines have minds. • Questions: what is a mind? • Will the answer be important for AI? • Objection 1: humans are not subject to Godel’s theorem • Objection 2: humans behavior cannot be modeled by rules • Objection 3: machines cannot be conscious (what is consciousness ?) • Can a “brain in a vat” have the same brain states as in a body? • Brain prosthesis experiment, are we a machine afterwards? • Chinese room: Does the Chinese room have a mind? • Do we need to give up the “illusion” that man is more than a machine? HW: read chapter 26 on philosophical foundations and read piece on intelligence. Form your own opinion and discuss this in class.ICS-171:Notes 1: 8 Meet HAL • 2001: A Space Odyssey – classic science fiction movie from 1969 http://www.youtube.com/watch?v=ukeHdiszZmE&feature=related • HAL – part of the story centers around an intelligent computer called HAL – HAL is the “brains” of an intelligent spaceship – in the movie, HAL can • speak easily with the crew • see and understand the emotions of the crew • navigate the ship automatically • diagnose on-board problems • make life-and-death decisions • display emotions • In 1969 this was science fiction: is it still science fiction? • http://www.youtube.com/watch?v=dKZczUDGp_IICS-171:Notes 1: 9 Ethics • People might lose jobs • People might have too much leasure time • People might lose sense of uniqueness • People might lose privacy rights • People might not be held accountable for certain actions • Machines may replace the human race...ICS-171:Notes 1: 10 Different Types of Artificial Intelligence • Modeling exactly how humans actually think – cognitive models of human reasoning • Modeling exactly how humans actually act – models of human behavior (what they do, not how they think) • Modeling how ideal agents “should think” – models of “rational” thought (formal logic) – note: humans are often not rational! • Modeling how ideal agents “should act” – rational actions but not necessarily formal rational reasoning – i.e., more of a black-box/engineering approach • Modern AI focuses on the last definition – we will also focus on this “engineering” approach – success is judged by how well the agent performs -- modern methods are also inspired by cognitive & neuroscience (how people think).ICS-171:Notes 1: 11 Acting humanly: Turing Test • Turing (1950) "Computing machinery and intelligence": • "Can machines think?" "Can machines behave intelligently?" • Operational test for intelligent behavior: the Imitation Game • Suggested major components of AI: - knowledge representation - reasoning, - language/image understanding, - learning Can you think of a theoretical system that could beat the Turing test yet you wouldn’t find it very intelligent?ICS-171:Notes 1: 12 Acting rationally: rational agent • Rational behavior: Doing that was is expected to maximize one’s “utility function” in this world. • An agent is an entity that perceives and acts. • A rational agent acts rationally. • This course is about designing rational agents • Abstractly, an agent is a function from percept histories to actions: [f: P* A] • For any given class of environments and tasks, we seek the agent (or class of agents) with the best performance •
View Full Document