CS 331: Artificial Intelligence Intelligent Agents

Similar documents
Chapter 2 Intelligent Agents

Chapter 2: Intelligent Agents

Introduction to Artificial Intelligence. Intelligent Agents

Artificial Intelligence

Agents: Rationality (2)

Intelligent Agents. Chapter 2. Chapter 2 1

Measuring the Performance of an Agent

cs171 HW 1 - Solutions

INTELLIGENT AGENTS 2.1 INTRODUCTION 2.2 HOW AGENTS SHOULD ACT

Introduction to Artificial Intelligence

A Review of Intelligent Agents

Course Outline Department of Computing Science Faculty of Science. COMP Applied Artificial Intelligence (3,1,0) Fall 2015

Appendices master s degree programme Artificial Intelligence

Formalizing Properties of Agents

CSC384 Intro to Artificial Intelligence

AI: A Modern Approach, Chpts. 3-4 Russell and Norvig

Proceedings of the 9th WSEAS International Conference on APPLIED COMPUTER SCIENCE

Intelligent Agents. Based on An Introduction to MultiAgent Systems and slides by Michael Wooldridge

6.080/6.089 GITCS Feb 12, Lecture 3

Designing Behavior-Based Systems

Digital Systems Based on Principles and Applications of Electrical Engineering/Rizzoni (McGraw Hill

A Survey on Software Agent Architectures

Simulation-based traffic management for autonomous and connected vehicles

AIBO as an Intelligent Robot Watchdog

Tutorial 1. Introduction to robot

Lecture 1: Introduction to Reinforcement Learning

Robot coined by Karel Capek in a 1921 science-fiction Czech play

Motivation. Motivation. Can a software agent learn to play Backgammon by itself? Machine Learning. Reinforcement Learning

Intelligent Flexible Automation

MECE 102 Mechatronics Engineering Orientation

By: M.Habibullah Pagarkar Kaushal Parekh Jogen Shah Jignasa Desai Prarthna Advani Siddhesh Sarvankar Nikhil Ghate

Toyota Prius Hybrid Display System Usability Test Report

Appendices master s degree programme Human Machine Communication

Overview and History of Operating Systems

Software Active Online Monitoring Under. Anticipatory Semantics

ROBOTICS 01PEEQW. Basilio Bona DAUIN Politecnico di Torino

CE801: Intelligent Systems and Robotics Lecture 3: Actuators and Localisation. Prof. Dr. Hani Hagras

Lab 3 Microcontroller programming Interfacing to Sensors and Actuators with irobot

Formal Verification and Linear-time Model Checking

Unit A451: Computer systems and programming. Section 2: Computing Hardware 4/5: Input and Output Devices

Teaching Introductory Artificial Intelligence with Pac-Man

Definitions. A [non-living] physical agent that performs tasks by manipulating the physical world. Categories of robots

3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving

Hybrid System for Driver Assistance

Industrial Automation Training Academy. PLC, HMI & Drives Training Programs Duration: 6 Months (180 ~ 240 Hours)

TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAA

Safe robot motion planning in dynamic, uncertain environments

COMP 590: Artificial Intelligence

CS 51 Intro to CS. Art Lee. September 2, 2014

The Basics of Robot Mazes Teacher Notes

Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina

Is a Data Scientist the New Quant? Stuart Kozola MathWorks

6.099, Spring Semester, 2006 Assignment for Week 13 1

TomTom HAD story How TomTom enables Highly Automated Driving

SD1 SPEECH DIALLER OPERATING INSTRUCTIONS SD1 ESC ENT

CAD Training Center Sooxma Technologies Ameerpet, Hyderabad. Ph: ,

What is Artificial Intelligence?

Robot Perception Continued

The Basics of Scanning Electron Microscopy

ICT deployment status in factory and development trend under the concept of Industry 4.0

SIP-PROJECT : Preliminary examination of traffic accident simulation to evaluate the benefits of safety systems for the reduction of traffic accidents

LOOP Technology Limited. vision. inmotion IMPROVE YOUR PRODUCT QUALITY GAIN A DISTINCT COMPETITIVE ADVANTAGE.

GEARBOX MONITOR MODEL NR USER MANUAL

Particles, Flocks, Herds, Schools

Find what matters. Information Alchemy Turning Your Building Data Into Money

How To Set Up A Wide Area Surveillance System

SYSTEMS, CONTROL AND MECHATRONICS

INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & TECHNOLOGY (IJCET)

3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving

Cloud Computing for Agent-based Traffic Management Systems

CIM Computer Integrated Manufacturing

About the Author. The Role of Artificial Intelligence in Software Engineering. Brief History of AI. Introduction 2/27/2013

Implementation of a Wireless Gesture Controlled Robotic Arm

Aguantebocawertyuiopasdfghvdslpmj klzxcvbquieromuchoanachaguilleanic oyafernmqwertyuiopasdfghjklzxcvbn mqwertyuiopasdfghjklzxcvbnmqwerty

Tools and Operational Data Available st RESOLUTE Workshop, Florence

Problem-Based Group Activities for a Sensation & Perception Course. David S. Kreiner. University of Central Missouri

ElastMan: Autonomic Elasticity Manager for Cloud-Based Key-Value Stores

Simulation in a Nutshell

ABA AND RDI. Therapy vs. RDI Life Style ABA

Organic Fault-Tolerant Control Architecture for Robotic Applications

University of Salford, UK Centre for Virtual Environments and Future Media. Carl Moore, Toby Duckworth and David Roberts

Safety Car Drive by Using Ultrasonic And Radar Sensors

The role of demand-responsive energy management systems for smart cities

GOAL: Get a Drivers License

Linux Driver Devices. Why, When, Which, How?

Adaptive Cruise Control System Overview

How cloud-based systems and machine-driven big data can contribute to the development of autonomous vehicles

SYSM 6304: Risk and Decision Analysis Lecture 5: Methods of Risk Analysis

Welcome to the Next Generation of Trading Solutions by PandeLa - Intelligent Automated Trading.

Laboratory work in AI: First steps in Poker Playing Agents and Opponent Modeling

Classroom Activities. These educational materials were developed by the Carnegie Science Center <

Graphical Environment Tool for Development versus Non Graphical Development Tool

CS 40 Computing for the Web

Development of effective Urban Road Traffic Management using workflow wechniques for upcoming metro cities like Lucknow (India).

BIOMETRIC SOLUTIONS 2013 ISSUE

May 6, 2014 Presented by: Kyle McNerney

UNDERGRADUATE DEGREE PROGRAMME IN COMPUTER SCIENCE ENGINEERING SCHOOL OF COMPUTER SCIENCE ENGINEERING, ALBACETE

Introduction to Robotics. Vikram Kapila, Associate Professor, Mechanical Engineering

Autonomous Advertising Mobile Robot for Exhibitions, Developed at BMF

Automatic Train Control based on the Multi-Agent Control of Cooperative Systems

Transcription:

CS 331: Artificial Intelligence Intelligent Agents 1 General Properties of AI Systems Sensors Reasoning Actuators Percepts Actions Env vironment This part is called an agent. Agent: anything that perceives its environment through sensors and acts on that environment through actuators 2 1

Example: Vacuum Cleaner Agent Percept Sequence Action [A, Clean] Right [A, Dirty] Suck [B, Clean] Left [B, Dirty] Suck [A, Clean],[A, Clean] Right [A, Clean],[A, Dirty] Suck : : [A, Clean], [A, Clean], [A, Clean] Right [A, Clean], [A, Clean], [A, Dirty] Suck : : Agent-Related Terms Percept sequence: A complete history of everything the agent has ever perceived. Think of this as the state of the world from the agent s perspective. Agent function (or Policy): Maps percept sequence to action (determines agent behavior) Agent program: Implements the agent function 4 2

Question du Jour What s the difference between the agent function and the agent program? 5 Rationality Rationality: do the action that causes the agent to be most successful How do you define success? Need a performance measure Eg. reward agent with one point for each clean square at each time step (could penalize for costs and noise) Important point: Design performance measures according to what one wants in the environment, not according to how one thinks the agent should behave 6 3

Rationality Rationality depends on 4 things: 1. Performance measure of success 2. Agent s prior knowledge of environment 3. Actions agent can perform 4. Agent s percept sequence to date Rti Rational agent: tfor each possible percept sequence, a rational agent should select an action that is expected to maximize its performance measure, given the evidence provided by the percept sequence and whatever built-in knowledge the agent has 7 Learning Successful agents split task of computing policy in 3 periods: 1. Initially, designers compute some prior knowledge to include in policy 2. When deciding its next action, agent does some computation 3. Agent learns from experience to modify its behavior Autonomous agents: Learn from experience to compensate for partial or incorrect prior knowledge 8 4

PEAS Descriptions of Agents Performance, Environment, Actuators, Sensors Example: Automated taxi driver Performance Measure Environment Actuators Sensors Safe, fast, legal, Roads, other Steering, accelerator, Cameras, sonar, comfortable trip, pedestrians, customers brake, signal, horn, speedometer, GPS, maximize profits display odometer, accelerometer, engine sensors, keyboard 9 Properties of Environments Fully observable: can access complete state of environment at each point in time Deterministic: if next state of the environment completely determined by current state and agent s action Episodic: agent s experience divided into independent, atomic episodes in which agent perceives and performs a single action in each episode. Static: agent doesn t need to keep sensing while decides what action to take, doesn t need to worry about ttime vs vs Vs vs Partially observable: could be due to noisy, inaccurate or incomplete sensor data Stochastic: a partially observable environment can appear to be stochastic. (Strategic: environment is deterministic except for actions of other agents) Sequential: current decision affects all future decisions Dynamic: environment changes while agent is thinking (Semidynamic: environment doesn t change with time but agent s performance does) Discrete: (note: discrete/continuous distinction applies to states, time, percepts, or actions) vs Continuous Single agent vs Multiagent: agents affect each others performance measure cooperative or competitive 5

Task Environment Crossword puzzle Examples of task environments Observable Deterministic Episodic Static Discrete Agents Fully Deterministic Sequential Static Discrete Single Chess with a Fully Strategic Sequential Semi Discrete Multi clock Poker Partially Stochastic Sequential Static Discrete Multi Backgammon Fully Stochastic Sequential Static Discrete Multi Taxi driving Partially Stochastic Sequential Dynamic Continuous Multi Medical diagnosis Partially Stochastic Sequential Dynamic Continuous Multi Image analysis Fully Deterministic Episodic Semi Continuous Single Part-picking robot Refinery controller Interactive English tutor Partially Stochastic Episodic Semi Continuous Single Partially Stochastic Sequential Dynamic Continuous Single Partially Stochastic Sequential Dynamic Discrete Multi Agent Programs Agent program: implements the policy Simplest agent program is a table-driven agent function TABLE-DRIVEN-AGENT(percept) returns an action static: percepts, a sequence, initially empty table, a table of actions, indexed by percept sequences, initially fully specific append percept to the end of percepts action LOOKUP(percepts, table) return action This is a BIG table clearly not feasible! 12 6

4 Kinds of Agent Programs Simplex reflex agents Model-based reflex agents Goal-based agents Utility-based agents 13 Simple Reflex Agent Selects actions using only the current precept Works on condition-action rules: if condition then action Not the entire percept history! function SIMPLE-REFLEX-AGENT(percept) returns an action static: rules, a set of condition-action rules state INTERPRET-INPUT(percept) rule RULE-MATCH(state, rules) action RULE-ACTION[rule] return action 14 7

Simple Reflex Agents 15 Advantages: Simple Reflex Agents Easy to implement Uses much less memory than the table-driven agent Disadvantages: Will only work correctly if the environment is fully observable Infinite loops 16 8

Model-based Reflex Agents Maintain some internal state that keeps track of the part of the world it can t see now Needs model (encodes knowledge about how the world works) function REFLEX-AGENT-WITH-STATE(percept) returns an action static: state, a desription of the current world state rules, a set of condition-action rules action, the most recent action, initially none state UPDATE-STATE(state, action, percept) rule RULE-MATCH(state, rules) action RULE-ACTION[rule] return action 17 Model-based Reflex Agents 18 9

Goal-based Agents Goal information guides agent s actions (looks to the future) Sometimes achieving i goal is simple eg. from a single action Other times, goal requires reasoning about long sequences of actions Flexible: simply reprogram the agent by changing goals 19 Goal-based Agents 20 10

Utililty-based Agents What if there are many paths to the goal? Utility measures which states are preferable to other state Maps state to real number (utility or happiness ) 21 Utility-based Agents 22 11

Learning Agents 23 Learning Agents Critic: Tells learning element how well the agent is doing with respect ot the performance standard (because the percepts don t tell the agent about its success/failure) Responsible for improving the agent s behavior with experience Suggest actions to come up with new and informative experiences 24 12

Learning Agents Think of this as outside the agent since you don t want it to be changed by the agent Maps percepts to actions 25 What you should know What it means to be rational Be able to do a PEAS description of a task environment Be able to determine the properties of a task environment Know which agent program is appropriate for your task 26 13

In-class Exercise Develop a PEAS description of the task environment for a face-recognition agent Performance Measure Environment Actuators Sensors 27 In-class Exercise Describe the task environment Fully Observable Deterministic Episodic Static Discrete Single agent Partially Observable Stochastic Sequential Dynamic Continuous Multi-agent 28 14

In-class Exercise Select a suitable agent design for the face- recognition agent 29 15