Multiagent Systems. Lecture 2: Conducting Empirical Investigations in Multiagent Systems and Parallel Models. MAS Evolutionary Learning

Similar documents
Cooperative Multi-Agent Learning: The State of the Art

Self-organized Multi-agent System for Service Management in the Next Generation Networks

Coevolution of Role-Based Cooperation in Multi-Agent Systems

Cloud Computing for Agent-based Traffic Management Systems

D A T A M I N I N G C L A S S I F I C A T I O N

CHAPTER 1 INTRODUCTION

MEng, BSc Computer Science with Artificial Intelligence

Estimation of the COCOMO Model Parameters Using Genetic Algorithms for NASA Software Projects

FC Portugal D Simulation Team Description Paper

A Sarsa based Autonomous Stock Trading Agent

Evolving Multimodal Behavior Through Subtask and Switch Neural Networks

Overview. Swarms in nature. Fish, birds, ants, termites, Introduction to swarm intelligence principles Particle Swarm Optimization (PSO)

Intelligent Agents Serving Based On The Society Information

Selection Procedures for Module Discovery: Exploring Evolutionary Algorithms for Cognitive Science

Hybrid Evolution of Heterogeneous Neural Networks

Automatic Generation of Intelligent Agent Programs

Distribution transparency. Degree of transparency. Openness of distributed systems

Artificial Intelligence and Politecnico di Milano. Presented by Matteo Matteucci

2010 Master of Science Computer Science Department, University of Massachusetts Amherst

Implementation of hybrid software architecture for Artificial Intelligence System

A strategy for stock trading based on multiple models and trading rules. CS 395T Project Report Subramanian Ramamoorthy December 5, 2003

A FORMALIZATION FOR MULTI-AGENT DECISION SUPPORT IN COOPERATIVE ENVIRONMENTS. A FRAMEWORK FOR SITUATED AGENTS. Salvador IBARRA MARTÍNEZ

Ad Hoc Autonomous Agent Teams: Collaboration without Pre-Coordination

On Fleet Size Optimization for Multi-Robot Frontier-Based Exploration

ISSUES IN RULE BASED KNOWLEDGE DISCOVERING PROCESS

Evolutionary robotics, neural networks, video game AI

The Network Structure of Hard Combinatorial Landscapes

Cooperative Multi-Agent Systems from the Reinforcement Learning Perspective Challenges, Algorithms, and an Application

A Stock Trading Algorithm Model Proposal, based on Technical Indicators Signals

Information Broker Agents in Intelligent Websites

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

Machine Learning Introduction

A Systems of Systems. The Internet of Things. perspective on. Johan Lukkien. Eindhoven University

MEng, BSc Applied Computer Science

Learning in Abstract Memory Schemes for Dynamic Optimization

ARTIFICIAL INTELLIGENCE: DEFINITION, TRENDS, TECHNIQUES, AND CASES

Building a Global Payroll Model. Charlotte N. Hodges, CPP

Introduction. Swarm Intelligence - Thiemo Krink EVALife Group, Dept. of Computer Science, University of Aarhus

Master s Program in Information Systems

Advanced Techniques for Solving Optimization Problems through Evolutionary Algorithms

Masters in Human Computer Interaction

Search Algorithm in Software Testing and Debugging

MULTI AGENT-BASED DISTRIBUTED DATA MINING

Data Integration using Agent based Mediator-Wrapper Architecture. Tutorial Report For Agent Based Software Engineering (SENG 609.

Masters in Networks and Distributed Systems

Masters in Computing and Information Technology

CS Master Level Courses and Areas COURSE DESCRIPTIONS. CSCI 521 Real-Time Systems. CSCI 522 High Performance Computing

Multi-agent System based Service Oriented Architecture for Supply Chain Management System (MAS-SOA-SCM)

Content Delivery Network (CDN) and P2P Model

Parallel Programming at the Exascale Era: A Case Study on Parallelizing Matrix Assembly For Unstructured Meshes

Computing Near Optimal Strategies for Stochastic Investment Planning Problems

EL Program: Smart Manufacturing Systems Design and Analysis

Doctor of Philosophy in Computer Science

Evolving Team Darwin United

University of Portsmouth PORTSMOUTH Hants UNITED KINGDOM PO1 2UP

Cloud Computing and Software Agents: Towards Cloud Intelligent Services

Masters in Information Technology

Introducing A Cross Platform Open Source Cartesian Genetic Programming Library

Knowledge Sharing in Software Development

An approach towards building human behavior models automatically by observation

Appendices master s degree programme Artificial Intelligence

ARTIFICIAL INTELLIGENCE METHODS IN EARLY MANUFACTURING TIME ESTIMATION

Simulate Grid Resource Trading via Cognitive Agent: A Case Study

Efficient Reward Functions for Adaptive Multi-Rover Systems

Material and some slide content from: - Software Architecture: Foundations, Theory, and Practice NFPs Reid Holmes Lecture 5 - Tuesday, Sept

HybrID: A Hybridization of Indirect and Direct Encodings for Evolutionary Computation

Master of Science in Computer Science

Beyond Reward: The Problem of Knowledge and Data

PhD in Computer Sciences

Multi-Objective Optimization using Evolutionary Algorithms

GA as a Data Optimization Tool for Predictive Analytics

CONNECTING DATA WITH BUSINESS

Service Oriented Architecture

Mechanical Design/Product Design Process

Masters in Advanced Computer Science

EFFICIENT KNOWLEDGE BASE MANAGEMENT IN DCSP

Masters in Artificial Intelligence

ReSIST NoE Resilience for Survivability in IST. Introduction

Trading. Theory and Practice

Agent Models of 3D Virtual Worlds 1

Multi-Agent Architecture for Implementation of ITIL Processes: Case of Incident Management Process

One for All and All in One

Reaching CMM Levels 2 and 3 with the Rational Unified Process

International Journal of World Research, Vol: I Issue XIII, December 2008, Print ISSN: X DATA MINING TECHNIQUES AND STOCK MARKET

Agent-Based Software and Practical Case of Agent-Based Software Development

Software that writes Software Stochastic, Evolutionary, MultiRun Strategy Auto-Generation. TRADING SYSTEM LAB Product Description Version 1.

6.254 : Game Theory with Engineering Applications Lecture 1: Introduction

ANT COLONY OPTIMIZATION APPLIED TO AN AUTONOMOUS MULTIAGENT GAME

Genetic algorithms for changing environments

walberla: Towards an Adaptive, Dynamically Load-Balanced, Massively Parallel Lattice Boltzmann Fluid Simulation

Draft dpt for MEng Electronics and Computer Science

Visionet IT Modernization Empowering Change

CLOUD COMPUTING: A NEW VISION OF THE DISTRIBUTED SYSTEM

How To Create A Text Classification System For Spam Filtering

A1 Introduction to Data exploration and Machine Learning

Use of Agent-Based Service Discovery for Resource Management in Metacomputing Environment

Applying 4+1 View Architecture with UML 2. White Paper

Agent-Oriented Software Engineering PORTO Methodology AIAD 2013/2014. António Castro and Eugénio Oliveira

Professional Organization Checklist for the Computer Science Curriculum Updates. Association of Computing Machinery Computing Curricula 2008

Review of Computer Engineering Research CURRENT TRENDS IN SOFTWARE ENGINEERING RESEARCH

Transcription:

Multiagent Systems Distributed Artificial Intelligence Lecture 2: Conducting Empirical Investigations in Multiagent Systems and Parallel Models Lecturers: Adrian Grajdeanu & Liviu Panait Distributed Problem Solving Agents: (Wooldridge and Jennings, 95) - weak agents - autonomy - social ability - reactivity - pro-activeness - strong agents - beliefs, desires, intentions - knowledge, commitments, trust Multiagent Systems (MAS) Multiagent Systems: (Jennings et al, 98) - agents with incomplete information - decentralized control - decentralized data - asynchronous computation Multiagent Learning MAS Evolutionary Learning Learning = improving performance with accumulated experience, as indicated by a metric measure (Mitchell, 97) Multiagent Learning = improving the performance of individual agents or of teams of agents in a MAS setting We assume learning affects more than one agent Throughout the lecture, EC is THE learning technique

Team Learning EC particularly suited for team learning An individual codes for the behavior of an entire team Relatively similar to standard EC Team composition domain specific (soccer) scalable to larger teams (MAV) Team Learning Approaches Homogeneous Team Learning an individual contains a single behavior used for all agents fast, scalable, possible suboptimal results Heterogeneous Team Learning an individual contains one behavior for each of the agents can potentially code for any homogeneous team slower, potentially non-scalable, allows agent specialization restricted inter-breeding may be better (Luke and Spector, 96) Hybrid Team Learning an individual codes for team behaviors composed of heterogeneous groups of homogeneous behaviors usually requires additional parameters for coding and manipulating the hybrid teams breeding for hybrid teams? learning team decomposition (Hara and Nagao, 99) Heterogeneity and Performance Adding heterogeneity increases performance IF ENOUGH TIME IS AVAILABLE and in domains that require task specialization (Balch, 98) in inherently decomposable domains (Bongard, 00) in domains that require increased number of different skills (Potter et al, 01) Empirical Investigations in Team Learning Empirical investigations in team learning are very similar to those in standard EC analysis of performance is straightforward best-so-far curves standard statistics/visualization tools How to measure scalability? plot learning curves for different numbers of agents plot performance versus number of agents How to measure heterogeneity? problematic in GP How to measure and quantify the relation between heterogeneity and domain features?

Teammate Learning Teammate Learning Introduction Research Directions Issues Conduction Empirical Investigations Teammate Learning Coevolution for MAS learning Agents performing own learning processes decentralized learning closer to the concept of MAS EC learning for agent 1 EC learning for agent 2 Teammate learning better than team learning (Iba, 96, 98) Teammate learning worse than team learning (Miconi, 03) Theoretical comparisons (Jansen and Wiegand, 03) EC learning for agent 3 EC learning for agent 4

Teammate Learning Teammate Learning Introduction Research Directions Issues Conduction Empirical Investigations Optimality Search influenced by performance and balance (Panait et al, 03) Cooperative tasks with joint reward functions standard algorithms not guaranteed to find optima, even with relaxed settings robustness of solutions? better when teamed with optimal collaborator better when teamed with many other collaborators Competitive tasks what is optimal? duel methodology renaissance-man methodology Good news: fertile area for future research Influences performance Influences heterogeneity locality of reward domain particularities Locality of Reward learning speed team heterogeneity performance Future research opportunities: automatic adjustment of locality

Cooperation or Competition Teammate Modeling No clear relation among agents, relations might change over time Learning opportunities manipulation exploitation of other agents faults mutual trust reciprocity Recursive modeling Flavors single focus of learning modeling combined with learning Initial beliefs are VERY important Teammate Learning Issues in Team Learning Introduction Research Directions Issues Conduction Empirical Investigations Search Space Red Queen Effect Exploration Credit Assignment Learning Cycles Loss of Gradient Mediocre Stability

Search Space for Team Learning Search Space for Teammate Learning Agent 2 R 11 b 2 R 13 R 21 b 3 R 23 R 31 Agent 1 a 1 a 2 a 3 b 1 R 12 R 22 R 32 R 33 Agent 1 Agent 2 a 1 a 2 a 3 R a1 R a2 R a3 b 1 b 2 b 3 R b1 R b2 R b3 Search Space for Teammate Learning Search Space for Teammate Learning Agent 1 Agent 2 a 2 a 3 b 1 b 2 b 3 R 0 a1 R 0 a2 R 0 a2 a 1 R 0 R 0 b1 R 0 b2 b2 R 1 a1 R 1 a2 R 1 a2 R 1 b1 R 1 b2 R 1 b2 R 2 a1 R 2 a2 R 2 a2 R 2 b1 R 2 b2 R 2 b2 R 3 a1 R 3 a2 R 3 a2 R 3 b1 R 3 b2 R 3 b2........................................................

Red Queen Effect Exploration An agent s exploration process affects the learning processes of other agents, with later repercussions on the agent s learning process Similar to an agent learning in a dynamic environment, where the dynamicity is directly related to the agent s behavior Agent 1 Agent 2 a 1 a 2 a 3 b 1 b 2 b 3 R 0 Change in a moving landscape may go unnoticed Individuals are evaluated in the context of other individuals Subjective performance metrics may hide progress, stagnation, or learning cycles R 0 a1 R 0 a2 R 0 a2 R 1 a1 R 1 a2 R 1 a2 R 2 a1 R 2 a2 R 2 a2 R 3 a1 R 3 a2 R 3 a2 R 0 b1 R 0 b2 b2 R 1 b1 R 1 b2 R 1 b2 R 2 b1 R 2 b2 R 2 b2 R 3 b1 R 3 b2 R 3 b2............................................ Credit Assignment Learning Cycles Inter- and intra- agent credit assignment Individual reinforcement information may influence agents to learn greedy strategies focused on individual, rather than team, performance wins against Agent 1 using strategy 1 Agent 2 using strategy 1 looses against Agent 1 using strategy 1 Agent 2 using strategy 2 wins against Agent 1 using strategy 2 Agent 2 using strategy 2 looses against Agent 1 using strategy 2 Agent 2 using strategy 1

Loss of Gradient Mediocre Stability Teammate Learning Introduction Research Directions Issues Conduction Empirical Investigations Empirical Investigations in Teammate Learning What is being measured? Performance is subjective (Red Queen Effect) possible solutions choose domains where objective performance measure is available (Panait and Luke, 02), (Bucci and Pollack, 03) use benchmarks dominance tournament (Stanley and Miikkulainen, 02) hall of fame? (Rosin and Belew, 97) measure for team heterogeneity? measure for sizes of basins of attractions? What is meant by better or best? (Panait and Luke, 02) duel methodology renaissance-man methodology

Empirical Investigations in Teammate Learning What are the assumptions of the experiments? global information does not guarantee optimality recommendations to restrict assumptions about other agents when their behaviors are unknown coevolution may be improved when assuming other agents are competing or cooperating How to select problem domains? my method is better than your method stage of investigation for theoretical analysis, use very simple domains (game matrixes) pay attention to assumptions Empirical Investigations in Teammate Learning Visualization visualization needs to capture the relation among different coevolutionary algorithms plot the trajectories of the search process search driven by balance and performance visualization of search space: basins of attraction? assess difficulty of domain based on sizes of basins of attraction for suboptimal peaks Statistical methods because performance assessment is subjective, the results of statistical tests will depend on the other components co-adaptation and learning cycles time may be an especially important characteristic assess performance based on final results for all agents Flavors Problem Decomposition task decomposition behavior decomposition layered learning shaping Questions: automatic problem decomposition decentralized problem decomposition

Communication MAS + unrestricted communication = centralized system (Stone and Veloso, 00) Via rapidly decaying information may increase the search space may improve performance emergent vocabularies Via slowly decaying information (example: pheromones) long-lasting shared information Via embodiment Empirical Investigations and Communication What are the assumptions? Additional parameters to tune range, bandwidth evaporation and diffusion rates communication topologies How to measure relation between learning algorithm and communication? Emergent vocabularies? Test communication via embodiment? Conclusions Conclusions Empirical investigations in team learning pretty much straightforward analysis of heterogeneity and scalability Empirical investigations in teammate learning subjective evaluation no clear performance criteria visualize and measure balance and its relation to performance as the components driving the search process assumptions about other agents are very important Empirical investigations and problem decomposition representations Empirical investigations and communication assumptions test of emergent vocabularies test of communication via embodiment

M. Wooldridge and N. Jennings, (1995), Intelligent Agents: Theory and Practice, The Knowledge Engineering Review, 10(2):115-152 N.R. Jennings, K. Sycara, and M. Wooldridge, (1998), A Roadmap of Agents Research and Development, Autonomous Agents and Multi-Agent Systems, 1:7-38 T. Mitchell, (1997), Machine Learning, McGraw-Hill S. Luke and L. Spector, (1996), Evolving Teamwork and Coordination with Genetic Programming, In Genetic Programming 1996: Proceedings of the First Annual Conference, MIT Press A. Hara and T. Nagao, (1999), Emergence of cooperative behavior using ADG; Automaticaly Defined Groups, In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO- 99) T. Balch, (1998), Behavioral Diversity in Learning Robot Teams, PhD Thesis, College of Computing, Georgia Institute of Technology J. Bongard, (2000), The Legion System: A Novel Approach to Evolving Heterogeneity for Collective Problem Solving, In Genetic Programming: Proceedings of EuroGP'2000, Springer-Verlag M. Potter, L. Meeden, and A. Schultz, (2001), Heterogeneity in the Coevolved Behaviors of Mobile Robots: The Emergence of Specialists, In Proceedings of The Seventeenth International Conference on Artificial Intelligence (IJCAI-2001) H. Iba, (1996), Emergent Cooperation for Multiple Agents Using Genetic Programming, In Parallel Problem Solving from Nature IV: Proceedings of the International Conference on Evolutionary Computation H. Iba, (1998), Evolutionary Learning of Communicating Agents, Information Sciences, Volume 108 T. Miconi, (2003), When Evolving Populations is Better than Coevolving Individuals: The Blind Mice Problem, In Proceedings of the Thirteenth International Joint Conference on Artificial Intelligence (IJCAI-03) T. Jansen and R. P. Wiegand, (2003), Exploring the Explorative Advantage of the Cooperative Coevolutionary (1+1) EA, In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2003) L. Panait, R. P. Wiegand, and S. Luke, (2003), Improving Coevolutionary Search for Optimal Multiagent Behaviors, In Proceedings of the Thirteenth International Joint Conference on Artificial Intelligence (IJCAI-2003) L. Panait and S. Luke, (2002), A Comparison of Two Competitive Fitness Functions, In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2002) A. Bucci and J. Pollack, (2003), Focusing versus Intransitivity: Geometrical Aspects of Co-evolution, In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2003) K. O. Stanley and R. Miikkulainen, (2002), The Dominance Tournament Method of Monitoring Progress in Coevolution, In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO-2002) Workshop Program C. Rosin and R. Belew, (1997), New Methods for Competitive Coevolution, Evolutionary Computation Journal, 5(1):1-29 P. Stone and M. Veloso, (2000), Multiagent Systems: A Survey from a Machine Learning Perspective, Autonomous Robots, 8(3):345-383