A Database of Empirical Evaluations of Adaptive Systems

Similar documents
Rosa Arruabarrena, Tomás A. Pérez, J. López-Cuadrado, J. Gutiérrez, and J.A. Vadillo

Developing Adaptive Internet Based Courses with the Authoring System NetCoach

Evaluation of Adaptive Web Systems

A Web-based Framework for User-centred Evaluation of End-User Experience in Adaptive and Personalized E-Learning Systems

Identifying Learning Styles in Learning Management Systems by Using Indications from Students Behaviour

Adapting to the Level of Experience of the User in Mixed-Initiative Web Self-Service Applications

MASSIVE OPEN ONLINE COURSES AN ADAPTIVE LEARNING FRAMEWORK

AN INTELLIGENT TUTORING SYSTEM FOR LEARNING DESIGN PATTERNS

An Approach to Adaptivity and Collaboration Support in a Web-Based Learning Environment

Conference Navigator 2.0: Community-Based Recommendation for Academic Conferences

Single Level Drill Down Interactive Visualization Technique for Descriptive Data Mining Results

Is the ACT-value a valid estimate for knowledge? An empirical evaluation of the inference mechanism of an adaptive help system

Methodology for Creating Adaptive Online Courses Using Business Intelligence

Research Methods & Experimental Design

Exploiting User and Process Context for Knowledge Management Systems

WEPS PEER AND AUTOMATIC ASSESSMENT IN ONLINE MATH COURSES

Cross-Domain Collaborative Recommendation in a Cold-Start Context: The Impact of User Profile Size on the Quality of Recommendation

Building a Moodle front-end for Greek language learning

Task-Model Driven Design of Adaptable Educational Hypermedia

DESIGN OF THE STUDY-RESEARCH METHODOLOGY

Usability metrics for software components

The Evaluation of Alternate Learning Systems: Asynchronous, Synchronous and Classroom

Chapter 2. Sociological Investigation

Modeling Peer Review in Example Annotation

Solar energy e-learning laboratory - Remote experimentation over the Internet

APPLICATION OF А FRENCH MODEL OF SOCIOCULTURAL ANIMATION AT ADOLESCENTS LEISURE TIME

Space project management

An Engagement Model for Learning: Providing a Framework to Identify Technology Services

An Evaluation of the Effects of Web Page Color and Layout Adaptations

AN ELECTRONIC LEARNING ENVIRONMENT FOR APPLIED STATISTICS: QUALITY CARE AND STATISTICS EDUCATION IN HIGHER EDUCATION

Newer and Early Career Researchers Conference Conference Information International Conference for Newer and Early Career Researchers

ORGANIZATIONAL DESIGN AND ADAPTATION IN RESPONSE TO CRISES: THEORY AND PRACTICE

The importance of using marketing information systems in five stars hotels working in Jordan: An empirical study

FoNS guidelines for writing a final project report July

Problem Decomposition Technique and Cooperative Learning in Enriching Software Development Projects

Performance Indicators for the Digital Library 1

Chapter 14: Usability testing and field studies. The aims: Usability testing

A Framework of Context-Sensitive Visualization for User-Centered Interactive Systems

Adaptive Navigation for Self-Assessment Quizzes

The Role of Controlled Experiments in Software Engineering Research

Pilot Testing and Sampling. An important component in the data collection process is that of the pilot study, which

Consulting projects: What really matters

Computer-aided web-based application to modular plant design

USING OPEN STUDENT MODELS IN DISTANCE LEARNING ENVIRONMENTS TO HELP TEACHERS PROVIDE ADAPTIVE TUTORING

Financial and Commercial Services HIRE OF MANAGEMENT IT/IS CONSULTANTS

Performance evaluation of large-scale data processing systems

A Group Formation Tool for e-learning Environments

Mobile GIS for Cadastral Data Collection in Ghana

Point and Interval Estimates

Correlation Map of LEARNING-FOCUSED to Marzano s Evaluation Model

Evaluating Software Products - A Case Study

Section 5 Methodology & Presenting Findings Of Research Proposal

WHAT WORKS IN INNOVATION AND EDUCATION IMPROVING TEACHING AND LEARNING FOR ADULTS WITH BASIC SKILL NEEDS THROUGH FORMATIVE ASSESSMENT STUDY OUTLINE

Prediction of Business Process Model Quality based on Structural Metrics

Designing an Adaptive Virtual Guide for Web Applications

A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling

Rosta Farzan Assisstant Professor School of Information Sciences University of Pittsburgh 135 North Bellefield Avenue, Pittsburgh, PA 15260

Completing a Peer Review

A PANEL STUDY FOR THE INFLUENTIAL FACTORS OF THE ADOPTION OF CUSTOMER RELATIONSHIP MANAGEMENT SYSTEM

One for All and All in One

Usability-Improving Mobile Application Development Patterns

SYSTEM DEVELOPMENT AND IMPLEMENTATION

A QTI editor integrated into the netuniversité web portal using IMS LD

Effect of the learning support and the use of project management tools on project success: The case of Pakistan

King Saud University. Deanship of Graduate Studies. College of Business Administration. Council of Graduate Programs in Business Administration

Nursing Journal Toolkit: Critiquing a Quantitative Research Article

Process-Family-Points

Adaptive Hypermedia Services for E-Learning

Syllabus. 1 Title Experimental Research in Marketing & Management

NORMIT: a Web-Enabled Tutor for Database Normalization

Experimental Design and Hypothesis Testing. Rick Balkin, Ph.D.

Survey Research Methods

Evaluation of the methodology in publications describing epidemiological design for dental research: a critical analysis

Providing Adaptive Courses in Learning Management Systems with Respect to Learning Styles *

Call for Papers Journal of Educational Technology & Society (ISSN: (online) and (print))

Strengthening Canada s Communities / Des communautés plus fortes au Canada

A Socio-Technical Approach towards Supporting Intra- Organizational Collaboration

Improving the outcome of e-learning using new technologies in LMS systems

Soft Skills Requirements in Software Architecture s Job: An Exploratory Study

An Investigation on Learning of College Students and the Current Application Situation of the Web-based Courses

An Empirical Study on Online MBA Cheating and Plagiarism Issues

A Book Proposal for. Building Help Systems That Work SAMPLE. Steven Radecki XXX Shadow Dance Drive San Jose, California (408)

NORTHWESTERN UNIVERSITY Department of Statistics. Fall 2012 Statistics 210 Professor Savage INTRODUCTORY STATISTICS FOR THE SOCIAL SCIENCES

RESEARCH METHODS IN I/O PSYCHOLOGY

The Role of Reactive Typography in the Design of Flexible Hypertext Documents

Chapter 1 Introduction. 1.1 Introduction

The Impact of Adapting Content for Students with Individual Differences

Sample 360 Degree Feedback Report. For Terry Smith. Designed by Human Resources Consulting For XYZ Corporation

The Financial Advisor Opportunity at Edward Jones

On the Design of an Advanced Web-Based System for Supporting Thesis Research Process and Knowledge Sharing

Web services to allow access for all in dotlrn

How To Create A Web Experiment List

Evaluation of Adaptive Course Construction Toolkit (ACCT)

Issues in Information Systems Volume 13, Issue 1, pp , 2012

A GUI Crawling-based technique for Android Mobile Application Testing

Evaluating an eportfolio System: the Case of a Hybrid University Course

NURSES ATTITUDES TOWARDS DEPRESSION: A STUDY IN SLOVENIA

NON-PROBABILITY SAMPLING TECHNIQUES

SSRC. Dissertation Proposal Development Fellowship. Mr. John Doe. Current University

Transcription:

A Database of Empirical Evaluations of Adaptive Systems Stephan Weibelzahl Gerhard Weber Pedagogical University Freiburg Kunzenweg 21 79117 Freiburg, Germany [weibelza,weber]@ph-freiburg.de The development of an online database for studies of empirical evaluations of adaptive systems (EASy-D) is proposed. Such a database will serve as reference for researchers in the field of adaptive systems and as guide for planning new evaluations. It aims at encouraging more evaluations that fulfill certain methodological requirements. The structure of the database records is discussed and the functionality of the web interface is described. 1 Current Evaluations of Adaptive Systems Empirical evaluations of adaptive systems are rare, e.g., only about one out of four articles published in User Modeling and User Adapted Interaction (UMUAI) reports significant empirical evaluations (Chin, 2001). Moreover, it is still controversial which experimental designs and criteria are useful. Researchers who want to evaluate their system empirically have to get an overview about current evaluations to plan their own study. In other fields, there are compendia and guidelines that provide such an overview. However, composing some simple guidelines for evaluation planning is not adequate for adaptive system evaluation at the moment if we consider the thin base of results that is currently available. Thus, a more promising approach is to improve the access to relevant studies by establishing an online database of empirical evaluations of adaptive systems. This paper describes the aims of such a database, what the record structure looks like, and how the system will be maintained and successively extended. 2 Developing a Database of Empirical Evaluations We developed a searchable online database for studies on empirical evaluations of adaptive systems which is available at http://art7.ph-freiburg.de/easy-d. 2.1 Aims The main goal of this database is to encourage empirical evaluations of adaptive systems. By providing a searchable categorized set of studies, interested people get suggestions of experimental designs, criteria, and experimental issues.

Name: the system s full name Function of system: rough category to describe the system. These categories include tailor information presentation recommend products or other objects help user to find information support learning give help adapt interface take over routine tasks support collaboration other function Task: description of the task user and system perform together. If possible the domain is reported as well. Adaptation rate: micro or macro adaptation (Cronbach, 1967); basically a question of whether a long-term or a short-term user model is used Purpose of adaptation: why is adaptation used; which goals or advantages have been the reason for using adaptation Method of adaptation: in which way does adaptation take place (e.g., selection of appropriate level of difficulty of tasks, change layout of display) Figure 1: Description of the system categorization Such a database will help to identify pitfalls in the planning process as well as in the analysis of collected data. Moreover, it will identify lacks in the state of art, e.g., a certain category of systems might appear to be not evaluated at all. For people outside the community the database will serve as reference for the usefulness (or insufficiency) of adaptive systems in general, of certain types of systems, or of a specific system as it describes the current state of the art. 2.2 Record Structure From the user s point of view, there are two kind of entries: experimental studies and systems. While each study evaluates one or more systems, a system which is categorized by a specific function and an adaptation mechanism might be evaluated in several studies. 2.2.1 Systems Each evaluated system is described in terms of its name, the function it fulfills, the task that it performs, and a brief description of the adaptation mechanism. This way of characterizing an adaptive system and most of the categories are adopted from Jameson (1999). See Figure 1 for a detailed description of the categories. The purpose and the method of adaptation are important to help the reader in understanding what the system does. For finding related systems the functioning and the task are probably more important.

References: references where the information about the system was drawn from Evaluated system: name of the evaluated system(s) Evaluation layer: according to the framework proposed by Weibelzahl (2001) a study can be assign to one or more of the following evaluation layers evaluation of input data evaluation of inference evaluation of adaptation decision evaluation of interaction Method of evaluation: a short description of the evaluation, using one of the following categories without running system results of previous research early exploratory studies knowledge acquisition from experts studies with a system controlled evaluations with users controlled evaluations with hypothetical (i.e., simulated) users experience with real world use Data type: brief description of the kind of analyzed data (e.g., observed behavior, questionnaire, interview, log-files, etc.) Criteria: which were the main criteria, and which measures where used, if possible measures (e.g., elapsed time) are grouped in reference to the abstract criterion (e.g., user satisfaction) Criteria categories: one or more of the following categories apply if at least one of the criteria belong to it efficiency effectiveness usability Figure 2: Description of the study categorization 2.2.2 Studies For each study the database provides a citation, a reference to the evaluated system, and a detailed description of the evaluation design (see Figure 2). For experimental studies statistical data and methods of analysis are reported as far as available (see Figure 3). The categories of evaluation method were also adopted from Jameson (1999). The studies are categorized in reference to an evaluation framework introduced by Weibelzahl (2001). This framework proposes to evaluate four different information processing steps in a so-called layered evaluation (Karagiannidis & Sampson, 2000). Recently, two alternative frameworks have been proposed (Brusilovsky, Karagiannidis, & Sampson, 2001; Paramythis, Totter, & Stephanidis, 2001) which could serve as additional categorization. 2.3 Online Interface Users may search for either a system or a study. For finding a related system it is most important to search for a specific function. Searching for a specific name, and full text search are supported as well.

N: number of subjects, sample size k: number of groups or conditions randomization: is the assignment of subjects to groups randomized or quasi-experimental statistical analysis: which statistical methods are used (e.g., ANOVA, MANOVA, correlation) Figure 3: Additional information for experimental studies The presentation of results includes the complete information that is available about this system, as well as a link to studies that evaluated this system. When searching for a related study the user might either fill in a method of evaluation, specify the evaluation layer, or limit the search to a certain data type or criterion. In principle other search criteria (e.g., sample size) would be easy to implement, but appear to be not very useful. In addition, there is a glossary that explains the categories and entries, and a form for authors to submit a new study. The submission procedure is explained in the next section. 3 Implementation and Maintenance EASy-D is based on MySQL and PHP. Currently the database contains 24 studies most of them from UMUAI and User Modeling Conferences. Of course this small number of records would not require a complete database and should be seen as a starting point only. However, we hope that other authors are interested in making studies (either their own studies or papers that are of importance) available in EASy-D. New records are submitted with an online form by categorizing and describing a study and as long as it is not available in the database the evaluated system. However, submissions are reviewed before being published to avoid abuse and to keep entries consistent in terms of language and format. Certainly, submission requires a published study. However, we strongly encourage authors to submit both successful and unsuccessful studies, because especially from the later ones others might learn a lot for planning their own study. Peer reviewed studies are preferred but not required as long as the publication is available for the community. 4 Summary Compared to a usual literature search users of EASy-D may search for system functions and evaluation methods very easy. Moreover, the studies are presented in a standardized way which gives a quick overview of the study. Another advantage is that related studies that evaluate the same system or a system with the same function are identify quickly. We hope that this database will become the central contact point for researchers who are planning empirical evaluations of their adaptive systems and invite everybody to enhance EASy-D by submitting studies or giving feedback. References Brusilovsky, P., Karagiannidis, C., & Sampson, D. (2001). The benefits of layered evaluation of adaptive applications and services. In S. Weibelzahl, D. N. Chin, & G. Weber (Eds.), Empir-

ical Evaluation of Adaptive Systems. Proceedings of workshop at the Eighth International Conference on User Modeling, UM2001 (pp. 1 8). Freiburg. Chin, D. N. (2001). Empirical evaluation of user models and user-adapted systems. User Modeling and User-Adapted Interaction, 11(1-2), 181 194. Cronbach, L. (1967). How can instruction be adapted to individual differences. In R. Gagné (Ed.), Learning and individual differences. Ohio: Columbus. Jameson, A. (1999). User-adaptive systems: An integrative overview. (Tutorial presented at the Seventh International Conference on User Modeling, Banff, Canada, June 20th 1999) Karagiannidis, C., & Sampson, D. G. (2000). Layered evaluation of adaptive applications and services. In P. Brusilovsky & C. S. Oliviero Stock (Eds.), Proceedings of International Conference on Adaptive Hypermedia and Adaptive Web-Based Systems, AH2000, Trento, Italy (p. 343-346). Berlin: Springer. Paramythis, A., Totter, A., & Stephanidis, C. (2001). A modular approach to the evaluation of adaptive user interfaces. In S. Weibelzahl, D. N. Chin, & G. Weber (Eds.), Empirical Evaluation of Adaptive Systems. Proceedings of workshop at the Eighth International Conference on User Modeling, UM2001 (pp. 9 24). Freiburg. Weibelzahl, S. (2001). Evaluation of adaptive systems. In M. Bauer, P. J. Gmytrasiewicz, & J. Vassileva (Eds.), User Modeling: Proceedings of the Eighth International Conference, UM2001 (pp. 292 294). Berlin: Springer.