Standards for Internet-Based Experimenting
|
|
|
- Donna McCoy
- 10 years ago
- Views:
Transcription
1 REPRINT, cite as Reips, U.-D. (2002). Standards for Internet-based experimenting. Experimental Psychology, 49 (4), Standards for Internet-Based Experimenting Ulf-Dietrich Reips Experimental and Developmental Psychology, University of Zürich, Switzerland Abstract. This article summarizes expertise gleaned from the first years of Internet-based experimental research and presents recommendations on: (1) ideal circumstances for conducting a study on the Internet; (2) what precautions have to be undertaken in Web experimental design; (3) which techniques have proven useful in Web experimenting; (4) which frequent errors and misconceptions need to be avoided; and (5) what should be reported. Procedures and solutions for typical challenges in Web experimenting are discussed. Topics covered include randomization, recruitment of samples, generalizability, dropout, experimental control, identity checks, multiple submissions, configuration errors, control of motivational confounding, and pre-testing. Several techniques are explained, including warm-up, high hurdle, password methods, multiple site entry, randomization, and the use of incentives. The article concludes by proposing sixteen standards for Internet-based experimenting. Key words: Internet-based experimenting, Web experiment, standards, experiment method, psychological experiment, online research, Internet research, Internet science, methodology Introduction We are in the midst of an Internet revolution in experimental research. Beginning in the mid-nineties of the last century, using the world-wide network for experimental research became the method of choice for a small number of pioneers. Their early work was conducted soon after the invention of forms on Web pages established user-server interaction (Musch & Reips, 2000). This medium holds the promise to achieve further methodological and procedural advantages for the experimental method and a previously unseen ease of data collection for scientists and students. Several terms are used synonymously for Internet-based experiments: Web experiment, on(-)line experiment, Web-based experiment, World Wide Web (WWW) experiment, and Internet experiment. Here the term Web experiment will be used most often because historically this term was used first and experiments delivered via the Web are clearly the most accessible and popular, as experiments using Internet I would like to thank Tom Buchanan, William C. Schmidt, Jochen Musch, Kevin O Neil and an anonymous reviewer for very helpful comments on earlier versions of this article. services other than the Web (such as , ICQ, Telnet, Gopher, FTP, etc.) are rarely conducted. Many of the issues discussed in this article apply to experiments conducted via these services as well, and even to nonexperimental Internet-based methods, such as Web surveying or nonreactive data collection (for examples see Reips & Bosnjak, 2001). Web experiments may be used to validate results from field research and from laboratory experiments (see Krantz & Dalal, 2000; Pohl, Bender, & Lachmann, 2002; Reips, 1997; Reips, Morger, & Meier, 2001), or they may be used for new investigations that could only be feasibly accomplished in this medium. For instance, in 2000, Klauer, Musch, and Naumer published an article on belief bias in Psychological Review that contains an example of a study that reasonably could only be conducted as an Internet-based experiment, because several thousand participants were needed to obtain accurate estimates of model parameters. Similarly, Birnbaum (2001) recruited experts in decision making via a decision making researchers list and sent these experts to a Web page where many of them contradicted their own theories in a Web experiment on choices between gambles. Experiments such as these would prove impractical and burdensome if delivered in another medium. DOI: // Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
2 244 Ulf-Dietrich Reips In addition to the benefit of increased access to participants, Internet-based experimenting always includes the possibility to use the programmed experimental materials in a traditional laboratory setting (Reips, 1997). In contrast, a laboratory experiment, even if built with Internet software technologies, cannot simply be turned into a Web experiment by connecting it to the Internet. Successful and appropriate use of the Web medium requires careful crafting and demands methodological, procedural, technical, and ethical considerations to be taken into account! While laboratory experiments can be built directly with Internet software technologies, it seems wise to conceptualize experiments as Web experiments whenever possible, given their many advantages. Sheer numbers, reduced cost, and accessibility of specific participants are only a few of the Internetspecific properties in Web experimenting that create an environment that has been greeted with great enthusiasm by experimental psychologists. When and When not to Conduct an Experiment on the Internet? Before standards for Internet-based experimenting can be established, a few words should be devoted to the question of criteria that should be used in deciding the mode an experiment is best conducted in. Implementing an Experiment: A General Principle Because many laboratory experiments are conducted on computers anyway, nothing is lost when an experiment is designed Web-ready: It can always also be used in the laboratory. In distributed Web experimenting, local collaborators recruit and assist participants who all log onto the same Internet-based experiment (Reips, 1999). Solving Long-Standing Issues in Experimental Research The experimental method has a long and successful tradition in psychological research. Nevertheless, the method has been criticized, particularly in the late 1960s and early 1970s (e.g., Chapanis, 1970; Orne, 1962; Rosenthal, 1966; Rosenthal & Fode, 1973; Rosenthal & Rosnow, 1969; Smart, 1966). This criticism is aimed in part at the validity of the method and in part at improper aspects of its realization; for instance experimenter effects, volunteer bias, and Experimental Psychology 2002; Vol. 49(4): 243Ð256 low power. A solution for many of these problems could be the implementation of experiments as Web experiments. In total, about eighteen advantages counter seven disadvantages of Web experimenting (Reips, 2000, see Table 1). Why Experimenters Relish Internet-Based Experimenting Speed, low cost, external validity, experimenting around the clock, a high degree of automation of the experiment (low maintenance, limited experimenter effects), and a wider sample are reasons why the Internet may be the setting of choice for an experiment. Seventy percent of those who have conducted a Web experiment intend to certainly use this method again (the other 30 % maybe). This result from a survey of many of the early pioneers in Web experimenting conducted by Musch and Reips (2000) is indirect evidence that learning and using the methods of Internet-based experimenting is certainly worthwhile. Surveyed Web experimenters rated large number of participants and high statistical power as the two most important factors why they made the decision to conduct a Web experiment. The survey conducted by Musch and Reips is in itself a good example that Internet-based research may be the method of choice if a special subpopulation is to be reached. Access to specific groups can be achieved through Internet newsgroups (Hewson, Laurent, & Vogel, 1996; Schmidt, 1997), Web site guestbooks, chat forums, or topic-related mailing lists. Eichstaedt (2002, Experiment 1) recruited persons using either Macintosh or Windows operating systems for his Web experiment via newsgroups devoted to the discussion of issues related to these operating systems. The participants performed a Javabased tachistoscopic word recognition task that included words typically used in ads for these computer systems. Word recognition was faster for words pertaining to a participant s computer system. Some target groups may be easier to study via Internet, because persons belonging to this group will only reveal critical information under the protection of anonymity, for example drug dealers (Coomber, 1997), or Ecstasy users (Rodgers, Buchanan, Scholey, Heffernan, Ling, & Parrott, 2001). When Not to Conduct an Experiment on the Internet Obviously, Web experiments are not the most suitable method for all research projects. For instance whenever physiological parameters of participants 2002 Hogrefe & Huber Publishers
3 Standards for Internet-Based Experimenting 245 Table 1. Web Experiments: Advantages, Disadvantages and Solutions (Adapted from Reips, 2000) Advantages of Web Experiments Disadvantages with Solutions (1) Ease of access to a large number of demographi- (1) Possible multiple submissions Ð can be avoided cally and culturally diverse participants (for an exam- or controlled by collecting personal identification ple of a study conducted in three languages with 440 items, by checking internal consistency as well as women from more than nine countries see in this vol- date and time consistency of answers (Schmidt, ume Bohner, Danner, Siebler, & Samson, 2002); 1997), and by using techniques such as sub-sampling, participant pools, or handing out passwords (Reips, (2)... as well as to rare and specific participant pop- 1999, 2000, 2000b). There is evidence that multiple ulations (Schmidt, 1997). submissions are rare in Web experiments (Reips, (3) Better generalizability of findings to the general 1997). population (Horswill & Coster, 2001; Reips, 1995). (2) Generally, experimental control may be an issue (4) Generalizability of findings to more settings and in some experimental designs, but is less of an issue situations (because of high external validity, e.g., when using between-subjects designs with random dis- Laugwitz, 2001). tribution of participants to experimental conditions. (5) Avoidance of time constraints. (3) Self-selection can be controlled by using the (6) Avoidance of organizational problems, such as multiple site entry technique. scheduling difficulties, as thousands of participants (4) Dropout is always an issue in Web experiments. may participate simultaneously. However, dropout can be turned into a detection device for motivational confounding. Also, dropout can (7) Highly voluntary participation. be reduced by implementing a number of measures, (8) Ease of acquisition of just the optimal number of such as promising immediate feedback, giving finanparticipants for achieving high statistical power while cial incentives, and by personalization (Frick, being able to draw meaningful conclusions from the Bächtiger, & Reips, 2001). experiment. (5) The reduced or absent interaction with partici- (9) Detectability of motivational confounding. pants during a Web experiment creates problems, if (10) Reduction of experimenter effects. instructions are misunderstood. Possible solutions are pretests of the materials and providing the partici- (11) Reduction of demand characteristics. pants with the opportunity for giving feedback. (12) Cost savings of laboratory space, personnel (6) The comparative basis for the Web experiment hours, equipment, administration. method is relatively low. This will change. (13) Greater openness of the research process (in- (7) External validity of Web experiments may be limcreases replicability). ited by their dependence on computers and networks. (14) Access to the number of nonparticipants. Also, many studies cannot be done on the Web. However, where comparable, results from Web and labora- (15) Ease of comparing results with results from a lo- tory studies are often identical (Krantz & Dalal, cally tested sample. 2000). (16) Greater external validity through greater technical variance. (17) Ease of access for participants (bringing the experiment to the participant instead of the opposite). (18) Public control of ethical standards. are to be measured directly, specialized hardware is required, or when a tightly controlled setting is important, then laboratory experiment administration is still required. A further basic limitation lies in Web experiments dependency on computers and networks having psychological, technical, and methodological im- plications. Psychologically, participants at computers will likely be subject to self-actualization and other influences in computer-mediated communication (e.g., Bargh, McKenna, & Fitzsimons, 2002; Joinson, 2001). Technically, more variance is introduced in the data when collected on the Internet than in the laboratory, because of varying network connection 2002 Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
4 246 Ulf-Dietrich Reips speed, varying computer speed, multiple software running in parallel, etc. (Reips, 1997, 2000). On first view one may think that the Internet allows for easy setup of intercultural studies, and it is certainly possible to reach people born into a wide range of cultures. However, it is a widespread misunderstanding that Internet-based cultural research would somehow render unnecessary the use of many techniques that have been developed by cross-cultural psychologists (such as translation Ð back translation, use of the studied cultures languages, and extensive communication and pretests with people from the cultures that are examined). Issues of access, self-selection, and sampling need to be resolved. In many cultures, English-speaking computer users are certainly not representative of the general population. Nevertheless, these people may be very useful in bridging between cultures, for instance, in cooperative studies based on distributed Web experimenting. Finally, for ethical reasons, many experiments cannot be conducted that require an immediate debriefing and adjunctive procedures through direct contact whenever a participant terminates participation. In the following section we will turn to central issues and resulting proposals for standards that specifically apply to Internet-based experimenting. Checks and Solutions for Methodological Challenges in Web Studies The following section contains methodological and technical procedures that will reduce or alleviate issues that are rooted within the very nature of Internet-based studies. Web Experiment Implementation Data collection techniques on the Internet can be polarized into server-side and client-side processing. Server-side methods (a Web server, often in combination with a database application, serves Web pages that can be dynamically created depending on a user s input, Schmidt, 2000) are less prone to platformdependent issues, because dynamic procedures are performed on the server so that they are not subject to technical variance. Client-side methods use the processing power of the participants computers. Therefore, time measurements do not contain error from network traffic and problems with server availability are less likely. Such measurements do rely on Experimental Psychology 2002; Vol. 49(4): 243Ð256 the user s computer configuration however (Schmidt, 2000). Combinations of server-side and client-side processing methods are possible; they can be used to estimate technical error variance by comparison of measurements. General experimental techniques apply to Web experimentation as well. For example, randomized distribution of participants to experimental conditions is a measure against confounding and helps avoiding order effects. Consequently, in every Web experiment at least one randomization technique should be used. In order of reliability, these are roughly: (1) CGI or other server-side solutions, (2) client-side Java, (3) Javascript, and (4) the birthday technique (participants pick their experimental conditions by mouse-clicking on their birthday or birthday month; Birnbaum, 2000; Reips, 2002b). Generating an Experiment For several years, Web experimenters created their experimental materials and procedures by hand. With enough knowledge about HTML and the services and structures available on the Internet conducting a Web experiment was only moderately complicated. However, many researchers hesitate before acquiring new technical skills. Fortunately, several recently developed applications considerably ease the development of a Web experiment. For within-subjects designs, Birnbaum (2000) developed FactorWiz, 1 a Web page that creates Web pages with items combined according to previously defined factorial designs. WEXTOR, 2 by Reips and Neuhaus (2002), a Web-based tool for generating and visualizing experimental designs and procedures, guides the user through a ten-step program of designing an experiment that may include between-subjects, withinsubjects, and quasi-experimental (natural) factors. WEXTOR automatically creates the experiments in such a way that certain methodological requirements of Internet-based experimentation are met (for example, nonobvious file naming [Reips, 2002a] is implemented for experimental materials and conditions, and a session ID is generated that helps identify submissions by the same participant). Recruitment A Web experiment can be announced as part of the collection of Web studies by the American Psycho- 1 factorwiz.htm Hogrefe & Huber Publishers
5 Standards for Internet-Based Experimenting 247 logical Society. 3 This Web site is maintained by John Krantz. A virtual laboratory for Web experiments is the Web Experimental Psychology Lab 4. This Web site is visited by about 4500 potential participants a month (Reips, 2001). Internet-based experiments should always be linked to the web experiment list, 5 a Web site that is intended to serve as an archive of links and descriptions of as many experiments conducted on the Internet as possible. Other ways of recruiting participants that may be combined with linking to experiment Web sites is the use of online panels, newsgroups, search engines, banners, and lists. Also, participants for Web experiments can be recruited offline (e.g., Bamert, 2002; Eichstaedt, 2002; Pohl et al., 2002; Reips et al., 2001; Ruppertsberg, Givaty, Van Veen, & Bülthoff, 2001). Generalizability Self Selection In many areas of Psychology self-selection is not considered much of a problem in research because theory testing is the underlying model of epistemology and people are not considered to vary much on the essential criteria, for example, in research on cognition and perception. However, at least in research more socially oriented, self-selection may interfere with the aim of the study at hand and limit its generalizability. The presence and impact of self-selection in an Internet-based study can be tested by using the multiple site entry technique (Reips, 2000, 2002b). Via log file analysis it is possible to determine a Web experiment s degree of appeal for participation for each of the samples associated with referring Web sites. The multiple site entry technique can be used in any Internet-based study (for a recent example of longitudinal trauma survey research implementing this technique see Hiskey & Troop, 2002). Several links to the study are placed on Web sites, in discussion groups, or other Internet forums that are likely to attract different types of participants. Placing identifying information in the published URLs and analyzing different referrer information in the HTTP protocol can be used to identify these sources (see Schmidt, 2000). Later the data sets that were collected are compared for differences in relative degree of appeal (measured via dropout), demographic data, central results, and data quality, as a function of referring location. Consequently, an estimate of biasing WebExpPsyLab.html 5 potential through self-selection can be calculated (Reips, 2000). If self-selection is not found to be a problem, results from psychological experiments should be the same no matter where participants are recruited, and should therefore show high generalizability. Dependence on Technology Limited generalizability of results from Internetbased research may also arise due to dependence on computers and networking technology. Web-based experimenting can be seen as a form of computermediated communication. Hence, differences found in comparisons between behaviors in computer-mediated situations and face-to-face situations (see, for example, Buchanan, 2002; Kiesler & Sproull, 1986; Postmes, Spears, Sakhel & DeGroot, 2001) need to be taken into account when results from online research are interpreted. Advantages Apart from the challenges to generalizability mentioned above, Internet-based experimenting has three major advantages in this respect: (1) Increased generalizability through nonlocal samples with a wider distribution of demographic characteristics (for a comparison of data from several Web experiments see Krantz & Dalal, 2000, and Reips, 2001). (2) Ecological validity: the experiment comes to the participant, not vice versa (Reips, 1995, 1997). Participants in Web experiments often remain in familiar settings (e.g., at their computer at home or at work) while they take part in an Internet-based experiment. Thus, any effects cannot be attributed to being in an unfamiliar setting. (3) The high degree of voluntariness Ð because there are fewer constraints on the decisions to participate and to continue participation, the behaviors observed in Internet-based experiments may be more authentic and therefore can be generalized to a larger set of situations (Reips, 1997, 2000). A high degree of voluntariness is a trade-off for potential effects of self-selection. Voluntariness refers to the voluntary motivational nature of a person s participation, during the initial decision to participate and during the course of the experiment session. It is influenced by external factors, for example, the setting, the experimenter, and institutional regulations. If participants in an experiment subjectively feel that they are participating entirely voluntarily as 2002 Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
6 248 Ulf-Dietrich Reips opposed to feeling somewhat coerced into participating (as is often the case in the physical laboratory), then they are more likely to initiate and complete an experiment, and to comply with the instructions. The participant s impression of voluntariness reduces effects of psychological reactance (actions following an unpleasant feeling brought about by a perceived threat to one s freedom Ð such as being made to do something by an experimenter); effects such as careless responding, deliberately false answers, and ceasing participation. Therefore, the dropout rate in Web experiments can serve as an indicator of the reactance potential of an experiment. Generally, participant voluntariness is larger in Web experiments than in laboratory experiments (Reips, 1997, 2000). Although Web experiments can introduce a reliance on technology, their improved ecological validity may be seen as a counterweight to this factor when contemplating arguments in favor and opposition of one s choice of method. Dropout: Avoiding it, Reducing its Impact, Using it Dropout curves (dropout progress), or at least dropout rates (attrition rates, or the opposite: return rates) should be reported for all studies conducted on the Internet, separately for all between-subjects experimental conditions. Dropout may pose similar problems as self-selection. However, dropout concerns the end of participation (a comprehensive discussion of other forms of nonresponse is found in Bosnjak, 2001) instead of a decision to initiate participation. If participants end their participation selectively, then the explanatory power of the experiment is severely compromised, especially if dropout varies systematically with levels of the independent variable(s) or with certain combinations of levels. For example, if the impact of two different tasks is to be measured, one of which is more boring, then participants are more likely to drop from that condition. In this case motivational confounding is present (Reips, 2000, 2002a, 2002b). Two types of evasive measures can be taken to counter the problem of dropout: (1) reducing dropout, and (2) reducing the negative impact of dropout. Measures to be Taken in Reducing Dropout Dropout is reduced by all measures that form a motivating counterweight to the factors promoting dropout. For instance, some dropout reducing measures include the announcement of chances for financial incentives and questions about personal information Experimental Psychology 2002; Vol. 49(4): 243Ð256 at the beginning of the Web experiment (Frick et al., 2001; Musch & Reips, 2000; but see O Neil & Penrod, 2001). It is potentially biasing to use scripts that do not allow participants to leave any items unanswered. If the average time needed to fill in an experiment questionnaire is longer than a few minutes, then data quality may suffer from psychological reactance or even anger by those participants who do not wish to answer all questions, because some participants will fill in answers randomly and others will drop out. If all or most of the items in a study are placed on a single Web page then meaningful dropout rates can not be reported: we do not know at which item people decide to drop out. In the terminology of Bosnjak (2001), such a design is not suited to distinguish unit nonresponders from item nonresponders, item nonresponding dropouts, answering dropouts, lurking dropouts, lurkers, and complete responders. They all appear as either dropouts or complete responders, and it becomes impossible to measure a decision not to participate separately from a decision to terminate participation. Consequently, a oneitem-one-screen design or at least a multipage design (with a parcel of items on each page) is recommended. Layout, design, loading time, appeal, and functionality of the experimental materials play a role as well (Reips, 2000, 2002b; Wenzel, 2001). Sophisticated technologies and software used in Internet research (e.g., Javascript, Flash, multimedia elements such as various audio or streaming video formats as part of Web experiments or experiments written in Authorware or as Java applets) may contribute to dropout (Schmidt, 2000). Potential participants unable or unwilling to run such software will be excluded from the start, and others will be dropped, if resident software on their computer interacts negatively with these technologies (Schwarz & Reips, 2001). Findings by Buchanan and Reips (2001) indicate that samples obtained using such technologies will inevitably incorporate biases: People using certain technologies differ psychologically from those who don t. Buchanan and Reips showed that there are personality differences between different groups of respondents: Mac users scored higher on Openness, and people in the Javascriptenabled condition had lower average education levels than those who could or did not have Javascript enabled in their browsers. Consequently, whenever possible, Internet studies should be conducted using only the most basic and widely available technology. Measures to be Taken in Reducing the Negative Impact of Dropout Sometimes, there might be no alternative to avoiding the negative impact of dropout by forgoing self-se Hogrefe & Huber Publishers
7 Standards for Internet-Based Experimenting 249 Table 2. Checklist High-Hurdle Technique Component Seriousness Personalization Impression of control Patience: Loading time Patience: Long texts Duration Privacy Preconditions Technical pretests Rewards Description Tell participants participation is serious, and that science needs good data. Ask for address and/or phone number, and other personal data. Tell participants that their identity can be traced (for instance, via their computer s IP address). Use image files to successively reduce the loading time of Web pages. Place most text on the first page of the Web experiment, and reduce amount page by page. Give an estimate of how long participation in the Web experiment will take. Prepare the participants for any sensitive aspects of your experiment (e.g., you will be asked about your financial situation ). Name software requirements (and provide hyperlinks for immediate download). Perform tests for compatibility of Java, JavaScript, and other technologies, if applicable. Indicate extra rewards for full compliance. lected Internet participation and choosing to conduct the study with a precommitted offline sample. If the absence of dropout is crucial, a dropout-oriented experimental design may help. Three simple techniques can be used to reduce dropout: high hurdle, seriousness check, and warm-up. All three of these techniques have been repeatedly used in the Web Experimental Psychology Lab (e.g., Musch & Klauer, 2002; Reips, 2000, 2001; Reips et al., 2001). In the high- hurdle technique, motivationally adverse factors are announced or concentrated as close to the beginning of the Web experiment as possible. On the following pages, the concentration and impact of these factors should be reduced continuously. As a result, the highest likelihood for dropout resulting from these factors will be at the beginning of the Web experiment. The checklist in Table 2 shows factors to be bundled and measures to be taken in the high-hurdle technique (also see Reips, 2000). A second precaution that can be taken to reduce dropout is asking for the degree of seriousness of a participant s involvement (Musch & Klauer, 2002) or for a probability estimate that one will complete the whole experiment. If it is pre-determined that only data sets from persons with a motivation for serious participation will be analyzed, then the resulting dropout rate is usually much lower. The warm-up technique is based on the observation that most dropout will take place at the beginning of an online study, forming a natural dropout curve (Reips, 2002b). A main reason for the initial dropout is the short orientation period many participants show before making a final decision on their participation (orientation often takes place even after clicking on a submit button confirming informed consent, because informed consent forms tend to be filled with abstract legalese while the study materi- als are concrete 6 ). To keep dropout low during the experimental phase, as defined by the occurrence of the experimental manipulation, it is wise to place its beginning several Web pages deep into the study. The warm-up period can be used for practice trials, piloting of similar materials or buildup of behavioral routines, and assurance that participants are complying with instructions. Figure 1 shows the resulting dropout during warm-up phase and experimental phase after implementation of the warm-up technique in the experiment by Reips et al. (2001) Warm-up phase Start Instr 1 Instr 2 Instr 3 Instr 4 Item 1 Item 12 Last Item Web page Experimental phase Figure 1. Percentage of remaining participants as a function of progress into the Web experiment by Reips et al. (2001). The implementation of the warmup technique resulted in very low dropout after introduction of the experimental manipulation (beginning with Item 1 ). 6 At central European research institutions, informed consent is often provided in the form of a brief statement that the person will become a participant in a study by going further, accompanied by some (again: brief) information about the research institution, the study and the researchers. Many researchers feel that participants are per 2002 Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
8 250 Ulf-Dietrich Reips Dropout may also be used as a dependent variable (Frick et al., 2001; Reips, 2002a), for example in usability research. Control Experimental Setting It can be seen as a major problem in Web experimenting that the experimenter has limited control of the experimental setting 7. For example, in Web-based perception experiments it is more difficult than in the laboratory to guarantee that stimuli are perceived as intended or even assess how they were presented (Krantz, 2000, 2001). While we cannot control this principal property of Web experimentation, we can certainly use technical measures to collect information about the computer-bound aspects of the setting at the participant s end of the line. Via HTTP protocol, Javascript, and Java, the following information can be accessed: (1) type and version of Web browser, (2) type and version of operating system, (3) screen width and height, (4) screen resolution, (5) color depth of monitor setting, (6) accuracy of the computer s timing response (Eichstaedt, 2001), and (7) loading times. Javascript, Java, and plug-in based procedures may inherently reduce experimental control, because these programming languages and routines often interact with other software on the client computer. They will lead to an increase in loading times of Web pages, and increase the likelihood of inaccessibility of the Web experiment. A study by Schwarz and Reips (2001) showed that an otherwise identical copy of a Web experiment had a 13 % higher dropout rate if Javascript was used to implement client-side routines for certain experimental procedures. Therefore, the cons of using Javascript and Java carefully need to be considered against the value of gathering the above-listed information. These methods can, however, produce timing capabilities unavailable to any other methods. se knowledgeable about their basic rights and need not be coerced into reading pages of legalese texts. This seems particularly true in Internet-based studies, where the end of participation is always only one mouse click away. Also, lengthy informed consent forms may have a biasing impact on the resulting data. 7 I prefer the view that reduced control of situational variables allows for wider generalizability of results, if an effect is found. Experimental Psychology 2002; Vol. 49(4): 243Ð256 Identity Multiple submissions and reduced quality of data (missing responses, etc.) create potential problems in Internet-based research, because a participant s identity can hardly be determined without doubt. Several scenarios are thinkable: Ð The same person uses the same computer (IP address) to participate repeatedly. Ð The same person uses different computers to participate repeatedly. Ð Different persons use the same computer to participate. Ð Multiple datasets are submitted from the same computer, but are assigned different IP addresses. Ð Multiple datasets are submitted from different computers, but some Web pages or page elements are assigned the same IP addresses by proxy servers. There are many indications that the rate of repeated participations (below 3 % in most studies) is not a threat to the reliability of Internet-based research (Krantz & Dalal, 2000; Musch & Reips, 2000; Reips, 1997). However, adjunctive measures should be taken and be reported. Data quality is better in Webbased studies if identifying information is asked at the study s onset (Frick et al. 2001). Many attempts at reducing the number of multiple submissions and increasing data quality concentrate on uncovering the identities of participants or computers. Also, in light of the high numbers of participants that can be recruited on the Internet certain data sets can easily be excluded following clear predetermined criteria. In order to solve the identity problem and resulting issues with data quality the techniques listed in Table 3 may be used in Web experimenting. These techniques can be grouped as techniques of avoidance and techniques of control of multiple submissions. Users will return to a Web experiment for repeated participation only when they are motivated by entertainment or other attractive features of the experiment. Multiple submissions are likely in Web experiments that are designed in game style (e.g., Ruppertsberg et al., 2001), with highly surprising or entertaining outcomes (for example, the magic experiment in the Web Experimental Psychology Lab, available at magic.htm) or with high incentives. Quality of Data Some of the techniques that were discussed earlier in this article are suited to maximize or improve data quality in Web experimentation. For example, Frick 2002 Hogrefe & Huber Publishers
9 Standards for Internet-Based Experimenting 251 Table 3. Avoidance and Control of Multiple Submissions Avoidance of Multiple Submissions Control of Multiple Submissions (1) Informing participants on the experiment s start (1) Collecting information that allows personal identipage that multiple submissions are unwanted and det- fication. rimental to the study s purpose. (2) Asking participants on the experiment s start page (2) Redirecting the second and all further requests whether they have participated before. for the experiment s start page coming from the same (3) Continuous session IDs created through dynamic IP address (this technique likely excludes a number hidden variables or Javascript, in combination with of people using Internet providers with dynamic IP analysis of HTTP information (e.g., browser type and addressing). operating system). (3) Implementation of password-dependent access. (4) Use of the sub-sampling technique (Reips, 2000, Every participant is given a personal password that 2002b): for a limited random sample from all data can only be used once (e.g., Schmidt, 2000). Of sets every measure is taken to verify the participants course, the password scheme should be not too obviverifiable responses (for example, age, sex, occupaous (e.g., do not use hgz3, hgz4, hgz5, etc. as tion) and identity, resulting in an estimate for the topasswords). tal percentage of wrong answers and multiple submis- (4) Limiting participation to members of a controlled sions. group (participant pool or online panel). (5) Limiting participation to members of a controlled group. (6) Controlling for internal consistency of answers. (7) Controlling for consistency of response times. (8) Placement of identifiers (cookies) on the hard disks of participants computers Ð however, this technique is of limited reliability and will exclude a large portion of participants. et al. (2001) found that participants who requested the Web page with questions about their age, gender, nationality, and address substantially more often completed responses when these personal data were collected at the beginning (95.8 %) rather than at the end of an experiment (88.2 %). The warm-up technique ensures that data collected in the experimental phase come from the mostly highly committed participants. Providing information about incentives for participation certainly helps, but may not be legal in some jurisdictions (for example, in Germany) if the incentive is in the form of sweepstakes and entry is dependent on complete participation. Missing responses can be reduced by making participants aware that they might have accidentally skipped the corresponding items or even by repeatedly presenting these items. Consistency checks help in detecting inappropriate or possibly erroneous answers, for example participants claiming to be 15 years old and having a Ph.D. Quality of data may also be ensured by excluding data sets that do not meet certain criteria (e.g., more than missing entries, overly long response times). In general, incomplete data sets are more frequent in Internet-based research than in offline research (unless the experimenter forces participants into a decision between complete responding or dropout). Fortunately, there is evidence that complete data sets collected online are mostly accurate: Voracek, Stieger, and Gindl (2001) were able to control responses for biological sex via university records in an online study conducted at the University of Vienna. In contradiction to the frequently heard online gender swapping myth, they found that the rate of seemingly false responses was below 3 %. Furthermore, the data pattern of the participants suspected of false responding was in accordance with their reported biological sex, leading to the conclusion that many of these cases were opposite-sex persons using a friend s or roommate s university account. Misunderstandings and Limited Ways of Interaction In laboratory experiments it is often necessary to explain some of the procedures and materials to the participants in an interactive communication. In the laboratory, experimenters are able to answer questions directly and verify through dialogue and read Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
10 252 Ulf-Dietrich Reips ing of nonverbal behavior that the instructions were understood. In Web experiments such interaction is difficult to achieve, and mostly not desired, because it would counter some of the method s advantages, for instance its low expenditure and the ability to collect data around the clock. Pretesting of the materials and collecting feedback by providing communication opportunities for participants are the most practical ways of avoiding misunderstandings. They should be used with care, in order to identify problems early on. Errors and Frequent Misconceptions in Internet- Based Experimenting Apart from preventing misunderstandings on part of participants careful pretesting and monitoring of an Internet-based experiment will help in detecting the presence of any of a number of frequently made configuration errors and possibly of misconceptions on part of the experimenter. Configuration Errors and Technical Limitations Configuration Errors There are five frequently observed configuration errors in Internet-based experimenting (Reips, 2002a) that can easily be avoided, if awareness of their presence is raised and precautions are taken. These errors are: Ð Allowing external access to unprotected directories (configuration error I). In most cases, placing a file named index.htm or index.html into each directory will solve the problem. Ð Public display of confidential participant data through URL Ð data may be written to a third party Web server (configuration error II). To avoid this error, do not collect data with the GET method 8 on any Web page that is two nodes away from a Web page with links to external sources on the last page. Ð File and folder names and/or field names reveal the experiment s structure (configuration error 8 The GET method is a request method used in the WWW transmission protocol. The two most often used methods for transmitting form data are GET and POST. Information from a form using the GET method is appended onto the end of the action address being requested; for example, in yes, the answer yes in an item webexp was appended to the URL of the Web page that a user s action (pressing a submit button, etc.) leads to. Experimental Psychology 2002; Vol. 49(4): 243Ð256 III). Use a mixture of logical and random letters and numbers to avoid this error. Ð Ignorance towards the technical variance present in the Internet (configuration error IV). Differences in Web browsers, versions of Web browsers, net connections, hardware components etc. need to be considered in the implementation. Ð Biased results from improper use of form elements. For example, if no neutral answer (e.g., Choose here ) is preselected (configuration error V), there may be errors of omission. For example, one might erroneously believe that many seniors participated in an experiment until one realizes that the preset answer in the pull-down menu for age was 69 or older... Biasing may also result from order effects or type of form element used (i.e., pull-down menus, select multiples in lists, etc.). Technical Limitations There are several technical and methodological sources of error and bias that need to be considered. Security holes in operating systems and Web server applications (see Reips, 2002a; Securityfocus.com, n.d.) may allow viruses to shut down the experiment or corrupt data or even let intruders download confidential participant information. While there are significant differences in vulnerability of operating systems, as monitored permanently by Securityfocus.- com, any operating system with a Web server collecting data from human participants over the Internet needs to be maintained with the highest degree of responsibility. Another issue of a researcher s responsibility is a frequently observed breach of the basic principle of record taking in research, if only aggregated data are collected. Raw logfiles with full information need to be stored for reanalysis by other researchers and for meta-analyses. Records on incomplete data sets (dropout figures) should be collected as well! As mentioned, sometimes dynamic IP addressing makes it more difficult to identify coherent data sets. In any event, drawing conclusions about real behavior from Internet data is sometimes difficult, because log data may be ambiguous. For instance, long response times can be caused by the participant s behavior and also by a slow computer or a slow network connection. High dropout rates caused by technical problems (e.g., when using Javascript, Schwarz & Reips, 2001) may in some cases bias results. As mentioned, personality types correlate with the use of certain technologies (Buchanan & Reips, 2001), leading to biases in cases of systematic technical incompatibilities. Finally, due to the public accessibility of many Web experiments, researchers need to be aware of 2002 Hogrefe & Huber Publishers
11 Standards for Internet-Based Experimenting 253 one more likely source of bias: participation of experts (colleagues). In an Internet-based experiment on a causal learning phenomenon in cognition (Reips, 1997) 15 % of participants indicated on an insider control item that they were working in or studying cognitive psychology. Providing participants with such an item that allows them to look at the experiment without having to fear potential corruption of a colleague s data avoids this bias. Misconceptions There are two misconceptions about Internet-based experimenting that carry dangerous implications that might lead to blind eyes against errors. Misconception 1: Experimenting on the Web is Just Like Experimenting in the Lab The Internet is not a laboratory. Internet participants may end their commitment at any time that they no longer have interest. Even though lab participants can terminate participation at any time as well, they are less likely to given that they are required to face another human and encounter a potentially embarrassing situation when doing so (e.g., Bamert, 2002; Reips, 1997, 2000). Web experiments are usually online around the clock, and usually they reach a much larger number of people than laboratory experiments. Therefore, overlooking even small factors may have wide consequences. Web experiments are dependent on the quality of networks and are subject to great variance in local settings of participants. Last but not least, experimenting on the Web is much more public than working in the laboratory. In addition to a wider public, one s colleagues will be able to inform themselves in a much more direct way about one s work. Misconception 2: Experimenting on the Web is Completely Different from Experimenting in the Lab Even though Web experiments are always dependent on networks and computers, most laboratory experiments are conducted on computers. Often a standardized user interface is used to display the experimental materials on the screen. Web browsers are highly standardized user interfaces. Experimental materials made with the help of tools like WEXTOR (Reips & Neuhaus, 2002) can be used in both laboratory and Internet-based experiments. Creating the materials is quite easy with basic knowl- edge in experimental design and handling of HTML. Fundamental ideas and methodological procedures are the same in Web and physical lab, and similar results have been produced in studies conducted in both settings (for a summary see Krantz and Dalal, 2000). As mentioned earlier, the combination of laboratory and Internet-based experimenting in distributed Web experimenting shows that Web and lab can be integrated in creative ways to arrive at new variations of the experimental method. Dynamics of Data Collection Internet-based experimenting creates its own dynamics. Once an experiment is online and linked to a variety of Web sites it will be present on the Internet for a long time, even if the site itself is removed (which it shouldn t be Ð it should be replaced by information about the experiment instead). Some search engines will cache the experiment s contents, and so do some proxy servers, even if anticaching meta tags are used in the Web pages. Internet-based laboratories often have large numbers of visitors and are linked extensively (Reips, 2001). As a consequence, they may even create pressure to offer new Web experiments within short periods of time to satisfy the crowd s desires. Meeting these desires creates the risk of producing superfluous data Ð an issue that is in need of being discussed by the community of Internet scientists. The flood of data bears the danger of losing the sense for best care and attention towards data and participants. Would this loss of reasonable diligence be a simple more is worth less phenomenon that could result in long-term attitude changes in researchers? Or would it reflect an interaction of the realm of possibilities of techno-media power with limited education in Internet-based experimenting? In any case, the scientific process in psychology is changing profoundly. Summary: Sixteen Standards for Internet-Based Experimenting In this article, a number of important issues in Internet-based experimenting were discussed. As a consequence, several routines and standards for Internet-based experimenting were proposed. When reporting Web experiments, the implementation of and specifics about the mentioned techniques should be included. The following list of recommendations summarizes most of what needs to be remembered and may be used as a standards check Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
12 254 Ulf-Dietrich Reips list when conducting an Internet-based experiment and reporting its results. Standard 1: Consider using a Web-based software tool to create your experimental materials. Such tools automatically implement standard procedures for Web experiments that can guard against many problems. Examples are WEXTOR and FactorWiz (see footnotes 1 and 2 for URLs). If you use FactorWiz, make sure to protect your participants data by changing the default procedure of storing them in a publicly accessible data file and be aware that FactorWiz creates only one-page Web experiments, so you will not be able to measure dropout in a meaningful way. Standard 2: Pretest your experiment for clarity of instructions and availability on different platforms. Standard 3: Make a decision whether the advantages of non-html scripting languages and plug-ins outweigh their disadvantages. Standard 4: Check your Web experiment for configuration errors (I-V; Reips, 2002a). Standard 5: Consider linking your Web experiment to several Internet sites and services (multiple site entry technique) to determine effects of self-selection and estimate generalizability. Standard 6: Run your experiment both online and offline, for comparison. Standard 7: If dropout is to be avoided, use the warm-up technique. Standard 8: Use dropout to determine whether there is motivational confounding. Standard 9: Use the high-hurdle technique, incentive information, and requests for personal information to influence time and degree of dropout. Standard 10: Ask filter questions (seriousness of participation, expert status, language skills, etc.) at the beginning of the experiment to encourage serious and complete responses. Standard 11: Check for obvious naming of files, conditions, and, if applicable, passwords. Standard 12: Consider avoiding multiple submissions by exclusively using participant pools and password techniques. Standard 13: Perform consistency checks. Standard 14: Keep experiment log and other data files for later analyses by members from the scientific community. Standard 15: Report and analyze dropout curves or at least dropout rates for experimental conditions separately for between-subjects factors. Experimental Psychology 2002; Vol. 49(4): 243Ð256 Standard 16: The experimental materials should be kept available on the Internet, as they will often give a much better impression of what was done than any verbal description could convey. Conclusion Internet-based experimenting is fast becoming a standard method and therefore it is a method that needs standards. Many established experimentalists as well as students are currently making their first attempts in using the new method. So far, in many universities there is no curriculum that teaches Internet-based experimenting. Still only few people have both the technical and the methodological experience to give advice to those who would like to commence with the venture of conducting Web experiments. Without established standards the likelihood is high for making grave errors that would result in loss or reduced quality of data, in biased results, or in breach of ethical practices. Consequently, in the present paper an attempt was made to collect and discuss what has been learned in Internet-based experimenting in order to make recommendations, warn about errors, and introduce useful techniques. As a result, a set of standards for Internet-based experimenting could be defined that hopefully will serve as a guide for future experiments on the Internet. Those who have done Web experiments keep conducting them. Many of those who haven t will do soon. And many of those who would never conduct an experiment on the Internet will be confronted with the methodology as reviewers and readers of publications or as teachers of students who ask for convenient and proper ways of collecting experimental data in that international communication network that allow us to investigate the psychology of the many distant human beings out there. Hopefully, with the guidance of standards and examples from the present special issue of Experimental Psychology, Internet-based experimenting will come one step closer to being established and used as an equivalent tool for scientists. References Bamert, T. (2002). Integration von Wahrscheinlichkeiten: Verarbeitung von zwei Wahrscheinlichkeitsinformationen [Integration of probabilities: Processing two pieces of probability information]. Unpublished master s thesis, University of Zurich, Switzerland. Bargh, J. A., McKenna, K. Y. A., & Fitzsimons, G. M. (2002). Can you see the real me? Activation and expression of the true self on the Internet. Journal of Social Issues, 58, 33Ð Hogrefe & Huber Publishers
13 Standards for Internet-Based Experimenting 255 Birnbaum, M. H. (2000). SurveyWiz and FactorWiz: JavaScript Web pages that make HTML forms for research on the Internet. Behavior Research Methods, Instruments, and Computers, 32, 339Ð346. Birnbaum, M. H. (2001). A Web-based program of research on decision making. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 23Ð 55). Lengerich, Germany: Pabst Science. Bohner, G., Danner, U. N., Siebler, F., & Samson, G. B. (2002, this volume). Rape myth acceptance and judgments of vulnerability to sexual assault: An Internet experiment. Experimental Psychology, 49 (4). Bosnjak, M. (2001). Participation in non-restricted Web surveys: A typology and explanatory model for item non-response. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 193Ð208). Lengerich, Germany: Pabst Science. Buchanan, T. (2002). Online assessment: Desirable or dangerous? Professional Psychology: Research and Practice, 33, 148Ð154. Buchanan, T., & Reips, U.-D. (2001, October 10). Platform-dependent biases in Online Research: Do Mac users really think different? In K. J. Jonas, P. Breuer, B. Schauenburg, & M. Boos (Eds.), Perspectives on Internet Research: Concepts and Methods. Retrieved December 27, 2001, from Chapanis, A. (1970). The relevance of laboratory studies to practical situations. In D. P. Schultz (Ed.), The science of psychology: Critical reflections. New York: Appleton Century Crofts. Coomber, R. (1997, June 30). Using the Internet for survey research. Sociological Research Online, 2, Retrieved June 16 th, 2002, from 2/2/2.html Eichstaedt, J. (2001). Reaction time measurement by JAVA-applets implementing Internet-based experiments. Behavior Research Methods, Instruments, & Computers, 33, 179Ð186. Eichstaedt, J. (2002). Measuring differences in preactivation on the Internet: The content category superiority effect. Experimental Psychology, 49 (4). Frick, A., Bächtiger, M. T., & Reips, U.-D (2001). Financial incentives, personal information, and dropout in online studies. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 209Ð219). Lengerich, Germany: Pabst Science. Hewson, M., Laurent, D., & Vogel, C. M. (1996). Proper methodologies for psychological and sociological studies conducted via the Internet. Behavior Research Methods, Instruments, and Computers, 28, 186Ð191. Hiskey, S., & Troop, N. A. (2002). Online longitudinal survey research: Viability and participation. Social Science Computer Review, 20 (3) 250Ð259. Horswill, M. S., & Coster, M. E. (2001). User-controlled photographic animations, photograph-based questions, and questionnaires: Three instruments for measuring drivers risk-taking behavior on the Internet. Behavior Research Methods, Instruments, and Computers, 33, 46Ð58. Joinson, A. (2001). Self-disclosure in computer-mediated communication: The role of self-awareness and visual anonymity. European Journal of Social Psychology, 31, 177Ð192. Kiesler, S., & Sproull, L. S. (1986). Response effects in the electronic survey. Public Opinion Quarterly, 50, 402Ð413. Klauer, K. C., Musch, J., & Naumer, B. (2000). On belief bias in syllogistic reasoning. Psychological Review, 107, 852Ð884. Krantz, J. H. (2000). Tell me, what did you see? The stimulus on computers. Behavior Research Methods, Instruments, and Computers, 32, 221Ð229. Krantz, J. H. (2001). Stimulus delivery on the Web: What can be presented when calibration isn t possible. In Reips, U.-D. & Bosnjak, M. (Eds.), Dimensions of Internet Science (pp. 113Ð130). Lengerich, Germany: Pabst Science. Krantz, J. H., & Dalal, R. S. (2000). Validity of Web-based psychological research. In M. H. Birnbaum (Ed.), Psychological experiments on the Internet (pp. 35Ð60). San Diego, CA: Academic Press. Laugwitz, B. (2001). A Web experiment on color harmony principles applied to computer user interface design. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 131Ð145). Lengerich, Germany: Pabst Science. Musch, J., & Klauer, K. C. (2002). Psychological experimenting on the World Wide Web: Investigating content effects in syllogistic reasoning. In B. Batinic, U.-D. Reips, & M. Bosnjak (Eds.), Online Social Sciences (pp. 181Ð212). Göttingen, Germany: Hogrefe. Musch, J., & Reips, U.-D. (2000). A brief history of Web experimenting. In M. H. Birnbaum (Ed.), Psychological experiments on the Internet (pp. 61Ð88). San Diego, CA: Academic Press. O Neil, K. M., & Penrod, S. D. (2001). Methodological variables in Web-based research that may affect results: Sample type, monetary incentives, and personal information. Behavior Research Methods, Instruments, and Computers, 33, 226Ð233. Orne, M. T. (1962). On the social psychology of the psychological experiment: With particular reference to demand characteristics and their implications. American Psychologist, 17, 776Ð783. Pohl, R. F., Bender, M., & Lachmann, G. (2002, this volume). Hindsight bias around the world. Experimental Psychology, 49 (4). Postmes, T., Spears, R., Sakhel, K., & DeGroot, D. (2001). Social influence in computer-mediated communication: The effect of anonymity on group behavior. Personality and Social Psychology Bulletin, 27, 1243Ð1254. Reips, U.-D (1995). The Web experiment method. Retrieved January 6, 2002, from Reips, U.-D. (1997). Das psychologische Experimentieren im Internet [Psychological experimenting on the Internet]. In B. Batinic (Ed.), Internet für Psychologen (pp. 245Ð265). Göttingen, Germany: Hogrefe. Reips, U.-D. (1999). Online research with children. In U.- D. Reips, B. Batinic, W. Bandilla, M. Bosnjak, L. Gräf, K. Moser, & A. Werner (Eds.). Current Internet science Ð trends, techniques, results. Aktuelle Online- Forschung Ð Trends, Techniken, Ergebnisse. Zürich: Online Press. Retrieved April 7, 2002 from dgof.de/tband99/ Reips, U.-D. (2000). The Web experiment method: Advantages, disadvantages, and solutions. In M. H. Birnbaum (Ed.), Psychological experiments on the Internet (pp. 89Ð114). San Diego, CA: Academic Press. Reips, U.-D. (2001). The Web Experimental Psychology Lab: Five years of data collection on the Internet. Behavior Research Methods, Instruments, and Computers, 33, 201Ð Hogrefe & Huber Publishers Experimental Psychology 2002; Vol. 49(4): 243Ð256
14 256 Ulf-Dietrich Reips Reips, U.-D. (2002a). Internet-based psychological experimenting: Five dos and five don ts. Social Science Computer Review, 20 (3), 241Ð249. Reips, U.-D. (2002b). Theory and techniques of conducting Web experiments. In B. Batinic, U.-D. Reips, & M. Bosnjak (Eds.), Online Social Sciences (pp. 229Ð250). Seattle: Hogrefe & Huber. Reips, U.-D., & Bosnjak, M. (2001). Dimensions of Internet Science. Lengerich, Germany: Pabst Science. Reips, U.-D., Morger, V., & Meier B. (2001). Fünfe gerade sein lassen : Listenkontexteffekte beim Kategorisieren [ Letting five be equal : List context effects in categorization]. Unpublished manuscript. Retrieved April 7, 2002 from genpsy/reips/papers/re_mo_me2001.pdf Reips, U.-D., & Neuhaus, C. (2002). WEXTOR: A Webbased tool for generating and visualizing experimental designs and procedures. Behavior Research Methods, Instruments, and Computers, 34, 234Ð240. Rodgers, J., Buchanan, T., Scholey, A. B., Heffernan, T. M., Ling, J., & Parrott, A. (2001). Differential effects of Ecstasy and cannabis on self-reports of memory ability: A web-based study. Human Psychopharmacology: Clinical and Experimental, 16, 619Ð625. Rosenthal, R. (1966). Experimenter effects in behavioral research. New York: Appleton-Century-Crofts. Rosenthal, R., & Fode, K. L. (1973). The effect of experimenter bias on the performance of the albino rat. Behavioral Science, 8, 183Ð189. Rosenthal, R., & Rosnow, R. L. (1969). Artifact in behavioral research. New York: Academic Press. Ruppertsberg, A. I., Givaty, G., Van Veen, H. A. H. C., & Bülthoff, H. (2001). Games as research tools for visual perception over the Internet. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 147Ð158). Lengerich, Germany: Pabst. Schmidt, W. C. (1997). World-Wide Web survey research: Benefits, potential problems, and solutions. Behavior Research Methods, Instruments, & Computers, 29, 274Ð279. Schmidt, W. C. (2000). The server-side of psychology Web experiments. In M. H. Birnbaum (Ed.), Psychological experiments on the Internet (pp. 285Ð310). San Diego, CA: Academic Press. Smart, R. (1966). Subject selection bias in psychological research. Canadian Psychologist, 7a, 115Ð121. Schwarz, S., & Reips, U.-D. (2001). CGI versus JavaScript: A Web experiment on the reversed hindsight bias. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 75Ð90). Lengerich, Germany: Pabst Science. Securityfocus.com. (n.d.). BUGTRAQ Vulnerability Database Statistics. Retrieved April 7, 2002, from Voracek, M., Stieger, S., & Gindl, A. (2001). Online replication of Evolutionary Psychology evidence: Sex differences in sexual jealousy in imagined scenarios of mate s sexual versus emotional infidelity. In U.-D. Reips & M. Bosnjak (Eds.), Dimensions of Internet Science (pp. 91Ð112). Lengerich, Germany: Pabst Science. Wenzel, O. (2001). Webdesign, Informationssuche und Flow: Nutzerverhalten auf unterschiedlich strukturierten Websites [Web design, search for information, and flow: User behavior on differently structured Web sites]. Lohmar, Germany: Eul. Ulf-Dietrich Reips Experimental and Developmental Psychology University of Zürich Attenhoferstr. 9 CH-8032 Zürich Switzerland Tel.: Fax: [email protected] Experimental Psychology 2002; Vol. 49(4): 243Ð Hogrefe & Huber Publishers
Financial incentives, personal information and dropout rate in online studies
Financial incentives, personal information and dropout rate in online studies A. Frick 1, M. T. Bächtiger & U.-D. Reips Whereas in a classical laboratory setting participants often feel compelled to stay
Internet Surveys. Examples
Web material accompanying The International Handbook of Survey Methodology Chapter 14 Internet Surveys Katja Lozar Manfreda Vasja Vehovar University of Ljubljana Examples Below follow examples and more
N-CAP Users Guide Everything You Need to Know About Using the Internet! How Firewalls Work
N-CAP Users Guide Everything You Need to Know About Using the Internet! How Firewalls Work How Firewalls Work By: Jeff Tyson If you have been using the internet for any length of time, and especially if
PLATO Learning Environment System and Configuration Requirements for workstations. October 27th, 2008
PLATO Learning Environment System and Configuration Requirements for workstations October 27th, 2008 Windows 2000 Professional with SP4 Windows XP Professional with SP2 Windows XP Home Edition with SP2
Recommended Practice Case Study: Cross-Site Scripting. February 2007
Recommended Practice Case Study: Cross-Site Scripting February 2007 iii ACKNOWLEDGEMENT This document was developed for the U.S. Department of Homeland Security to provide guidance for control system cyber
How to Monitor and Identify Website Issues. 2013, SolarWinds Worldwide, LLC. All rights reserved. Share:
How to Monitor and Identify Website Issues 2013, SolarWinds Worldwide, LLC. All rights reserved. Share: What Affects a Website s Performance Website performance issues can come from anywhere the page elements
The following multiple-choice post-course assessment will evaluate your knowledge of the skills and concepts taught in Internet Business Associate.
Course Assessment Answers-1 Course Assessment The following multiple-choice post-course assessment will evaluate your knowledge of the skills and concepts taught in Internet Business Associate. 1. A person
Protected PDF Common Installation Issues
Protected PDF Common Installation Issues July 14, 2011 Version 1.1 TABLE OF CONTENTS Basic Instructions for Readers ------------------------------------------------------------------------- 2 Common Questions
Chapter-1 : Introduction 1 CHAPTER - 1. Introduction
Chapter-1 : Introduction 1 CHAPTER - 1 Introduction This thesis presents design of a new Model of the Meta-Search Engine for getting optimized search results. The focus is on new dimension of internet
Paper 10-27 Designing Web Applications: Lessons from SAS User Interface Analysts Todd Barlow, SAS Institute Inc., Cary, NC
Paper 10-27 Designing Web Applications: Lessons from SAS User Interface Analysts Todd Barlow, SAS Institute Inc., Cary, NC ABSTRACT Web application user interfaces combine aspects of non-web GUI design
PLATO Learning Environment 2.0 System and Configuration Requirements. Dec 1, 2009
PLATO Learning Environment 2.0 System and Configuration Requirements Dec 1, 2009 Table of Contents About this document... 3 Document Change Log... 4 System & Configuration Requirements... 5 Workstation
WHEN YOU CONSULT A STATISTICIAN... WHAT TO EXPECT
WHEN YOU CONSULT A STATISTICIAN... WHAT TO EXPECT SECTION ON STATISTICAL CONSULTING AMERICAN STATISTICAL ASSOCIATION 2003 When you consult a statistician, you enlist the help of a professional who is particularly
ONLINE EXTERNAL AND SURVEY STUDIES
ONLINE EXTERNAL AND SURVEY STUDIES Before reading this document, be sure you are already familiar with the Instructions for using the School of Psychological Sciences Participant Pool available on the
PRIVACY POLICY. I. Introduction. II. Information We Collect
PRIVACY POLICY school2life, Inc. ( school2life ) Privacy Policy is designed to provide clarity about the information we collect and how we use it to provide a better social gaming experience. By accepting
Privacy Policy. Effective Date: November 20, 2014
Privacy Policy Effective Date: November 20, 2014 Welcome to the American Born Moonshine website (this Site ). This policy describes the Privacy Policy (this Policy ) for this Site and describes how Windy
THE FUTURE OF INTERNET-BASED SURVEY METHODS
Chapter Seven CONCLUSIONS In this chapter, we offer some concluding thoughts on the future of Internet-based surveys, the issues surrounding the use of e-mail and the Web for research surveys, and certain
The Sales Lead System
The Sales Lead System Contents Intro... 3 Network Marketing Ad Basics... 3 Blogging... 4 Article Marketing... 4 Video Marketing... 5 E-mail Marketing... 6 Pay Per Click... 7 Banner Ads... 8 Social Marketing...
126.47. Web Design (One Credit), Beginning with School Year 2012-2013.
126.47. Web Design (One Credit), Beginning with School Year 2012-2013. (a) General requirements. Students shall be awarded one credit for successful completion of this course. This course is recommended
Using the Internet to Collect Data
C h a p t e r 17 Using the Internet to Collect Data Ulf-Dietrich Reips The Internet is not one single, monolithic medium. It consists of many services with different functions and needs for input, even
SONA SYSTEMS RESEARCHER DOCUMENTATION
SONA SYSTEMS RESEARCHER DOCUMENTATION Introduction Sona Systems is used for the scheduling and management of research participants and the studies they participate in. Participants, researchers, principal
Comparative Analysis Report:
Comparative Analysis Report: Visualization Tools & Platforms By Annabel Weiner, Erol Basusta, Leah Wilkinson, and Quenton Oakes Table of Contents Executive Summary Introduction Assessment Criteria Publishability
Using audio-conferencing for language learning tasks
Using audio-conferencing for language learning tasks Audio-conferencing tools enable users to talk to (and sometimes see) one another in real time over the Internet. In this activity we will see how you
Web Design Competition 2013. College of Computing Science, Department of Information Systems. New Jersey Institute of Technology
COMPETITION PURPOSE The Web is the most transformable invention of our time. This competition features the creation of high-quality, well-designed and original Websites, while seeking to identify and encourage
What is Web Security? Motivation
[email protected] http://www.brucker.ch/ Information Security ETH Zürich Zürich, Switzerland Information Security Fundamentals March 23, 2004 The End Users View The Server Providers View What is Web
PLATO Learning Environment System and Configuration Requirements. for workstations. April 14, 2008
PLATO Learning Environment System and Configuration Requirements Version 1.1 (for use with Academic Systems Algebra only) for workstations April 14, 2008 Windows 2000 Professional with SP4 Windows XP Professional
How To Test A Web Based System
Testing Web-Based Systems-Checklists Testing Web-Based Systems -Checklist Overview-: Web-based testing should be RISK ORIENTED. This article describes the risks, presents the types of testing that can
Search Engine Optimization Glossary
Search Engine Optimization Glossary A ALT Text/Tag or Attribute: A description of an image in your site's HTML. Unlike humans, search engines read only the ALT text of images, not the images themselves.
CHAPTER 20 TESING WEB APPLICATIONS. Overview
CHAPTER 20 TESING WEB APPLICATIONS Overview The chapter describes the Web testing. Web testing is a collection of activities whose purpose is to uncover errors in WebApp content, function, usability, navigability,
You're reading an excerpt. Click here to read official APPLE REMOTE DESKTOP 1.2 user guide http://yourpdfguides.com/dref/1168427
You can read the recommendations in the user guide, the technical guide or the installation guide for APPLE REMOTE DESKTOP 1.2. You'll find the answers to all your questions on the APPLE REMOTE DESKTOP
Miami University: Human Subjects Research General Research Application Guidance
Miami University: Human Subjects Research General Research Application Guidance Use the accompanying Word template for completing the research description. You must provide sufficient information regarding
Device Fingerprinting and Fraud Protection Whitepaper
Device Fingerprinting and Fraud Protection Whitepaper 1 of 6 Table Of Contents 1 Overview... 3 2 What is Device Fingerprinting?... 3 3 Why is Device fingerprinting necessary?... 3 4 How can Device Fingerprinting
GUIDE TO WEBSITES AND E-COMMERCE
GUIDE TO WEBSITES AND E-COMMERCE Version 1.0, 26-Sept-01 This document is available from www.webcentro.com.au 2001, WebCentro WebCentro Guide To Websites And E-commerce CONTENTS 1. What is a Website? 1
Sona Systems, Ltd. EXPERIMENT MANAGEMENT SYSTEM Master Documentation Set
Sona Systems, Ltd. EXPERIMENT MANAGEMENT SYSTEM Master Documentation Set Version 2.65 Copyright 2006 Sona Systems, Ltd., All Rights Reserved About This Manual This manual covers usage of the system from
CREATING YOUR ONLINE PRESENCE
CREATING YOUR ONLINE PRESENCE Congratulations on signing up for your webhosting package, you ve just completed the first and most important step in establishing your online presence. There are just a few
Distance-Learning Remote Laboratories using LabVIEW
Distance-Learning Remote Laboratories using LabVIEW Introduction Laboratories, which are found in all engineering and science programs, are an essential part of the education experience. Not only do laboratories
Parallels Plesk Automation. Customer s Guide. Parallels Plesk Automation 11.5
Parallels Plesk Automation Customer s Guide Parallels Plesk Automation 11.5 Last updated: 17 March 2015 Contents Quick Start with Hosting Panel 4 Set Up Your First Website... 4 1. Create Your Site... 5
INTRODUCTION We respect your privacy and are committed to protecting it through our compliance with this privacy policy.
Privacy Policy Last Modified: July 20, 2015 INTRODUCTION We respect your privacy and are committed to protecting it through our compliance with this privacy policy. This privacy policy describes the types
Configuring, Customizing, and Troubleshooting Outlook Express
3 Configuring, Customizing, and Troubleshooting Outlook Express............................................... Terms you ll need to understand: Outlook Express Newsgroups Address book Email Preview pane
Remote PC Guide for Standalone PC Implementation
Remote PC Guide for Standalone PC Implementation Updated: 2007-01-22 The guide covers features available in NETLAB+ version 3.6.1 and later. IMPORTANT Standalone PC implementation is no longer recommended.
Learning Activity Management System TEACHER S GUIDE. October 2006 LAMS V2.0 for The First International LAMS Conference 2006
Learning Activity Management System TEACHER S GUIDE October 2006 LAMS V2.0 for The First International LAMS Conference 2006 James Dalziel (amended by Bronwen Dalziel) LAMS Teacher s Guide, October 2006,
Responsive Web Design. vs. Mobile Web App: What s Best for Your Enterprise? A WhitePaper by RapidValue Solutions
Responsive Web Design vs. Mobile Web App: What s Best for Your Enterprise? A WhitePaper by RapidValue Solutions The New Design Trend: Build a Website; Enable Self-optimization Across All Mobile De vices
DEPLOYMENT GUIDE. Deploying F5 for High Availability and Scalability of Microsoft Dynamics 4.0
DEPLOYMENT GUIDE Deploying F5 for High Availability and Scalability of Microsoft Dynamics 4.0 Introducing the F5 and Microsoft Dynamics CRM configuration Microsoft Dynamics CRM is a full customer relationship
Comparison of Moodle and ATutor LMSs
Comparison of Moodle and ATutor LMSs Péter Lengyel - Miklós Herdon - Róbert Szilágyi University of Debrecen CAS-FAERD Contents Introduction (Moodle, ATutor) Evaluation aspects Technical flexibility Learning
FAQs for Oracle iplanet Proxy Server 4.0
FAQs for Oracle iplanet Proxy Server 4.0 Get answers to the questions most frequently asked about Oracle iplanet Proxy Server Q: What is Oracle iplanet Proxy Server (Java System Web Proxy Server)? A: Oracle
Appendix B Checklist for the Empirical Cycle
Appendix B Checklist for the Empirical Cycle This checklist can be used to design your research, write a report about it (internal report, published paper, or thesis), and read a research report written
4 Understanding. Web Applications IN THIS CHAPTER. 4.1 Understand Web page development. 4.2 Understand Microsoft ASP.NET Web application development
4 Understanding Web Applications IN THIS CHAPTER 4.1 Understand Web page development 4.2 Understand Microsoft ASP.NET Web application development 4.3 Understand Web hosting 4.4 Understand Web services
The web server administrator needs to set certain properties to insure that logging is activated.
Access Logs As before, we are going to use the Microsoft Virtual Labs for this exercise. Go to http://technet.microsoft.com/en-us/bb467605.aspx, then under Server Technologies click on Internet Information
SRCSB General Web Development Policy Guidelines Jun. 2010
This document outlines the conventions that must be followed when composing and publishing HTML documents on the Santa Rosa District Schools World Wide Web server. In most cases, these conventions also
VPAT Summary. VPAT Details. Section 1194.22 Web-based Internet information and applications - Detail
Date: October 8, 2014 Name of Product: System x3755 M3 VPAT Summary Criteria Status Remarks and Explanations Section 1194.21 Software Applications and Operating Systems Section 1194.22 Web-based Internet
Survey research. Contents. Chaiwoo Lee. Key characteristics of survey research Designing your questionnaire
Survey research Chaiwoo Lee Postdoctoral Associate MIT AgeLab [email protected] agelab.mit.edu Contents Key characteristics of survey research Designing your questionnaire Writing the questions Putting them
PARCC Technology Troubleshooting FAQs, Version 1.0
PARCC Technology Troubleshooting FAQs, Version 1.0 Released 3/6/14 This document contains Frequently Asked Questions to help districts and schools troubleshoot issues regarding technology set up for the
Clinical Data Management (Process and practical guide) Nguyen Thi My Huong, MD. PhD WHO/RHR/SIS
Clinical Data Management (Process and practical guide) Nguyen Thi My Huong, MD. PhD WHO/RHR/SIS Training Course in Sexual and Reproductive Health Research Geneva 2013 OUTLINE Overview of Clinical Data
PSYCHOLOGY AND PROJECT MANAGEMENT
PSYCHOLOGY AND PROJECT MANAGEMENT Human aspects of the project management Ver 1.0 2 360 Degree Approach of IT Infrastructure Projects PSYCHOLOGY AND PROJECT MANAGEMENT Abstract Project Management usually
Sona Systems, Ltd. EXPERIMENT MANAGEMENT SYSTEM Master Documentation Set
Sona Systems, Ltd. EXPERIMENT MANAGEMENT SYSTEM Master Documentation Set Version 2.74 Copyright 2010 Sona Systems, Ltd., All Rights Reserved About This Manual This manual covers usage of the system from
Digital media glossary
A Ad banner A graphic message or other media used as an advertisement. Ad impression An ad which is served to a user s browser. Ad impression ratio Click-throughs divided by ad impressions. B Banner A
SVSU Websites Style Guide. Need help? Call the ITD Lab, x7471
SVSU Websites Style Guide SVSU Websites Style Guide Need help? Call the ITD Lab, x7471 1 SVSU Websites Style Guide Contents Writing for the Web...3 Questions to Ask Yourself...3 Who is our priority audience?...3
LiveTalk Call Center solution
LiveTalk Call Center solution I. Introduction LiveTalk enables real-time interaction between callers and a pool of technical and customer support or sales agents via a completely web based interface. With
Check list for web developers
Check list for web developers Requirement Yes No Remarks 1. Input Validation 1.1) Have you done input validation for all the user inputs using white listing and/or sanitization? 1.2) Does the input validation
Data Driven Success. Comparing Log Analytics Tools: Flowerfire s Sawmill vs. Google Analytics (GA)
Data Driven Success Comparing Log Analytics Tools: Flowerfire s Sawmill vs. Google Analytics (GA) In business, data is everything. Regardless of the products or services you sell or the systems you support,
How to Use Motion Detection in ACTi Cameras
ACTi Knowledge Base Category: Installation & Configuration Note Sub-category: Application Model: All Firmware: N/A Software: N/A Author: Ando.Meritee Published: 2010/11/19 Reviewed: 2011/03/02 How to Use
Privacy Policy. If you have questions or complaints regarding our Privacy Policy or practices, please see Contact Us. Introduction
Privacy Policy This Privacy Policy will be effective from September 1 st, 2014. Please read Pelican Technologies Privacy Policy before using Pelican Technologies services because it will tell you how we
Regain Your Privacy on the Internet
Regain Your Privacy on the Internet by Boris Loza, PhD, CISSP from SafePatrol Solutions Inc. You'd probably be surprised if you knew what information about yourself is available on the Internet! Do you
Best Practices for Protecting Individual Privacy in Conducting Survey Research
Best Practices for Protecting Individual Privacy in Conducting Survey Research CONTENTS Foreword... 1 Introduction... 2 Privacy Considerations at Each Stage of a Survey Research Project... 5 Stage 1: Issue
How Web Browsers Work
144 PART 4 HOW THE WORLD WIDE WEB WORKS CHAPTER 18 How Web Browsers Work 145 LIKE much of the Internet, the World Wide Web operates on a client/server model. You run a web client on your computer called
If you have any questions about our privacy practices, please refer to the end of this privacy policy for information on how to contact us.
c4m Privacy Policy Last Modified: July 20, 2015 Colbette II Ltd., Block 1, 195-197 Old Nicosia-Limassol Road, Dali Industrial Zone, Cyprus 2540 (hereinafter "c4m", Colbette we", "our" or "us") is always
DEPLOYMENT GUIDE Version 1.2. Deploying the BIG-IP System v10 with Microsoft IIS 7.0 and 7.5
DEPLOYMENT GUIDE Version 1.2 Deploying the BIG-IP System v10 with Microsoft IIS 7.0 and 7.5 Table of Contents Table of Contents Deploying the BIG-IP system v10 with Microsoft IIS Prerequisites and configuration
DEPLOYMENT GUIDE Version 2.1. Deploying F5 with Microsoft SharePoint 2010
DEPLOYMENT GUIDE Version 2.1 Deploying F5 with Microsoft SharePoint 2010 Table of Contents Table of Contents Introducing the F5 Deployment Guide for Microsoft SharePoint 2010 Prerequisites and configuration
Version 1.0 January 2011. Xerox Phaser 3635MFP Extensible Interface Platform
Version 1.0 January 2011 Xerox Phaser 3635MFP 2011 Xerox Corporation. XEROX and XEROX and Design are trademarks of Xerox Corporation in the United States and/or other countries. Changes are periodically
CMPT 471 Networking II
CMPT 471 Networking II Firewalls Janice Regan, 2006-2013 1 Security When is a computer secure When the data and software on the computer are available on demand only to those people who should have access
Installation and configuration guide
Installation and Configuration Guide Installation and configuration guide Adding X-Forwarded-For support to Forward and Reverse Proxy TMG Servers Published: May 2010 Applies to: Winfrasoft X-Forwarded-For
Connecting with Computer Science, 2e. Chapter 5 The Internet
Connecting with Computer Science, 2e Chapter 5 The Internet Objectives In this chapter you will: Learn what the Internet really is Become familiar with the architecture of the Internet Become familiar
Creating a User Profile for Outlook 2013
Creating a User Profile for Outlook 2013 This document tells you how to create a user profile for Outlook 2013 on your computer (also known as the Outlook client). This is necessary, for example, when
INTERNET AND COMPUTER ACCEPTABLE USE POLICY (AUP)
4510.2 INTERNET AND COMPUTER ACCEPTABLE USE POLICY (AUP) The Board of Education is committed to the optimization of student learning and teaching. The Board considers computers to be a valuable tool for
Why HTML5 Tests the Limits of Automated Testing Solutions
Why HTML5 Tests the Limits of Automated Testing Solutions Why HTML5 Tests the Limits of Automated Testing Solutions Contents Chapter 1 Chapter 2 Chapter 3 Chapter 4 As Testing Complexity Increases, So
Testing Websites with Users
3 Testing Websites with Users 3 TESTING WEBSITES WITH USERS Better Practice Checklist Practical guides for effective use of new technologies in Government www.agimo.gov.au/checklists version 3, 2004 Introduction
The Internet, Intranets, and Extranets. What is the Internet. What is the Internet cont d.
C H A P T E R 7 The Internet, Intranets, and Extranets What is the Internet Millions of computers, all linked together on a computer network. A home computer usually links to the Internet using a phone
How To Study Information Security
Assessing The Relative Importance of Information Security Governance Processes Master Thesis Stockholm, Sweden 2011 XR-EE-ICS 2011:002 ASSESSING THE RELATIVE IMPORTANCE OF INFORMATION SECURITY GOVERNANCE
WhatsUp Gold v16.3 Installation and Configuration Guide
WhatsUp Gold v16.3 Installation and Configuration Guide Contents Installing and Configuring WhatsUp Gold using WhatsUp Setup Installation Overview... 1 Overview... 1 Security considerations... 2 Standard
Comparing online and lab methods in a problem-solving experiment
Behavior Research Methods 2008, 40 (2), 428-434 doi: 10.3758/BRM.40.2.428 Comparing online and lab methods in a problem-solving experiment Frédéric Dandurand, Thomas R. Shultz, and Kristine H. Onishi McGill
Managing Special Authorities. for PCI Compliance. on the. System i
Managing Special Authorities for PCI Compliance on the System i Introduction What is a Powerful User? On IBM s System i platform, it is someone who can change objects, files and/or data, they can access
ETHICAL ELECTRIC PRIVACY POLICY. Last Revised: December 15, 2015
ETHICAL ELECTRIC PRIVACY POLICY Last Revised: December 15, 2015 This policy ("Policy") covers the privacy practices that Ethical Electric, Inc. ("Ethical Electric", "we" or "us") employs with respect to
TERMS AND CONDITIONS OF USE OF WARWICK UNIVERSITY S SONA SYSTEM AND WARWICK EXPERIMENTAL LABORATORIES
TERMS AND CONDITIONS OF USE OF WARWICK UNIVERSITY S SONA SYSTEM AND WARWICK EXPERIMENTAL LABORATORIES In order to be granted access to Warwick s SONA System (https://warwick.sona-systems.com/) for managing
Lesson 1 Quiz. 3. The Internet is which type of medium? a. Passive b. Broadcast c. One-to-one d. Electronic print
Lesson 1 Quiz 1. Which technology trend in Web development can be defined as a group of XMLbased technologies that enable computers using different operating systems and software to easily exchange information
Introduction site management software
Web Testing Introduction Making a web site does not end with putting all the media and software together. Actually, web site work never ends. When all the design is done, you have to test the site first
National Cyber Security Month 2015: Daily Security Awareness Tips
National Cyber Security Month 2015: Daily Security Awareness Tips October 1 New Threats Are Constantly Being Developed. Protect Your Home Computer and Personal Devices by Automatically Installing OS Updates.
