PROCESS-ORIENTED AND PRODUCT-ORIENTED ASSESSMENT OF EXPERIMENTAL SKILLS IN PHYSICS: A COMPARISON



Similar documents
Teaching Patterns of Scientific Inquiry: A Video Study of Chemistry Lessons in Germany and Sweden

Use of E-Learning Tools in Context of Students Labs

THEORETICAL FRAMEWORK Students perception of instructional practice

AUTHOR ABSTRACT. The development of velocity and acceleration in sprints A comparison of elite and juvenile female sprinters. By Stefan Letzelter

RISK MANAGEMENT IN COMPANIES A QUESTIONNAIRE AS AN INSTRUMENT FOR ANALYSING THE PRESENT SITUATION

E-learning in the Introductory Physics Lab

Use of E-Learning Tools in Context of Students Labs - Preparation, Blended Learning, Substitute

HOW TO SUCCESSFULLY USE SOFTWARE PROJECT SIMULATION FOR EDUCATING SOFTWARE PROJECT MANAGERS

SPLITTING TUTOR ROLES: SUPPORTING ONLINE-LEARNERS WITH GROUP TUTORS AND SUBJECT TUTORS

Mit einem Auge auf den mathema/schen Horizont: Was der Lehrer braucht für die Zukun= seiner Schüler

How To Determine Your Level Of Competence

Supporting Self-Regulated Learning of Scientific Concepts through Interactive Learning Tasks Product and Process Findings

Implementation requirements for knowledge management components into ERP Systems: Comparison of software producers and companies

Dr. Hella Schick Im Aeuelchen Bonn. phone mobile: phone work:

User-centered Requirements Elicitation for Business Intelligence Solutions

CURRICULUM VITAE. M. Sc. Anne-Katharina Schiefele

What Makes a Good Online Dictionary? Empirical Insights from an Interdisciplinary Research Project

and Hung-Wen Chang 1 Department of Human Resource Development, Hsiuping University of Science and Technology, Taichung City 412, Taiwan 3

Total Quality Management in the food industry Current situation and potential in Germany

Review of Environmental Informatics Education in Germany and selected European Universities 2

ASSESSING MIDDLE SCHOOL STUDENTS CONTENT KNOWLEDGE AND SCIENTIFIC

Modularising Multilingual and Multicultural Academic Communication Competence for BA and MA level

Teachers General Pedagogical/Psychological Knowledge: Conceptualization and Test Construction

Artikel I Änderung der Studienordnung In der Anlage zur Studienordnung wird Punkt 2 wie folgt neu gefasst:

Master of Arts Program Sociology European Societies. Institute of Sociology Freie University Berlin

What does it mean to provide process-oriented training?

MARKUS ARNOLD Director Managerial Accounting Section Institute for Accounting University of Bern

3.- What atom s particle moves through a conductor material? 4.- Which are the electric components of an elemental electric circuit?

INTERNATIONAL CONFERENCE ON ENGINEERING DESIGN ICED 03 STOCKHOLM, AUGUST 19-21, 2003

Research into competency models in arts education

Strahlenschutzkommission. Geschäftsstelle der Strahlenschutzkommission Postfach D Bonn.

Publikationen und Vorträge

Evaluation of Selection Methods for Global Mobility Management Software

Pumping up Moodle via Integrated Content Authoring, Sharing and Delivery Tools The Educanext LTI Case Study

University of Münster, Institute of Political Science, Scharnhorststraße 100, Münster, Germany.

EFFICIENT ELEARNING COURSE DESIGN AND MEDIA PRODUCTION

Targeted Advertising and Consumer Privacy Concerns Experimental Studies in an Internet Context

Anne Kathrin Mayer & Günter Krampen. ZPID Leibniz Institute for Psychology Information, Trier/Germany

Dr. Anna Maria Schneider

Boom and Bust Cycles in Scientific Literature A Toolbased Big-Data Analysis

Functional Modelling in secondary schools using spreadsheets

Curriculum Vitae Professor Dr. Hans-Peter Blossfeld

Preparing teachers to teach physics and physical science by inquiry

Standard quality of format and guidelines for thesis writing at the Berlin School of Economics and Law for International Marketing Management

Curriculum vitae. I. Personal details Carsten Burhop Date of birth: 02 November 1973 Nationality: German address:

Understanding IT Governance

How to teach about transition processes and other more complex factors in so-called simple electric circuits Abstract Keywords: Introduction

Figure 1. Experiment 3 The students construct the circuit shown in Figure 2.

Experiment #3, Ohm s Law

Core Curriculum to the Course:

MEANINGS CONSTRUCTION ABOUT SAMPLING DISTRIBUTIONS IN A DYNAMIC STATISTICS ENVIRONMENT

Practical work: making it more effective

Regina Egetenmeyer, Susanne Lattke

Visual Sociology for Research and Concept Development

FORMAT OF THE PROPOSAL FOR A THESIS Revised and adapted from Practical Guide to Research Methods by Lang and Heiss.

Summary: Natalia Futekova * Vladimir Monov **

Task-Model Driven Design of Adaptable Educational Hypermedia

Self-assessment and placements tests - a worthwhile combination?

PHYSICS 111 LABORATORY Experiment #3 Current, Voltage and Resistance in Series and Parallel Circuits

Master Program in Entrepreneurship and Technology Management in Estonia

THE IMPACT OF USING ACCOUNTING INFORMATION SYSTEMS ON THE QUALITY OF FINANCIAL STATEMENTS SUBMITTED TO THE INCOME AND SALES TAX DEPARTMENT IN JORDAN

Lightning current simulation in the laboratory - parameters, procedures and test equipment -

Didactic Framework and Inventory of Methods

MISTAKE-HANDLING ACTIVITIES IN THE MATHEMATICS CLASSROOM: EFFECTS OF AN IN-SERVICE TEACHER TRAINING ON STUDENTS PERFORMANCE IN GEOMETRY

Mathematical Knowledge as a Social Construct of Teaching- / Learning Processes The Epistemology Oriented Mathematical Interaction Research

Resource-constrained Scheduling of a Real Project from the Construction Industry: A Comparison of Software Packages for Project Management

User Guidance in Business Process Modelling

A Case Retrieval Method for Knowledge-Based Software Process Tailoring Using Structural Similarity

Semantic Web. Semantic Web: Resource Description Framework (RDF) cont. Resource Description Framework (RDF) W3C Definition:

6,011,024 Jan. 4, 2000

A CASE STUDY ON SOFTWARE PROJECT MANAGEMENT IN INDUSTRY EXPERIENCES AND CONCLUSIONS

Transcription:

PROCESS-ORIENTED AND PRODUCT-ORIENTED ASSESSMENT OF EXPERIMENTAL SKILLS IN PHYSICS: A COMPARISON Nico Schreiber 1, Heike Theyßen 1 and Horst Schecker 2 1 University of Duisburg-Essen 2 University of Bremen Abstract: The acquisition of experimental skills is widely regarded as an important part of science education. Models describing experimental skills usually distinguish between three dimensions of an experiment: preparation, performance and data evaluation. Valid assessment procedures for experimental skills have to consider all these three dimensions. Hands-on tests can especially account for the performance dimension. However, in large-scale assessments the analysis of students performances is usually only based on the products of the experiments. But does this test format sufficiently account for a student s ability to carry out experiments? A processoriented analysis that considers the quality of students actions, e.g. while setting up an experiment or measuring, provides a broader basis for assessments. At the same time it is more timeconsuming. In our study we compared a process-oriented and a product-oriented analysis of hands-on tests. Results show medium correlations between both analysis methods in the performance dimension and rather high correlations in the dimensions preparation and data evaluation. Keywords: experimental skills, assessment, process-oriented analysis, product-oriented analysis, science education BACKGROUND AND FRAMEWORK The acquisition of experimental skills is widely regarded as an important part of science education (e.g. AAAS 1993, NRC 2012). Thus, there is a demand for assessment tools that allow for a valid measurement of experimental skills. In our study we compared a process-oriented and a product-oriented analysis of students performances in hands-on tests. The test instrument refers to a specific model of experimental skills. Modelling experimental skills In the literature there is a broad consensus concerning typical experimental skills, like create an experimental design, set up an experiment, observe / measure and interpret the results (e.g. NRC 2012, DEE 1999). These skills can be assigned to three dimensions of an experimental investigation: preparation, performance and data evaluation. Most models of experimental skills are structured along these three dimensions with different accentuations (Klahr & Dunbar 1988, Hammann 2004, Mayer 2007, Emden & Sumfleth 2012). Our model uses the three dimensions, too. In contrast to other models, it accentuates the performance dimension (Figure 1 adapted from Schreiber, Theyßen, & Schecker 2009).

Figure 1: Model of experimental skills: 3 dimensions, 6 components (adapted from Schreiber, Theyßen & Schecker 2009). At school, an experimental question is usually given to the students and not developed by them. So students have to interpret and clarify the given task. In non-cookbook types of situations, students have to create the experimental design themselves. During the performance, students set up the experiment, they measure and document data. During data evaluation, students process data, give a solution and interpret the results. This description might suggest a linear order of steps. However, this is not intended. The steps can occur in different orders and loops. Measuring experimental skills Written tests are established instruments to assess experimental skills. But especially with regard to the performance dimension, their validity is in question (e.g. Shavelson, Ruiz-Primo, & Wiley 1999). Other approaches for the assessment of performance skills seem to be necessary (Ruiz- Primo & Shavelson 1996, Stebler, Reusser & Ramseier 1998, Garden 1999). Here, hands-on tests show their potential. However, in large-scale assessments the analysis of students performance in experimental tasks is usually only based on the products of experimenting, mostly documented in lab sheets (e.g. Stebler, Reusser, & Ramseier 1998, Ramseier, Labudde & Adamina 2011, Gut 2012). The processes of experimentation, like setting up the apparatus and making measurements, are considered only indirectly, insofar as they affect the products. On the one hand, it is in question whether a product-oriented analysis which neglects process aspects of experimenting yields adequate ratings compared to a process-oriented analysis. On the other hand, a process-oriented analysis that considers the quality of students actions in the performance dimension (e.g. Neumann 2004) is very resource-consuming. In order to justify the additional effort for a process based analysis, it has to be shown that ratings from a product-based analysis are insufficient predictors for ratings from a process-based analysis of the same hands-on test at least for the performance phase.

RATIONALE AND METHODS Hypotheses In our study we investigate correlations between ratings from a product-based and a processbased analysis of students performances in hands-on tests. In the performance dimension students get direct feedback from the experimental setup. A non-functional electric circuit e.g. may result in a series of changes, until the setup finally works. The lab sheet will only show the final result. Similar processes may occur during measurement. A product-based analysis only evaluates the documented (i. e. usually the final) results, while a process-based approach looks at the chain of students actions. Our first hypothesis is: H1: Concerning the performance dimension, ratings from a product-oriented analysis are not highly correlated with scores from a process-oriented analysis. Students prepare the experiment and evaluate their data mostly in written form, without handling experimental devices. We assume a close relationship between what they do and what they document. Thus, our second hypothesis is: H2: Concerning the preparation and evaluation of the experiment, ratings from a productoriented analysis are highly correlated with scores from a process-oriented analysis. As a high correlation we define a correlation above 0.7 (Kendall Tau-b) Methods Tasks For the comparison of the product- und process-based analysis we developed two experimental tasks for the domain of electric circuits in secondary school curricula (Schreiber et al., 2012). The first task is Here are three bulbs. Find out the one with the highest power at 6 V. In the second task the students get a set of wires and have to find the best conductor from three metals. Students have a set of apparatus and a pre-structured lab sheet at their disposal. The lab sheet is structured along our model of experimental skills, requesting to plan a suitable experiment, assemble the experimental setup, perform measurements, evaluate the data and draw conclusions. Both tasks are open-ended and the students have to structure their paths towards the solutions on their own. They are only assisted by written information on the necessary physics content knowledge. Design Table 1 shows the design of the study. It was embedded in a more extensive study concerning the comparison of different assessment tools for experimental skills (Schreiber 2012).

Table 1 Design of the study The videos and the lab sheets were analysed according to the components of experimenting shown in Figure 1. The process-oriented analysis leads to a quality index for each student in each of these assessment categories. In a first step students actions in the videotape are assigned to one of the six components. A second step of analysis codes the qualities of intermediate stages (e.g. whether an experimental setup is corpre-test cognitive skills, content knowledge, selfconcept training introducing the hands-on test 45 min 20 min Hands-on tests group 1 and task 1 highest power group 2 and task 2 best conductor 30 min 138 upper secondary students, aged about 16 to 17, took part in this study. In a pre-test we measured personal variables that are supposed to have an influence on students test performances: cognitive skills, self-concept concerning physics and experimenting in physics, and the content knowledge in the field of electricity. Established tests and questionnaires were adapted for this pre-test (Heller & Perleth 2000, Engelhardt & Beichner 2004, von Rhöneck 1988, Brell 2008). In the hands-on test the students worked on one of the two tasks described above. The use of two different tasks was due to the design of the more extensive project into which this study was embedded. The students were assigned to the two groups based on their pre-test results in such a way that a sufficient and similar variance of the personal variables was realized in both groups. In a training session the students were introduced to the hands-on test (structures of the tasks and handling of the devices). The training task was also taken from the domain of electric circuits (measuring the current-voltage characteristic of a bulb). In the hands-on test, students worked with a set of electric devices and a pre-structured lab sheet (Figure 2, task 1). In the situation shown in Figure 2, the student documents his (inadequate) setup with two multimeters, a battery and a bulb in the lab sheet. The pre-structured lab sheet demands to clarify the question, to document the setup, to perform measurements and to interpret the results. Students can choose when and in which order they fill in the sheet. The lab sheet does not specify a particular solution or approach. Students actions were videotaped and the lab sheets were collected. Figure 2: A student performed task 1 with the hands-on test. Process-oriented analysis

rect, imperfect or wrong) and the development (e.g. whether an imperfect setup is detected and improved) (cf. Schreiber, Theyßen & Schecker 2012, Theyßen et al. 2013). The flow chart in Figure 3 illustrates an example of how the rating decisions are made. The result is a quality index on an ordinal scale with five levels. To secure validity and reliability of this analysis, several studies with high inferent expert-ratings and interviews were conducted (details: Dickmann 2009, Holländer 2009, Fichtner 2011, Dickmann, Schreiber & Theyßen 2012, Schreiber 2012). The evaluation of double coding yields a high objectivity of the ratings (Cohen s Kappa.67). Figure 3: Formal analysis scheme of the sequence analysis specified for setup skills. Product-oriented analysis For the product-oriented analysis only the students documentations in the lab sheets were analysed with regard to the same six model components (skill in Figure 1). Each entry in the lab sheet is directly associated with an assessment category. The single criterion is the correctness of the entry. A development cannot be assessed since in most cases only one result is documented in the sheets. Thus, using the formal analysis scheme (Figure 3), in the product-oriented analysis only the levels 1, 2, and 5 can be scored. Again the objectivity in each assessment category is satisfying (Cohen s Kappa >.62). RESULTS To test the hypotheses, rank correlations (Kendall-Tau b, ) between the quality parameters from the product-oriented and the process-oriented analysis were calculated for each category (Tab. 2). In all the four assessment categories that can be assigned to the preparation and the evaluation dimensions, the correlations are high (. For components of the performance dimension, we found only medium or low correlations (. Thus, both hypotheses can be confirmed. The high correlations in the planning and data evaluation dimensions can be explained by the data basis: In these dimensions the process-oriented analysis also refers mainly to the documentations in the lab sheets. Only in a few cases the videos provided further information concerning

Table 2 Correlations (Kendall-Tau b, ) between a product-oriented and a process-oriented analysis. The correlations are highly significant ( ** ) or significant ( * ). The assessment categories are assigned to the three experimental dimensions: preparation, performance and data evaluation. n: sample size. dimension assessment categories n preparation apprehend the task.877 ** 138 create the experimental design.728 ** 138 performance set up the experiment.499 ** 130 perform & document measurements.221 ** 122 data evaluation process the data & give a solution.960 ** 122 interpret the results.775 ** 117 developments. Thus, regardless of the method of analysis, in the dimensions of planning and data evaluation the scores 1, 2 and 5 dominate. In contrast, in both assessment categories belonging to the performance dimension the process-oriented analysis largely profits from the videos. The videos provide relevant information about the developments while the students work on the set up and make measurements. The low correlations between the process-oriented and the product-oriented analysis are obviously caused by an information gap between the documented setups and measurements on the one hand and the actual setupand measurement-processes in the video on the other hand. A further result can be derived from Table 2: the sample size per category decreases over the course of the experiment. Whereas 138 students clarified the questions and created an experimental design in the beginning, only 117 students interpreted the results in the end. This is a noticeable dropout of about 15 %. The reason is the use of an open task format (Fig. 2). The students had to structure the approach on their own and without any assistance. Students who e.g. did not complete the setup were in the following not able to measure and to document data. CONCLUSIONS We draw two conclusions from our results: 1. Comparison of a process-oriented and a product-oriented analysis A product-oriented analysis seems to be sufficient to analyse students skills of preparing an experiment and evaluating data. But in order to account for performance skills adequately, hands-on tests with a process-oriented analysis of students actions seem to be necessary. These findings should be considered for the development of more valid assessment procedures. 2. Open task format and sample size The use of an open task format in testing experimental skills ( Find out ) causes a noticeable dropout of students during the test. For assessing the full range of experimental skills, we suggest a guided test with non-interdependent sub-tasks. Each sub-task should refer to a specific experimental skill. To allow for a non-interdependent assessment, the item should present a sample solution of the preceding step, e.g. a measurement-item should provide a complete experimental setup. We have started to work on such a test format.

REFERENCES American Association for the Advancement of Science (AAAS) (Ed.) (1993). Benchmarks for Science Literacy. New York: Oxford University Press. Brell, C. (2008). Lernmedien und Lernerfolg - reale und virtuelle Materialien im Physikunterricht. Empirische Untersuchungen in achten Klassen an Gymnasien zum Laboreinsatz mit Simulationen und IBE. In H. Niedderer, H. Fischler & E. Sumfleth (Eds.), Studien zum Physik- und Chemielernen, Vol. 74. Berlin: Logos. Department for Education and Employment (DEE) (Ed.) (1999). Science - The National Curriculum for England. London: Department for Education and Employment. Dickmann, M. (2009). Validierung eines computergestützten Experimentaltests zur Diagnostik experimenteller Kompetenz (unpublished bachelor thesis). Dortmund: Technische Universität Dortmund. Dickmann, M., Schreiber, N. & Theyßen, H. (2012). Vergleich prozessorientierter Auswertungsverfahren für Experimentaltests. In S. Bernholt (Ed.), Konzepte fachdidaktischer Strukturierung für den Unterricht, (pp. 449 451). Münster: LIT. Emden, M. & Sumfleth, E. (2012). Prozessorientierte Leistungsbewertung des experimentellen Arbeitens. Zur Eignung einer Protokollmethode zur Bewertung von Experimentierprozessen. Der mathematische und naturwissenschaftliche Unterricht (MNU), 65 (2), 68-75. Engelhardt, P. V. & Beichner, R. J. (2004). Students understanding of direct current resistive electrical circuits. American Journal of Physics 72 (1), 98 115. Fichtner, A. (2011). Validierung eines schriftlichen Tests zur Experimentierfähigkeit von Schülern (unpublished master thesis). Bremen: Universität Bremen. Garden, R. (1999). Development of TIMSS Performance Assessment Tasks. Studies in Educational Evaluation 25(3), 217 241. Gut, C. (2012). Modellierung und Messung experimenteller Kompetenz. Analyse eines largescale Experimentiertests. In H. Niedderer, H. Fischler & E. Sumfleth (Eds.), Studien zum Physik- und Chemielernen, Vol. 134. Berlin: Logos. Hammann, M. (2004). Kompetenzentwicklungsmodelle: Merkmale und ihre Bedeutung - dargestellt anhand von Kompetenzen beim Experimentieren. Der mathematische und naturwissenschaftliche Unterricht 57(4), 196 203. Heller, K. A. & Perleth, C. (2000). Kognitiver Fähigkeitstest für 4.-12. Klassen, Revision (KFT 4-12+ R). Göttingen: Hogrefe. Holländer, L. K. (2009). Validierung eines Experimentaltests mit Realexperimenten zur Diagnostik experimenteller Kompetenz (unpublished bachelor thesis). Dortmund: Technische Universität Dortmund. Klahr, D. & Dunbar, K. (1988). Dual Space Search During Scientific Reasoning. Cognitive Science 12, 1 48. Neumann, K. (2004). Didaktische Rekonstruktion eines physikalischen Praktikums für Physiker. In H. Niedderer, H. Fischler & E. Sumfleth (Eds.), Studien zum Physik- und Chemielernen, Vol. 38. Berlin: Logos.

Mayer, J. (2007). Erkenntnisgewinnung als wissenschaftliches Problemlösen. In D. Krüger & H. Vogt (Eds.), Theorien in der biologiedidaktischen Forschung (pp. 177-186). Berlin, Heidelberg: Springer. National Research Council (NRC) (Ed.) (2012). A Framework for K-12 Science Education: Practices, Crosscutting Concepts, and Core Ideas. Washington, DC: The National Academies Press. Ramseier, E., Labudde, P. & Adamina, M. (2011). Validierung des Kompetenzmodells HarmoS Naturwissenschaften: Fazite und Defizite. Zeitschrift für Didaktik der Naturwissenschaften, 17, 7 33. Ruiz-Primo, M. A. & Shavelson, R. J. (1996). Rhetoric and Reality in Science Performance Assessments: An Update. Journal of Research in Science Teaching 33 (10), 1045 1063. Schreiber, N. (2012). Diagnostik experimenteller Kompetenz - Validierung technologiegestützter Testverfahren im Rahmen eines Kompetenzstrukturmodells. In H. Niedderer, H. Fischler & E. Sumfleth (Eds.), Studien zum Physik- und Chemielernen, Vol. 139. Berlin: Logos. Schreiber, N., Theyßen, H., Schecker, H. (2009). Experimentelle Kompetenz messen?! Physik und Didaktik in Schule und Hochschule, 8 (3), 92-101. Schreiber, N., Theyßen, H. & Schecker, H. (2012). Experimental Competencies In Science: A Comparison Of Assessment Tools. In C. Bruguière, A. Tiberghien & P. Clément (Eds.), E- Book Proceedings of the ESERA 2011 Conference, Lyon France. Retrieved from: http://www.esera.org/media/ebook/strand10/ebook-esera2011_schreiber-10.pdf (29.11.2013). Shavelson, R. J., Ruiz-Primo, M. A. & Wiley, E. W. (1999). Note on Sources of Sampling Variability in Science Performance Assessments. Journal of Educational Measurement, 36 (1), 61-71. Stebler, R., Reusser, K. & Ramseier, E. (1998). Praktische Anwendungsaufgaben zur integrierten Förderung formaler und materialer Kompetenzen - Erträge aus dem TIMSS- Experimentiertest. Bildungsforschung und Bildungspraxis 20 (1), 28 54. Theyßen, H., Schecker, H., Gut, C., Hopf, M., Kuhn, J., Labudde, P., Müller, A., Schreiber, N., Vogt, P. (2013). Modelling and Assessing Experimental Competencies in Physics. In C. Bruguière, A. Tiberghien & P. Clément (Eds.), 9th ESERA Conference Contributions: Topics and trends in current science education - Contributions from Science Education Research (pp. 321 337). Dordrecht: Springer. von Rhöneck, C. (1988). Aufgaben zum Spannungsbegriff. Naturwissenschaften im Unterricht - Physik/Chemie, 36(31), 38 41.