Student Evaluation of Teaching (SET) in Higher Education: How to Use SET More Effectively and Efficiently in Public Affairs Education



Similar documents
UNH Graduate Education Department. Quarterly Assessment Report

Student Response to Instruction (SRTI)

PROCEDURES AND EVALUATIVE GUIDELINES FOR PROMOTION OF TERM FACULTY DEPARTMENT OF PSYCHOLOGY VIRGINIA COMMONWEALTH UNIVERSITY MARCH 31, 2014

MSSW Application Requirements

How To Determine Teaching Effectiveness

Strategies for Teaching Undergraduate Accounting to Non-Accounting Majors

MSW Application Requirements. General Requirements...2. Additional Requirements by Program 4. Admissions Essays...5. International Applicants 7

MPA Program Assessment Report Summer 2015

Comparison of Student and Instructor Perceptions of Best Practices in Online Technology Courses

NORTHERN ILLINOIS UNIVERSITY. College: College of Business. Department: Inter-Departmental. Program: Master of Business Administration

Students Perception Toward the Use of Blackboard as a Course. Delivery Method. By Dr. Ibtesam Al mashaqbeh

Adjunct Faculty Orientation and Professional Development Custom Research Brief

Master of Arts in Teaching/Science Education Master of Arts in Teaching/Mathematics Education

Department of History Policy 1.1. Faculty Evaluation. Evaluation Procedures

Annual Assessment Report - The College of Information Systems Program

A Modest Experiment Comparing HBSE Graduate Social Work Classes, On Campus and at a. Distance

University Undergraduate Teaching Quality Chapter 3 Section. Background. Ministry of Training, Colleges and Universities

Student Success in Business Statistics

UNIVERSITY OF BELGRADE FACULTY OF PHILOSOPHY. Part two: INFORMATION ON DEGREE PROGRAMS

University Senate Task Force on Faculty Evaluations FINAL REPORT January 9, 2015

EXECUTIVE SUMMARY. List all of the program s learning outcomes: (regardless of whether or not they are being assessed this year) Learning Outcome 1

Principal Appraisal Overview

ATTITUDES OF ILLINOIS AGRISCIENCE STUDENTS AND THEIR PARENTS TOWARD AGRICULTURE AND AGRICULTURAL EDUCATION PROGRAMS

FACULTY PEER ONLINE CLASSROOM OBSERVATIONS AA

students to complete their degree online. During the preliminary stage of included Art, Business, Computer Science, English, Government, History, and

Enrollment Data Undergraduate Programs by Race/ethnicity and Gender (Fall 2008) Summary Data Undergraduate Programs by Race/ethnicity

THE MASTER OF ARTS PROGRAM IN INDUSTRIAL/ORGANIZATIONAL PSYCHOLOGY GRADUATE SCHOOL OF ARTS AND SCIENCE NEW YORK UNIVERSITY

Wasserman Center for Career Development LIFE BEYOND THE SQUARE

G E N E R A L I N F O R M A T I O N F O R G R A D U A T E S T U D E N T S

The Influence of a Summer Bridge Program on College Adjustment and Success: The Importance of Early Intervention and Creating a Sense of Community

RESEARCH ARTICLES A Comparison Between Student Ratings and Faculty Self-ratings of Instructional Effectiveness

Running head: CREDIT RECOVERY 1. Report Submitted to the Academic Partnerships

VANDERBILT UNIVERSITY DEPARTMENT OF SOCIOLOGY GRADUATE PROGRAM GUIDE

1. List proposed program changes related to merging, splitting, renaming or modifying a program. Community Counseling and School Counseling.

Graduate Faculty Committee Doc. No. 1075A Approved April 20, 2009

Program Assessment: School of Criminology and Criminal Justice.

PROGRAM HANDBOOK

STUDENTS PERCEPTIONS OF ONLINE LEARNING AND INSTRUCTIONAL TOOLS: A QUALITATIVE STUDY OF UNDERGRADUATE STUDENTS USE OF ONLINE TOOLS

An Analysis of IDEA Student Ratings of Instruction in Traditional Versus Online Courses Data

CHAPTER 1 INTRODUCTION. stakeholders, and a subject that goes beyond the world of researchers, given its social,

Annual Goals for Math & Computer Science

OVERVIEW AND ADMISSION REQUIREMENTS

Special Requirements of the Educational Administration

Review of the M.A., M.S. in Psychology

IDEA. Student Ratings of Instruction Survey Report Fall Guam Community College

Evaluation in Online STEM Courses

The University of Arizona

V. Course Evaluation and Revision

National Standards. Council for Standards in Human Service Education (2010, 1980, 2005, 2009)

Examples of Teacher-Designed/Scored Feedback Questionnaires A Guided Self-Analysis for Beginning Instructors

GRADUATE DEGREE REGULATIONS

CTL 2009 ADVANCED PROGRAM REPORT

University of Missouri-Columbia. MU Sinclair School of Nursing. GUIDELINES for APPOINTMENT, REAPPOINTMENT, and PROMOTION of NON-REGULAR FACULTY

Cerritos College Associate Degree Nursing Program

. P S Y C H O L O G Y.

Outcome: Compare and contrast different research methods used by psychologists including their respective advantages and disadvantages.

Background. 1 During the fall semester of 2008, graduate faculty members reviewed all syllabi for graduate

Frequently Asked Questions Howard University APA Accredited Clinical Psychology PhD Program

パンフレット Okinawa International University (Key phrases) Peace and Coexistence / Individuality and Creativity / Autonomy and Development

INTENSIVE TEACHING IN LAW SUBJECTS

IPP Learning Outcomes Report. Faculty member completing template: Greg Kim Ju, Marya Endriga (Date: 1/17/12)

THE Ph.D. PROGRAM IN MARKETING. The Smeal College of Business The Pennsylvania State University

Tennessee State Board of Education February 1, 2013 First Reading Item: III. A. Teacher Licensure Standards

OMANI SCIENCE TEACHERS USE OF COOPERATIVE WORK IN BASIC EDUCATION SCHOOLS

Program Assessment Report. Unit Psychology Program name: Clinical Psychology MA Completed by David Grilly May 2007

Changing Practice in an Ed.D. Program

Does Macro/Micro Course Sequencing Affect Student Performance in Principles of Economics Courses?

Doctoral Degree Programs in Special Education

Virginia Tech Department of Accounting and Information Systems Ph.D. Program GENERAL INFORMATION

Quality Measurement and Good Practices in Web-Based Distance Learning:

Reproductions supplied by EDRS are the best that can be made from the original document.

Summary of the Case American Intercontinental Universit y1 Master of Education Program

Division of Communication Disorders AP4 Strategic Plan ( )

Onsite Peer Tutoring in Mathematics Content Courses for Pre-Service Teachers

SELF-STUDY FORMAT FOR REVIEW OF EXISTING DEGREE PROGRAMS

Chapter 5. Summary, Conclusions, and Recommendations. The overriding purpose of this study was to determine the relative

STUDENT ATTITUDES TOWARD WEB-BASED COURSE MANAGEMENT SYSTEM FEATURES

Comparing the Perceptions of Online Learning between Students with Experience and Those New to Online Learning

Department of Psychology

Educational Leadership Advising Handbook: Addendum to the Graduate Handbook

Realizeit at the University of Central Florida

Psychology. Forensic Psychology Master of Science (M.S.)

THE MASTER OF ARTS PROGRAM IN GENERAL PSYCHOLOGY GRADUATE SCHOOL OF ARTS AND SCIENCE NEW YORK UNIVERSITY. Information Booklet for Applicants

An Analysis of Student Satisfaction in Higher Education Courses Delivered Online and in the Live Classroom

Presented at the 2014 Celebration of Teaching, University of Missouri (MU), May 20-22, 2014

Hiring Guidelines for Teaching Associates & Student Assistants

Third Year Report: Graduate Education Programs

Section Two: Ohio Standards for the Teaching Profession

Course Evaluations at the Kelley School of Business

Recommended Course Sequence MAJOR LEADING TO PK-4. First Semester. Second Semester. Third Semester. Fourth Semester. 124 Credits

Cognitive Area Program Requirements 10/4/12 1

The coach-team approach: An introductory accounting instructional alternative

19K660. Brooklyn, NY Jocelyn Badette

2. State the job opportunities for graduates of the proposed program. Include evidence documenting those opportunities.

Section 2: Program Summary Economics (CA): Secondary Major and Minor

Journal of College Teaching & Learning July 2008 Volume 5, Number 7

The Impact of Web-Based Instruction on Performance in an Applied Statistics Course

ACT CAAP Critical Thinking Test Results

High-Impact Practices and Experiences from the Wabash National Study

UMD Department of Mechanical and Industrial Engineering

Transcription:

Student Evaluation of Teaching (SET) in Higher Education: How to Use SET More Effectively and Efficiently in Public Affairs Education Koichiro Otani, B. Joon Kim, and Jeong-IL Cho Indiana University Purdue University Fort Wayne Abstract The present study examined the impact of 13 aspects involved in student evaluation of teaching (SET). The study used student course evaluations of 1,410 university students in the Division of Public and Environmental Affairs, Indiana University Purdue University Fort Wayne. These evaluations were collected in Fall 2009, Spring 2010, and Summer 2010. The goal was to identify and prioritize influential variables in the SETs. The four items rated as most influential in the overall effectiveness of instructors teaching are ranked as follows: (1) clear explanation, (2) effective use of class time, (3) positive learning environment, and (4) stimulating course materials. The results showed that evaluation items in SETs are not equally influential in their effects on students overall rating of teaching effectiveness. Thus this study revealed that a prioritized list of aspects needs to be developed, and efforts should be concentrated on improving these aspects to increase effectiveness and efficiency in rating of teaching. Improvement strategies for instructors as well as future implications of the effective use of SET at the department or college level are discussed. Student evaluation of teaching (SET) in higher education has a long history, originating in the 1920s. Much research has suggested that educators should make better use of the routinely collected SET data (Campbell & Bozeman, 2008; Griffin & Cook, 2009; Wiers-Jenssen, Stensaker, & Grøgaard, 2002; Wolfer & Johnson, 2003). Data from SETs can be more effectively and efficiently used in improving the quality of teaching at program, department, and college levels, when institutions identify influential aspects on the overall teaching performance of their instructors, disseminate the outcomes, and make a timely response as recommended at the United Kingdom (UK) 2007 National Conference on Student Evaluation (Griffin & Cook, 2009). Such JPAE 18(3), 531 544 Journal of Public Affairs Education 531

K. Otani, B. Joon Kim, & J. Cho shared information would provide groundwork to initiate practical and proactive conversation among college and university instructors to meet the unique needs of their diverse student body and ultimately to improve the overall quality of teaching in their courses and learning environments for their students. The reason for lack of effective responsiveness to SET is a result of the traditional focus on the utilization of SET data: privately owned data and personal improvement efforts. The outcome from SET informs individual instructors about aspects of their teaching that need to be improved. Data from SET can assist the instructors to be conscious of their strengths and weaknesses in teaching, and it can eventually contribute to improving their teaching effectiveness and enhance the quality of student learning (McKeachie, 1997; Pike, 1998). The instructors are personally aware of areas to be improved upon in their own teaching based on the results of SET from their own classes. They may share such evaluation information with their mentor, if any, and seek their advice on improving their teaching effectiveness. However, it is commonly found that individual instructors do not have a collective knowledge and understanding of their teaching effectiveness in the classroom as a group (e.g., department and college). In most institutions, data from SETs remains as an instructor s private material, exclusively shared with administrators. It is not mentioned or discussed in a collective manner such as during department-wide training or discussion. Such utilization of SET data is in some way due to the sensitive nature of the SET. The SET has been used as a primary indicator of teaching effectiveness of college and university instructors in the process of promotion and tenure decisions, annual reviews, and reappointment determinations (d Apollonia & Abrami, 1997; McKeachie, 1997; Pike, 1998). Both individual and administrative uses of the SETs have been well accepted in an academic community and justify the collection of SETs. College and university students routinely complete a summative course evaluation toward the end of each semester throughout their degree or nondegree programs. Questionnaires are generally the most frequently used format of SETs using rating scales. It is reasonable for students to assume that their feedback has an impact on the modification, transformation, and improvement in future lecture formats and materials, course requirements, assessment methods, and classroom management and interactions. College and university students, however, seldom see the changes they expect in their future courses that would probably be taught by either the same or a different instructor (Campbell & Bozeman, 2008; Griffin & Cook, 2009). Such nonresponsiveness toward their feedback could cause frustration among students and discourage them from providing genuine and considered feedback on their classes. By using the SET data, instructors can easily identify weak aspects of teaching, since these aspects are often rated lower relative to other aspects. However, this simple prioritizing approach fails to provide which aspect(s) need 532 Journal of Public Affairs Education

Student Evaluation of Teaching to be improved first and urgently. A study conducted by Wolfer and Johnson (2003) emphasized an importance of effective uses of SET data for individual teaching improvement and reported a challenge of identifying patterns of areas needed to be improved. Those areas can be used in a departmental level of training for teaching improvement of instructors. Without a systematically combined, prioritized set of information, instructors could arbitrarily choose some aspects over others based on their individual evaluation data with a hope that their future students overall satisfaction with their teaching and improvement efforts would be positive. Extensive studies have investigated a range of variables on teaching effectiveness and documented inconsistent and inconclusive results of their effects on SETs (Campbell, Steiner, & Gerdes, 2005; McKeachie, 1997; Whittington, 2001; Wolfer & Johnson, 2003). Most of these variables (e.g., class size, curricular area, prior interest in the course topic, expected grades, workload/difficulty, whether a course is required or elective, challenging level of course, instructor characteristics, course format whether online or traditional) are not in an instructor s control and are not usually collected in routine SETs. Moreover, many studies disagreed on specific variables identified as more influential on SET scores than others. According to a study of Campbell and her colleagues (2005), SET scores are influenced by accumulated effects of a wide variety of variables that are both uncontrollable (e.g., gender, age, and race of instructors; class size; starting time; class with primarily quantitative content) and controllable (e.g., instruction delivery mode, time allocation for different activities) by instructors. At the UK 2007 National Conference on Student Evaluation: Dissemination and Debate, higher education professionals proposed strategies to effectively use data from SET that are routinely collected from students (Griffin & Cook, 2009). This conference was a collaborative attempt to respond to the underutilization issue of the SET data and to redirect attention to what instructors actually can do to improve teaching effectiveness using the outcomes of the routinely collected SET data. Prioritizing actions by ranking the problematic aspects of teaching reported in SETs is one of the practical recommendations in using SETs more effectively and efficiently. These actions focus on variables, mostly under the instructors control, to improve the quality of teaching and students overall satisfaction with teaching. Instead of distracting educators attention with a wide range of variables that may accumulatively affect the SET data, there is a need to redirect our focus on variables under the instructors control. Indeed, in a comprehensive review of related literature, Campbell and her colleagues (2005) reported the need to pay more attention to variables under the control of instructors to improve teaching effectiveness. Then, the next question would be Which controllable aspects are more influential? The present study specifically addresses this question. Journal of Public Affairs Education 533

K. Otani, B. Joon Kim, & J. Cho The SET measure that is employed by the Public Affairs program at Indiana University Purdue University Fort Wayne (IPFW) contains one global rating item Overall, I would rate the instructor of this course as outstanding that refers to the overall satisfaction of an instructor s teaching effectiveness as recommended by Cashin and Downey (1992). This study examined the impact of the 13 aspects (independent variables) on the overall satisfaction of the instructor s teaching effectiveness (dependent variable) with the intention of distinguishing and prioritizing influential variables on the SET. Influential variables within SETs in public affairs education deserve more attention, and identifying variables that are within instructors control would be potentially crucial in improving teaching and, ultimately, in enhancing student learning. Methodology Students combine each of their experiences (aspects) to arrive at their overall evaluation of the instructor. Some aspects may be good and pleasant, while others may not. Some aspects are more influential than others to the students in forming their overall satisfaction level of an instructor s teaching effectiveness. Thus, when they have a positive experience with those salient aspects, their overall satisfaction is likely to be positive. On the other hand, if they have a negative experience with salient aspects, they are likely to be unsatisfied even when other less salient aspects were positive. In the field of marketing and other studies in related areas, a number of researchers use the Fishbein model to predict which product with multiple aspects would be selected by customers (Fishbein & Ajzen, 1975). This model has been used by many studies in customer satisfaction research including, but not limited to, patient satisfaction (Otani, Kurz, & Harris, 2005). Thus this well-established model is also relevant to Student Evaluation of Teaching studies. This model would suggest that to improve student overall evaluation, it is critical to improve more influential aspects than less influential aspects. Data Sources This study used student course evaluations from the Division of Public and Environmental Affairs (DPEA) at Indiana University Purdue University Fort Wayne (IPFW). Thirteen aspects (elements) were used to assess the course, and these items were considered as independent variables for the purpose of the study. The dependent variable was Overall, I would rate the instructor of this course as outstanding. The current set of items, including the overall evaluation item, has been used for SET since Fall 2008 at DPEA-IPFW. All items came from a pool of suggested items at the School of Public and Environmental Affairs (SPEA) program in the Indiana University System, and the pool of suggested items has been used for more than three decades. The data were collected in Fall 2009, Spring 2010, and Summer 2010 at DPEA-IPFW. The SET survey uses a Likert-type scale, and the operationalization of the survey for each item is as follows: 5 = Strongly Agree, 4 = Agree, 3 = Neither Agree nor Disagree, 2 = Disagree, and 1 = Strongly Disagree. 534 Journal of Public Affairs Education

Student Evaluation of Teaching Instructors distributed the SET survey at the end of each semester in class, but the instructor was not allowed to stay in the classroom when students were responding to the survey. One designated student collected the completed surveys and took them directly to the secretary of the division. Data Analysis The study analyzed the combining process of the 13 aspects in arriving at an overall rating of the course and the instructor. In other words, it evaluated the 13 aspects and attempted to find which aspects have more influence on students overall rating. To analyze the relative importance of the 13 aspects, a general regression model was used. The general model for the i th case is n Y = a + S b i x i +e i = 1 where Y is overall students evaluations of teaching, a is the intercept, b i is a coefficient, x i is an experience of the i th aspect, and e is an error term. The significance of b i and the value of R 2 are examined to test the model fit. Results There were 1,410 student responses. Among them, 545 cases (38.7%) were taught by limited-term lecturers, and 859 cases (60.9%) were taught by full-time faculty members. Six cases are missing values. The data set includes three university terms, and the distribution is as follows: 677 cases, 48.0% (Fall 2009); 680 cases, 48.2% (Spring 2010); and 53 cases, 3.8% (Summer 2010). To prevent possible identification of students and to lead to more honest and candid responses, the SET survey did not ask for students demographic backgrounds (age, gender, race, etc.). Instead of the sample statistics, we offer our division information. In the fall semester, 2009, there were 498 students enrolled in the Division of Public and Environmental Affairs (DPEA) at IPFW. Among them, 229 were male and 269 were female. The average age for male students was 23.8, and that for female students was 26.9. There were 453 undergraduate students, and 45 graduate students. Among undergraduate students, 76.2% were full-time students, and 23.8% were part-time students. Among graduate students, 35.6% were full-time students, and 64.4% were part-time students. Among all 498 students, white students accounted for 403 followed by black students (57), Hispanic students (11), Asian students (7), American Indian students (2), and others (12). There were 6 international students. The Division offers five majors: Criminal Justice, Environmental Policy, Health Services Administration, Legal Studies, and Public Management. The numbers of students in classes range from 6 to 45, and a typical class size is 30. There were nine faculty members in the division. Among them, three were female faculty members, and five were senior faculty members (associate or full professors). There were also 15 limited-term lecturers, although each of them typically teaches only one course. Journal of Public Affairs Education 535

K. Otani, B. Joon Kim, & J. Cho In the data, average students responses for independent variables (evaluation items) are generally high; they range from 4.10 to 4.57. The average overall evaluation (dependent variable) is 4.23, and its standard deviation is 1.127. The descriptive statistics and the description of all aspects (items) are shown in Table 1. Table 1. Descriptive Statistics of Question Items Question Items Mean SD N Independent Variables Q1: My instructor is well prepared for class meetings. 4.57 0.829 1405 Q2: My instructor explains the subject clearly. 4.21 1.121 1408 Q3: My instructor is enthusiastic about teaching 4.48 0.906 1406 this course. Q4: Course materials were thought-provoking 4.10 1.101 1405 and stimulating. Q5: My instructor is available for consultation. 4.37 0.929 1407 Q6: I know what is expected of me in this course. 4.34 1.013 1407 Q7: The exams cover the most important aspects 4.24 1.060 1405 of the course. Q8: My instructor evaluated student work in fair 4.35 1.001 1405 and appropriate ways. Q9: This course fulfilled the objectives described 4.43 0.904 1407 in the syllabus. Q10: My instructor created an environment in 4.44 1.006 1405 which students felt comfortable asking questions and expressing their views. Q11: My instructor encouraged students to 4.44 0.908 1403 participate in their learning. Q12: My instructor made effective use of class time. 4.33 1.050 1407 Q13: I acquired new knowledge in this course. 4.37 1.001 1403 Dependent Variable Q14: Overall, I would rate the instructor of this course as outstanding. 4.23 1.127 1389 The multiple linear regression analysis was conducted, and the result is shown in Table 2. The R 2 is 0.828, and thus the model explains 82.8% of the variance in Y. The analysis revealed that certain variables are more influential than others when students rate their overall evaluation. The magnitude of influence is determined by the value of the coefficient. The larger the value of the coefficient, the more influence. The most influential variable is Q2 536 Journal of Public Affairs Education

Student Evaluation of Teaching (The coefficient is 0.364) followed by Q10, Q12, Q4, Q13, Q7, Q8, and Q5 (coefficients ranging from 0.134 to 0.052) in this order. All of them are statistically significant at the level of α = 0.05, and they are positively related. Other variables (Q1, Q3, Q6, Q9, and Q11) are not statistically significant at the level ofα= 0.05. Compared to full-time faculty members, limited-term lecturers have more positive overall rating, and this relationship is statistically significant. Table 2. Parameter Estimates of All Question Items and Rank Order of Significant Items Independent Variables Parameter Estimates Standard Error P Value Intercept 0.491 0.088 0.000 Q1. Class preparation 0.029 0.026 0.276 Q2. Clear explanation 0.364 0.023 0.000 1 Q3. Enthusiasm 0.030 0.025 0.229 Q4. Stimulating course materials 0.105 0.019 0.000 4 Q5. Instructor availability 0.052 0.020 0.010 8 Q6. Student understanding of expectation 0.034 0.024 0.168 Q7. Comprehensive exam 0.087 0.021 0.000 6 Q8. Fair assessment 0.072 0.024 0.003 7 Q9. Course objectives 0.009 0.027 0.734 Q10. Positive learning environment 0.134 0.023 0.000 2 Q11. Encouragement in participation 0.004 0.024 0.882 Q12. Effective use of class time 0.117 0.023 0.000 3 Q13. Gaining new knowledge 0.091 0.024 0.000 5 Full-time 0.066 0.028 0.019 Note. N = 1368; R 2 = 0.828. Limited-term lecturers are a reference group. Rank Order Discussion As recommended in the UK 2007 National Conference in Student Evaluation, the present research attempted to prioritize aspects (variables) of teaching embedded in SETs that are more influential on students overall satisfaction of instructors teaching ( Overall, I would rate the instructor of this course as outstanding ) in SETs. The results of the study showed that evaluation items on the SET are not equally influential in their effects on students overall rating of teaching effectiveness. Among 13 evaluation items, eight items were statistically significant and positively influential, and five items were not statistically significant on the rating of the overall item addressing instructors teaching effectiveness. Journal of Public Affairs Education 537

K. Otani, B. Joon Kim, & J. Cho Consistent with previous findings (Cashin & Downey, 1992), certain evaluation items are critically more influential to students overall rating of teaching than other items (see Table 2 and Figure 1). It was revealed that the most influential evaluation item is the clarity of instructor s explanations of the course subject (Q2: My instructor explains the subject clearly). It was noteworthy that this clarity item is far more influential than the second most influential item, positive and risk-free learning environment (Q10: My instructor created an environment in which students felt comfortable asking questions and expressing their views). The following items, in the order shown, are less influential on students overall rating of teaching, but gradually less influential. The rank order of these aspects would help readers to clearly understand their relative importance. Third most influential item an effective use of class time (Q12: My instructor made effective use of class time) Fourth most influential item stimulating course materials (Q4: Course materials were thought-provoking and stimulating) Fifth most influential item gaining new knowledge (Q13: I acquired new knowledge in this course) Sixth most influential item comprehensive exams (Q7: The exams cover the most important aspects of the course) Seventh most influential item fair assessment (Q8: My instructor evaluated student work in fair and appropriate ways) Eighth most influential item availability for consultation (Q5: My instructor is available for consultation) Figure 1. Relative Importance of Each Salient Aspect in Student Evaluations of Teaching Parameter estimate Salient aspects 538 Journal of Public Affairs Education

Student Evaluation of Teaching The top four ranked items (i.e., clear explanations, positive learning environment, effective use of class time, and stimulating course materials; Questions 2, 10, 12, and 4 respectively) are, to some extent, related to class preparation and are generally what instructors can attempt to control in their teaching. This result supports the previous finding that student ratings of teaching effectiveness are closely associated with identifiable skills of instructors, such as organization and grading skills (Jirovec, Ramanathan, & Alvarez, 1998). This finding provides a concrete pattern of areas needing to be improved that can be used in a departmental training for teaching improvement of instructors (Wolfer & Johnson, 2003). Such collective utilization of SETs can not only improve teaching quality among instructors in a department but also enhance student learning experience in a Public Affairs program. However, an item directly related to class preparation (Q1: My instructor is well prepared for class meetings) was not shown to be statistically significant on the students overall rating of teaching. This direct measure of instructor preparedness was not identified as the top-ranking influential item. This result may indicate that instructor preparation is a prerequisite and that students may judge the outcomes of instructor preparation more than just the state of preparation. To improve the overall quality of class preparation, an instructor can devote his or her time and effort to comprehending the topic and being prepared to explain the subjects clearly; contribute to creating a positive learning environment where students feel comfortable in sharing their opinions and posing questions; and plan the use of class time carefully. Moreover, these three items are also found in the Purdue University s Course and Instructor Appraisal System that has been developed prudentially to create a comprehensive pool of SET questions. According to the system, these items can be categorized under (a) clarity and effectiveness of presentation, (b) student interest/involvement learning, and (c) respect and rapport. The fourth-ranked item (Q4: Course materials were thought-provoking and stimulating), however, may or may not be under a direct control of an instructor, but an instructor may be able to stimulate student interest by exposing them to real-world examples such as conducting a research project of local sources and having a guest speaker with extensive experience and knowledge in a certain topic. The fifth-ranked item (Q13: I acquired new knowledge in this course) addresses students learning and has long been debated by extensive research (Denhardt, 2001) due to its measurement complication using the SET data alone. The sixth- and seventh-ranked items (Q7: The exams cover the most important aspects of the course; Q8: My instructor evaluated student work in fair and appropriate ways) among the eight influential items are about assessment-related skills and are rather objective evaluations of the course. This finding is consistent with previous studies (Jirovec et al., 1998). Based on the results of the present study, it is clear that when students think they have Journal of Public Affairs Education 539

K. Otani, B. Joon Kim, & J. Cho gained a new skill, they give credit to their instructors. There are many things instructors can do to respond to this finding. An instructor may be able to positively influence assessment and student learning by carefully developing exam questions, using fair grading systems, and proactively assuring students learning through constant review of key points, particularly during exam review sessions (Campbell et al., 2005). However, having close interactions with students, which is partially addressed in the eighth-ranked item (Q5: My instructor is available for consultation) indicates being available for consultation. To improve the quality of teaching among instructors at departmental and college levels, they can work collaboratively and provide input on development and implementation of comprehensive examinations, fair grading systems, and effective teaching and student learning strategies. Among all the evaluation items, five did not show statistical significance. Thus these aspects are not likely to be associated with students overall rating of instructors teaching effectiveness (Instructors preparedness: Q1: My instructor is well prepared for class meetings; enthusiasm in teaching: Q3: My instructor is enthusiastic about teaching this course; students understanding of the course expectation: Q6: I know what is expected of me in this course; accomplishment of course objectives: Q9: This course fulfilled the objectives described in the syllabus; and instructor s encouragement in course participation: Q11: My instructor encouraged students to participate in their learning). It is noted that two items address student learning behaviors (i.e., student understanding of the course expectation and course participation) and are used to solicit students evaluation of their own learning behaviors exclusively and partially. As noted earlier, students perceived learning of new knowledge is ranked as the fifth most influential item. These findings show that students gave credit to instructors for their perceived final outcome of the course (Q13: I acquired new knowledge in this course), but did not give credit to instructors regarding their process of learning. The students might have obtained new knowledge that was possibly influenced and encouraged by instructors, but they were not able to perceive the instructor s influence. It can be partially explained by the previous finding that teaching behaviors of instructors that help students learn new knowledge might be evaluated differently from those that help students participate more in class activities and understand the course expectations (Cashin & Downey, 1992). Instructors enthusiasm and the fulfillment of course objectives (Q3 and Q9) were not influential in the students overall satisfaction with instructors, when enthusiasm is typically considered as a positive asset of an individual, and the fulfillment of course objectives is another part of positively identifiable and observable skills. These findings imply that instructors enthusiasm in teaching the subject and a diligent delivery of planned course contents may not be highly valued by students as an important quality of instructors. 540 Journal of Public Affairs Education

Student Evaluation of Teaching Limitations and Suggestions This study was carefully conducted with solid methodology; however, there are some limitations. First, the data were collected at only one school in the university. Even though there are 1,410 cases and this number is quite large, it is not a random sample of all students. Thus it is not possible to generalize our results to all departments and all universities in the United States. Future studies are encouraged to conduct this type of study in different schools and universities. Second, because this is a cross-sectional study and not an experimental study, this type of design cannot establish a cause-and-effect relationship. Nevertheless, this study used a well-established model (Fishbein model) to support the causeand-effect relationship where independent variables influence the dependent variable of overall evaluation of the instructor. If a future study employs an intervention design where influential aspects of teaching are improved, the cause-and-effect relationship can be more clearly established. Third, this SET survey used a Likert-type scale. The variables are, thus, ordinal. Some researchers argue the appropriateness of using a general regression model on ordinal data. However, research studies show that the use of regression models on ordinal data is appropriate and useful, and the results are generally robust (Labovits, 1970; O Sullivan & Rassel, 1989). Many research studies have used regression models on ordinal data. Fourth, this study used the SET survey that had been carefully developed and used for more than three decades at Indiana University Purdue University Fort Wayne. However, it is possible that the survey may not contain all salient aspects in teaching, and that limitation may bias the estimates. Further refinement and examination of the survey items are warranted. As recommended in the UK 2007 National Conference on Student Evaluation (Griffin & Cook, 2009), proactive, ongoing discussions among university instructors need to take place on how to use the evaluation results and how to improve teaching effectiveness. To improve overall quality of the program and to provide recommendations and directions for their instructors, university administrators need to focus not only on numerical means or medians of SETs of individual instructors, but on areas that need improvement. The SET data of instructors in a department or college need to be combined and analyzed to prioritize actions, disseminate the outcomes, and make timely responses to students (Griffin & Cook, 2009). Both experienced and new instructors would benefit from this shared knowledge about teaching, and they could have an opportunity to reflect on and reevaluate their teaching. In an effort to enhance the quality of teaching in the Public Affairs program at IPFW and other universities when appropriate, instructors need to follow up on these prioritized items in a timely, unified, and collaborative manner, disseminate relevant results to other instructors and students using multiple delivery means, including department or college websites and online learning tools (e.g., Blackboard), identify necessary improvement goals, prepare a clear Journal of Public Affairs Education 541

K. Otani, B. Joon Kim, & J. Cho and detailed timeline, promote a positive use of the evaluation results, and solicit active student involvement in achieving goals (Griffin & Cook, 2009). Furthermore, the Public Affairs program needs to obtain multiple perspectives and data sources in identifying issues by using multiple measures (e.g., interviews, observations, focus group discussion, and curriculum review; Griffin & Cook, 2009). All of these improvement efforts can yield a meaningful result when all involved parties including instructors, students, and administrators in the program understand the value of a unified effort in improving the quality of teaching and are open to receiving constructive feedback from each other. Conclusion This study advances the existing knowledge of SETs that all salient aspects of teaching do not equally influence students overall evaluation of the instructor in Public Affairs education. Results from this study suggest that the clarity of instructors explanation of the subject is far more influential than other aspects in the SET, and it should be emphasized as a critical quality of university instructions to improve the overall evaluation of teaching. This finding is logical because students take a course to understand the material, learn the skills necessary to pass the course, and ultimately compete in the real world. An instructor s clear explanation of the subject is the first key step of student learning. For clear explanation, it is necessary for the instructor to take time to be well prepared for the class. An instructor should know the subject well; but as the saying goes, a good player is not always a good coach. Accordingly, even though the instructor may know the subject very well, it cannot be assumed that the instructor is able to explain the subject clearly to students. In addition to the explanation side of teaching, other aspects, especially positive learning environment, the effective use of class time, and stimulating course materials should be pursued to improve teaching and accommodate better learning opportunities for students. These findings clearly show that an instructor needs to take time to prepare for the class and develop a solid and achievable plan of teaching before going to class. By prioritizing actions based on the scientific evidence and having a systematic improvement plan, instructors as a group can collaboratively and constructively assist each other on these issues related to teaching and learning in higher education. This approach would enhance the instructors overall teaching effectiveness and ultimately improve overall program quality. Acknowledgment The authors thank Tamara G. Davich for help with preparation of the data, and we are grateful to Dawn Adams and Shelley Hart for their comments. 542 Journal of Public Affairs Education

Student Evaluation of Teaching References Campbell, H. E., Steiner, S., & Gerdes, K. (2005). Student evaluations of teaching. Journal of Public Affairs Education, 11(3), 211 231. Campbell, J., & Bozeman, W. C. (2008). The value of student ratings: Perceptions of students, teachers, and administrators. Community College Journal of Research & Practice, 32(1), 13 24. Cashin, W. E., & Downey, R. G. (1992). Using global student rating items for summative evaluation. Journal of Educational Psychology, 84(4), 563 572. Denhardt, R. B. (2001). The big questions of public administration education. Public Administration Review, 61(5), 526 534. d Apollonia, S., & Abrami, P. C. (1997). Navigating student ratings of instruction. American Psychologist, 52(1), 1198 1208. Fishbein, M., & Ajzen, I. (1975). Belief, attitude, intention and behavior: An introduction to theory and research. Reading, MA: Addison-Wesley Publishing. (* Thank you.) Griffin A., & Cook, V. (2009). Acting on evaluation: Twelve tips from a national conference on student evaluations. Medical Teacher, 31, 101 104. Jirovec, R. L., Ramanathan, C. S., & Alvarez, A. R. (1998). Course evaluation: What are social work students telling us about teaching effectiveness? Journal of Social Work Education, 34(2), 229 236. Labovitz, S. (1970). The assignment of numbers to rank order categories. American Sociological Review, 35(3), 515 524. McKeachie, W. J. (1997). Student ratings: The validity of use. American Psychologist, 52, 1218 1225. O Sullivan, E. & Rassel, G.R. (1989). Research methods for public administrators. White Plains, NY: Longman. Otani, K., Kurz, R. S., & Harris, L. E. (2005). Managing primary care using patient satisfaction measures. Journal of Healthcare Management, 50(5), 311 325. Pike, C. K. (1998). A validity study of an instrument designed to measure teaching effectiveness. Journal of Social Work Education, 34, 261 271. Whittington, L. A. (2001). Detecting good teaching. Journal of Public Affairs Education, 7(5) 5 8. Wiers-Jenssen, J., Stensaker, B., & Grøgaard, J. B. (2002). Student satisfaction: Towards an empirical deconstruction of the concept. Quality in Higher Education, 8(2), 183 195. Wolfer, T., & Johnson, M.M. (2003). Reevaluating student evaluation of teaching: The teaching evaluation form. Journal of Social Work Education, 39(1), 111 121. Journal of Public Affairs Education 543

K. Otani, B. Joon Kim, & J. Cho Koichiro Otani is an associate professor in the Department of Public Policy at Indiana University Purdue University Fort Wayne. He received a PhD in Health Services Research from Saint Louis University. B. Joon Kim is an assistant professor in the Department of Public Policy at Indiana University Purdue University Fort Wayne. He received a PhD in Public Administration and Policy from Virginia Polytechnic Institute and State University. Jeong-IL Cho is an assistant professor in the Department of Professional Studies at Indiana University Purdue University Fort Wayne. She received a PhD in Teaching and Learning from the University of Iowa. 544 Journal of Public Affairs Education