Online Student Evaluation of Teaching for Distance Education: A Perfect Match?



Similar documents
The Commission on Colleges, Southern Association of Colleges and Schools and The IDEA Student Ratings of Instruction System

Online Collection of Midterm Student Feedback

An Analysis of how Proctoring Exams in Online Mathematics Offerings Affects Student Learning and Course Integrity

Student Response to Instruction (SRTI)

University of Missouri-Columbia. MU Sinclair School of Nursing. GUIDELINES for APPOINTMENT, REAPPOINTMENT, and PROMOTION of NON-REGULAR FACULTY

Chapter 14: Request for Institutional Change

V. Course Evaluation and Revision

College of Health Professions Dean s Review of Undergraduate Programs August 15, 2011

M. CRITERIA FOR FACULTY EVALUATION

Creating Quality Developmental Education

Implementing Guided Pathways at Miami Dade College: A Case Study

Guidelines for Massachusetts Early Educator Preparation Programs Participating in the Early Childhood Educator Scholarships Program.

Criterion two. ASQ Higher Education Brief December 2011 (Vol. 4, No. 6)

THE UIS MODEL FOR ONLINE SUCCESS

THE UTILIZATION OF COURSE MANAGEMENT SYSTEMS IN BUSINESS SCHOOLS: SOME RECENT EVIDENCE

University of Delaware College of Health Sciences Department of Behavioral Health and Nutrition

Community-Based Program Review at Indiana University-Purdue University Indianapolis (IUPUI) Trudy W. Banta

Instructions for Faculty Evaluation System: Full-time Continuing Faculty

Student Perceptions of Online Learning: A Comparison of Two Different Populations

Strategic Plan San Luis Obispo County Community College District

Report on the Florida State University School of Nursing Seminar on Excellence in Nurse Education

INTRODUCTION AND SUGGESTIONS FOR VITA PREPARATION COLLEGE OF SCIENCES AND HUMANITIES

Proposal to Change an Academic Program. Bachelor of Manufacturing Management for Online Delivery

Baker College - Master of Business Administration Program Assessment Report CGS Assessment Report: MBA Program

GRADUATE PROGRAM REVIEW POLICY. Texas Southern University

A Collaborative Professional Development Approach to Improving Student Outcomes

Proposal to Change an Academic Program. Bachelor of Science in Business for Online Delivery

CHAPTER 1 INTRODUCTION. stakeholders, and a subject that goes beyond the world of researchers, given its social,

TROY UNIVERSITY Sorrell College of Business Response to ACBSP Evaluation Feedback Report November 11, 2008

MCNY DL Case-Study: Paradigm-Shift in Business Program s Approach to Serving Predominantly Black Minority Students

GUIDELINES FOR CONDUCTING EVALUATIONS OF ACADEMIC PROGRAMS

Rio Salado College Assessment of Student Learning Executive Summary: Annual Report

COURSE COORDINATION: A MAINSTAY IN ACCOUNTING EDUCATION

2013 Review of the MS in Sport Management

QUESTIONING THE HYBRID MODEL: STUDENT OUTCOMES IN DIFFERENT COURSE FORMATS

Online Instruction

SPEA CRIMINAL JUSTICE 3 RD YEAR REPORT MAY 2008

The Organization: Leadership, Resources, Faculty and Staff Worksheet

Performance Management for Faculty and Staff

ACADEMIC PROGRAM REVIEW GUIDELINES

Department of Child & Family Development Promotion and Tenure Guidelines November 2004

Policies for Evaluating Faculty: Recommendations for Incorporating Student and Peer Reviews in the Faculty Evaluation Process DRAFT

D R A F T. Faculty Senate Ad Hoc Committee on Quality in Online Learning.

SACS REQUIREMENTS AND STANDARDS. Core Requirements

DISTANCE EDUCATION STANDARDS AND GUIDELINES

Online Course Delivery - A Santa Barbara City College Perspective

INNOVATION WITH IMPACT. Creating a Culture for Scholarly and Systematic Innovation in Engineering Education

CTL 2009 ADVANCED PROGRAM REPORT

Journal of Technology Education Vol. 13 No.1, Fall Comparing Computer Usage by Students in Education Programs to Technology Education Majors

Definitions SECTION ACADEMIC AFFAIRS 333 DISTANCE LEARNING

Date Submitted: October 1, Unit: The Office of Institutional Research, Planning, and Effectiveness I. MISSION STATEMENT

THE UTILIZATION OF PART-TIME FACULTY IN BUSINESS SCHOOLS: ARE THEY DIFFERENT?

IUPUI Program Review and Assessment Committee Assessment Project Proposal. Cover Sheet. Assistant Clinical Professor Assistant Clinical Professor

The MetLife Survey of

DRAFT (February 7, 2000) Bert Garza. Faculty and Office for Computing and Information Science: Administrative and Management Structure

Associate Dean, Graduate Academic & Faculty Affairs College of Professional Studies Boston, MA

California State University, Stanislaus Doctor of Education (Ed.D.), Educational Leadership Assessment Plan

Annual Goals for Math & Computer Science

Academic Unit Action Plan: SOCIAL WORK

An Evaluation of Student Satisfaction With Distance Learning Courses. Background Information

Student Success at the University of South Carolina: A comprehensive approach Category: Academic Support

Continuing Education Credit and Special Learning Activities (2.G., 2.H)

Principal Appraisal Overview

Department of History Policy 1.1. Faculty Evaluation. Evaluation Procedures

Charting the Future Bachelor of Applied Science in Public Service. I. The Past: Reflecting on our Heritage

One Faculty Serving All Students. An Issue Brief by the Coalition on the Academic Workforce

Commission on Colleges Southern Association of Colleges and Schools. Best Practices For Electronically Offered Degree and Certificate Programs

The University of Arizona

GUIDELINES FOR GRADUATE COURSES OFFERED FOR PROFESSIONAL DEVELOPMENT OF SCHOOL PERSONNEL

EVALUATION OF DEPARTMENT CHAIR

GUIDELINES FOR ACADEMIC PROGRAM REVIEW For self-studies due to the Office of the Provost on October 1, 2015 GRADUATE PROGRAMS

The University of West Florida. Department of Criminology & Criminal Justice. Bylaws

Frequently Asked Questions

Asynchronous Learning Networks in Higher Education: A Review of the Literature on Community, Collaboration & Learning. Jennifer Scagnelli

ASSOCIATION FOR GENERAL AND LIBERAL STUDIES 2008 AGLS Awards for Improving General Education: Effective Program Processes

POLICIES CONCERNING ADJUNCT FACULTY

NORTH CAROLINA AGRICULTURAL AND TECHNICAL STATE UNIVERSITY

Rules of Organization and Bylaws Gladys A. Kelce College of Business

Fall Summer W e s t e r n C o n n e c t i c u t S t a t e U n i v e r s i t y

Program Assessment: School of Criminology and Criminal Justice.

DEPARTMENT OF MARKETING COLLEGE OF BUSINESS ADMINSTRATION POLICY ON REAPPOINTMENT, TENURE, AND PROMOTION (RTP)

Effective Online Instructional Competencies as Perceived by Online University Faculty and Students: A Sequel Study

North Dakota State University College of Pharmacy, Nursing, and Allied Sciences Strategic Plan

COLLEGE OF ARTS AND LETTERS ORGANIZATION AND PROCEDURES

Kurt F. Geisinger. Changes in Outcomes Assessment for Accreditation: The Case of Teacher Education in the United States

Examining the Role of Online Courses in Native Hawaiian Culture and Language at the University of Hawaii

THE GRADUATE SCHOOL CREIGHTON UNIVERSITY. By-Laws. ARTICLE I Definitions

Evolving expectations for teaching in higher education in Canada

NEW UNITS OF INSTRUCTION, PUBLIC SERVICE, AND RESEARCH AT PUBLIC UNIVERSITIES

EXECUTIVE SUMMARY. List all of the program s learning outcomes: (regardless of whether or not they are being assessed this year)

Bert C. Bach PROFILE :

Strategic Planning for the Alabama Community College System

Indiana STEM Teacher Recruitment Fund Grant Proposal

Research Project. Abstracts AAP ACADEMY ADMINISTRATION PRACTICE. Growth Starts With Knowledge

Julian R. Cowart EDUCATION

AC : ART2STEM: BUILDING A STEM WORKFORCE AT THE MIDDLE SCHOOL LEVEL

Summary of Critical Success Factors, Action Items and Performance Measures

Textbook Steering Committee Report. April, 2009

IU Southeast Program Development Guidelines: New Degree Online

Internet Learning. Volume 1, Number 1 Fall 2012 Article Policy Studies Organization

Transcription:

Online Student Evaluation of Teaching for Distance Education: A Perfect Match? Charles F. Harrington, University of Southern Indiana Saxon G. Reasons, University of Southern Indiana Abstract Given the limited number of currently available resources, one public mid-western university is working to develop and implement an effective and appropriate means for online student evaluation of teaching in distance education courses that is useful and beneficial for all stakeholders student, faculty, and administration. Because traditional methods have proven insufficient in addressing the breadth of instructional delivery and course design methods, the university began a pilot project supporting online evaluation for distance education courses in spring 2003. This paper outlines the issues considered and challenges faced in the search to employ reliable and valid online student evaluation for distance education courses, to maximize the benefits associated with electronic data gathering and reporting, and to meet the organizational and logistical challenges inherent in such an endeavor. The Journal of Educators Online, Volume 2, Number 1, January 2005 1

Introduction Distance education as a vehicle for the delivery of course content, as well as complete academic programs, continues to grow rapidly in American higher education. According to the recent NCES Report, Distance Education at Degree-Granting Postsecondary Institutions: 2000-2001 (Waits & Lewis, 2003), during the 12-month 2000-2001 academic year, 56% of all twoyear and four-year Title IV-eligible, degree-granting institutions offered distance education courses, with public institutions more likely to offer distance education courses than private institutions. During the 2000 2001 academic year, nearly 90% of public two-year and public four-year institutions offered distance education courses, compared with 16% of private two-year and more than 40% of private four-year institutions. With these statistics in mind, several pressing questions arise. How are institutions measuring and ensuring the quality of their distance education curricula? How thoroughly is teaching in distance education courses being evaluated? Are standardized evaluations of teaching forms being used, and if so, with what result? Are students completing the same evaluation forms as they would for on-campus courses? Who is responsible for developing the forms, and have the forms been evaluated for statistical reliability and validity? Who oversees survey and instrument administration, data collection processes, and dissemination of results? For many institutions, the answers to these questions are not obvious. Just as distance education frequently makes use of existing and emerging technologies, especially the Internet, for delivery of curriculum, another trend in higher education s leveraging of technology is the increasing use of the Internet in facilitating student evaluation of teaching, for both distance education and campus-based courses (Achtemeier, Morris, & Finnegan, 2003). Online student evaluation of teaching is a fairly recent topic of scholarly interest, with practitioners collaborating to build a foundation of best principles and practices for other institutions considering the migration from paper-and-pencil evaluation and assessment. Sorenson and Reiner (2003) provide a helpful overview of the online student evaluation trend, including references to its pervasiveness in higher education. The authors outline a number of benefits, such as time savings, flexibility in design and reporting of data, increased quantity and quality of student responses, and lowered instructional and support costs. Additionally, the The Journal of Educators Online, Volume 2, Number 1, January 2005 2

authors highlight a number of logistical and organizational challenges to online evaluation systems and make recommendations for meeting these challenges. In the same vein, our public mid-western university is currently laboring to develop an effective and appropriate means for online student evaluation of teaching that is useful and beneficial for all stakeholders student, faculty, and administration. Because traditional methods (paper-andpencil assessments) have proven insufficient in addressing the breadth of instructional delivery and course design methods, our university began a pilot project supporting online evaluation for distance education courses during the spring of 2003. Outlined in this paper are the more pressing issues considered and challenges faced in our attempt to develop and employ reliable and valid online student evaluation for distance education courses, to maximize the benefits associated with electronic data gathering and reporting, and to meet the organizational and logistical challenges along the way. Student Evaluation of Teaching, Distance Education, and Online Evaluation Systems There is no shortage of research on the subject of student evaluation of teaching (SET). A casual search of the ERIC database for student evaluation of teaching performance reveals over 5,000 citations (Cashin (1999) found over 2,000). Research examining instrument development, validity, and reliability has been conducted by a number of notable scholars on these issues (Arreola, 1995; Aleamoni, 1987; Feldman, 1978). Theall and Franklin (1990) reviewed voluminous studies that date as far back as the late 1920 s. Not surprisingly, colleges and universities differ in their approaches to the student evaluation of teaching in distance education courses. In his study of institutions offering telecourses, Compora (2003) found that institutions consider evaluation of teaching important and that many had mechanisms in place for conducting evaluations; however, methods and success varied among institutions. In some cases, faculty members were responsible for conducting evaluations; in other cases, the distance education program director coordinated instrument administration. An evaluation of teaching via distance education was not found to be mandatory at all of the institutions studied. The author further proposed that at minimum, summative The Journal of Educators Online, Volume 2, Number 1, January 2005 3

evaluations of individual distance education courses be conducted every term, and that preferably both formative and summative evaluations are conducted. Consistent with this sage advice, our university is currently attempting to develop a consistent, effective, and appropriate means for student evaluation of teaching in distance education courses, which recognizes the unique necessity of combining both formative and summative measures. With an enrollment approaching 10,000, our university offers three undergraduate degree programs, three graduate degree programs, and more than 150 courses via distance education. Although these courses are offered in a variety of academic disciplines, complete degree programs are offered only through the School of Nursing and Health Professions. Instructional delivery is primarily Internet-based, with course enrollments increased steadily since 1996, when the distance education programs began. This past spring semester, 145 faculty members taught 177 distance education course sections generating 2586 course enrollments. Historically, our end-of-term course evaluations have been handled through the same cumbersome paper-and-pencil method used for on-campus courses. In spring 2003, we began a pilot project to examine the feasibility of moving student evaluation of teaching into an online environment. Goals for the project included: Utilizing the Internet to administer and collect student evaluation data; Decreasing the time necessary for administration, data collection, analysis, and report dissemination; Providing faculty teaching distance education course with course evaluation items more suitable to their pedagogical and course design needs; Providing students with an alternative to paper-and-pencil student evaluation of teaching form completion; and Facilitating the analysis of SET results across disciplines and over time. In theory, a natural fit would seem to exist with distance education courses and student evaluation of teaching facilitated online. In practice, a great deal of effort is needed to make this work. Tasks such as coordinating the work of a project steering committee, establishing and testing the evaluation items, determining access permissions to the evaluation data, assigning The Journal of Educators Online, Volume 2, Number 1, January 2005 4

responsibility for process details, uploading course enrollment date into survey software, providing technical support for students and faculty, and collecting and reporting the findings require cooperation and individual dedication to the project. Addressing Reliability and Validity of SET Instruments The assurance of validity and reliability in a teaching evaluation instrument is of obvious importance, whether for on-campus or distance education courses. Despite decades of extensive research, questions persist related to validity and reliability. A prior study conducted by Harrington and Schibik (2003) produced national baseline data concerning evaluation forms used by students to evaluate teaching in on-campus courses. Specific research questions addressed the origin of forms, frequency of forms utilization, and the evidence of the reliability and validity of forms. The authors found that nearly 90% of all institutions surveyed indicated that student evaluation of teaching were required of all faculty, both full- and part-time. Hence, SETs remain the primary source of information available to those involved with the evaluation of faculty teaching performance. More than two-thirds (69%) of all institutions indicated that their campuses had adopted a single, campus-wide student evaluation form with a common core of evaluation items. Many of these respondents indicated that the common evaluation form was supplemented with a school and/or departmental specific SET instrument. The mean number of separate SETs used within the university during any given term was 7.5. Perhaps most significantly, despite the availability of commercially available evaluation tools, more than 80% of institutions using a supplemental form had designed their own student evaluation of teaching survey instrument. Most of these supplemental and home grown evaluation forms were developed to address faculty preferences. The utilization of user-created forms is not a major cause for concern. However, the fact that only approximately 9% of survey respondents (30 of 334) had conducted studies on the reliability and/or validity of their own student evaluation of teaching instruments presents a significant dilemma. To avoid this problem with our endeavor, a project steering committee comprised of faculty, academic administrators, and distance education support professionals was convened to facilitate The Journal of Educators Online, Volume 2, Number 1, January 2005 5

the development and testing of potential assessment and evaluation items. Faculty with expertise in statistical analysis, particularly psychometrics, tested suggested items for reliability and validity. Statistical analyses were run to calculate inter-item correlations, covariance, and Cronbach alphas, and the Tukey test for additivity for all items. Additionally, survey items were field-tested with student focus groups to ascertain appropriateness and readability. Increasing Consistency of Evaluation of Distance Education Courses As faculty and students collectively define appropriate means of evaluation for distance education courses, questions about consistency will likely arise. What purposes do student evaluations of teaching serve? Can and should all distance education courses, regardless of academic discipline, utilize the same evaluation instrument? Who should have access to the findings - individual faculty, department chairpersons, program directors, or academic deans? Regarding distance education evaluation, Cheung (1998) suggested the following best principles relative to teaching evaluation: Provide diagnostic feedback for improving the academic quality of the course, Provide information for personnel and administrative decision making, Allow students to express their needs and views formally and systematically, Advance research on distance teaching, Monitor the quality of distance teaching for the sake of accountability, Collect information for accreditation purposes, Provide comparative data across different courses so as to monitor the consistency of standards, and Facilitate staff development. Cheung further stressed the importance of using evaluation tools with solid psychometric properties. Facing the Challenges of Distance Education Course Evaluation As those who teach or administer distance education courses and programs can attest, there are many inherent challenges to evaluating distance education courses, some of which include the facts that: The Journal of Educators Online, Volume 2, Number 1, January 2005 6

Adjunct instructors frequently teach distance education courses. What is the institution s policy toward SET for part-time faculty? Is this policy the same for on-campus and distance education courses? Distance education instructors frequently employ teaching strategies mediated by technology that serve to bridge the separation of student and instructor. Are instructors properly prepared to teach at a distance so that favorable SETs are within reach? Student evaluations typically figure into the promotion and tenure process. Has the academic dean or department chairperson specified evaluation expectations and rewards for faculty teaching distance education courses? Distance education courses are often the product of a team of professionals, including instructional designers, producer/directors, and technical specialists, working alongside faculty. The technology used in instruction is integral to the teaching and learning processes; therefore, it is important to assess the technology s effectiveness alongside the faculty member s teaching effectiveness. Because these two elements are intertwined, it is understandable that the evaluation process can become very complex. How can evaluation forms be constructed so that all interested parties get the information that they need without making the process excessively burdensome for the student? Institutions offering distance education courses are likely to experience these common challenges. However, it is possible to overcome them with careful planning, collaboration, creativity, and individual dedication to the project. Inevitably, the principle question that institutions must answer for themselves is, how can we develop and implement a distance education course and instructor evaluation system that supports a campus culture of assessment? Collaboration is critical to project success. Faculty must be involved in all aspects of project development, such as item selection and testing, deployment and data collection strategies, and the presentation of results. Faculty evaluation is a topic fraught with potential peril and one that should not be taken lightly or shortsightedly. Involving campus assessment professionals in discussions relative to the formative assessment items strengthens the project further. The Journal of Educators Online, Volume 2, Number 1, January 2005 7

Maximizing Benefits and Meeting Challenges While the perceived benefits of migrating to an online evaluation system, such as time savings, flexibility in design and reporting of data, increased quantity and quality of student responses, and lowered costs, may seem obvious to some, proposing a change in an institution s process for student evaluation of teaching has far-reaching implications and can elicit strong, and potentially negative, opinion (Sorenson & Reiner, 2003). Even a scaled-down pilot project, in our case limited to distance education courses, requires widespread buy-in and long-term planning. Before committing to an expensive and long-term contract for course evaluation software, we wanted to focus attention on the creation of a culture of assessment. To this end, members of the university community, including institutional research professionals, academic affairs administrators, deans, department chairs, faculty, and instructional technology support staff, came together in order to discuss a series of fundamental questions: Why are we doing this in the first place? What do we need to measure? What would we like to measure? What do the results need to look like? How will they be used? Once the project steering committee had identified a point of departure, the pilot project was launched in spring 2003. The following is a general outline of project logistics: 1. An inexpensive Internet-based survey software was licensed on a monthly basis to collect, store, and report SET data. 2. Evaluation items were developed and tested for validity and reliability. A sample of student evaluation of teaching items included: Formative Items: I found course objectives and assignments to be clearly stated and easily understood. I felt connected to the instructor and other students in this course. The instructor used a number of teaching techniques to involve me in learning. The way in which this course was taught required me to think in new and different ways. The Journal of Educators Online, Volume 2, Number 1, January 2005 8

The instructor was able to clearly explain the relationships among the various course topics. The instructor makes difficult course material understandable. Parts of this course were designed to make difficult material thoroughly understandable. Summative Items: In this course, the instructor's teaching required me to do my best work. In this course, the instructor is able to explain difficult materials in ways that I can understand. In this course, the assignments given in class were challenging. This instructor is one of the best I've had at this University. Of all the courses I've taken at this university, this course if one of the best I've had. 3. A maximum of 40 items per evaluation was established, comprised of 20 mandatory items that all participating distance education faculty would use plus a bank of 50 optional items from which faculty could select an additional 20. These items included both summative and formative evaluation questions. 4. The timeline of events was established, with particular attention given to the window of opportunity for students to respond. 5. Distance education faculty were invited via email to participate in the online evaluation project as an option to the paper-and-pencil method. 6. Participating course rosters and student email information were merged into spreadsheets and loaded into the software, which then generated customized email messages containing student-specific Internet links to the appropriate evaluation instrument for their course. 7. Technical support for both faculty and students was provided as needed. 8. Participating faculty were sent periodic progress reports. 9. SET data were gathered by the Office of Academic Affairs and reported to deans, department chairs, and faculty. The Journal of Educators Online, Volume 2, Number 1, January 2005 9

10. The members of the project steering committee met to debrief and plan the next phase in the evaluation project. The following is descriptive information concerning the online evaluation project. Faculty teaching classes via distance education were provided an opportunity to participate, but were not mandated to do so. Table 1. Online Evaluation Project Descriptive Information Faculty Course Student Response Rate Participants Sections Respondents Spring 2003 11 20 202/332 60.8% Fall 2003 35 96 1129/1791 63.0% Spring 2004 33 77 947/1461 64.8% (Note: Fewer overall course sections are offered in spring terms compared to fall.) The student response rate has increased steadily each term. As faculty and students become more comfortable with this new approach to SET, we anticipate a continued increase in the response rates. It is also posited that faculty will begin to use the SET item bank for formative classroom assessment. Perhaps the greatest single benefit from this project involves the ability to separate summative from formative assessment. Historically, the two have been combined. Using the paper-and-pencil method, potentially helpful formative assessment data were not available to faculty until the release of the summative results, typically weeks following the close the academic semester and far too late to make substantive changes to the course. By migrating to an Internet-based evaluation tool, we gain the ability to share with faculty the formative assessment results in a timely and helpful fashion, without compromising student confidentiality in the development of the summative evaluation data. Our university has benefited significantly from an introduction to online student evaluation of teaching in distance education courses. As this project evolves and matures, we look forward to The Journal of Educators Online, Volume 2, Number 1, January 2005 10

an increased ability to examine the growing data set in a multitude of ways. Because the data are electronic, longitudinal analysis of course-specific or instructor-specific data is more easily conducted. Similarly, data are readily available, and data reporting is faster, more flexible, and customizable. Conclusion and Future Directions University faculty and administrators realize that student evaluation of teaching is critical for the continued improvement and success of distance education courses. We are giving careful attention to this issue by the development of a new and improved evaluation system. Prior studies on this topic have examined institutional evaluation practices and instruments used for student evaluation of teaching in distance education courses; however, opportunities exists for institutions to examine their own practices and compare them with benchmark institutions, whether statewide, nationally, or both. Higher education institutions have a responsibility to their students for ensuring effective and quality teaching practices, and this can be a complex and difficult matter for distance education courses. Our project will continue to provide attention and focus in an attempt to exploit this Internetbased assessment and evaluation approach as a way to improve curricular and pedagogical quality. Embedding classroom assessment items in the formative survey items, helping faculty discover ways in which they can capitalize on the results of these assessments to make immediate and substantive improvements in their teaching, developing a longitudinal assessment and evaluation database, and increasing faculty and student satisfaction and comfort with this new evaluation approach will dominate our project agenda in the near term. The Journal of Educators Online, Volume 2, Number 1, January 2005 11

References Aleamoni, L. (1987). Student rating: Myths versus research facts. Journal of Personnel Evaluation in Education, 1, 111-119. Arreola, R. (1995). Developing a comprehensive faculty evaluation system. Bolton, Massachusetts: Anker Publishing Company, Inc. Achtemeier, S., Morris, L., & Finnegan, C. (2003). Considerations for developing evaluations on online courses. Journal of Asynchronous Learning Networks 7(1). Retrieved August 22, 2003, from http://sloan-c.org/publications/jaln/v7n1/v7n1_achtemeier.asp Cashin, W. (1999). Student Rating of Teaching: Uses and Misuses. In P. Seldin (Ed.), Changing Practices in Evaluating Teaching: A Practical Guide to Faculty Performance and Promotion/Tenure Decisions (pp. 25-44). Boston: Anker Publishing Company, Inc. Cheung, D. (1998). Developing a student evaluation instrument for distance teaching. Distance Education 19(1). Retrieved August 22, 2003, from ProQuest database. Compora, D. (2003). Current Trends in Distance Education: An Administrative Model. Online Journal of Distance Learning Administration 6(2). Retrieved August 22, 2003, from http://www.westga.edu/%7edistance/ojdla/summer62/compora62.html Feldman, K. (1978). Course characteristics and college students ratings of their teachers: What we know and what we don t. Research in Higher Education, 9, 199-242. Harrington, C., & Schibik, T. (2003, April). Student evaluation of teaching: What every institutional researcher should know. Paper presented at the 17 th Annual Meeting of the Indiana Association for Institutional Research, Nashville, IN. Sorenson, L., & Reiner, C. (2003). Charting the Uncharted Seas of Online Student Ratings of Instruction. New Directions for Teaching and Learning, Winter2003 Issue 96, 1-24. Theall, M. & Franklin, J. (Eds.) (1990). Student ratings of instruction: Issues for improving practice. New Directions for Teaching and Learning, Fall1990 Issue 43. Waits, T. and Lewis, L. (2003). NCES Report, Distance Education at Degree-Granting Post- Secondary Institutions; 2000-2001. Retrieved July 21, 2003, from http://nces.ed.gov/pubsearch/pubsinfo.asp?pubid=2003017 The Journal of Educators Online, Volume 2, Number 1, January 2005 12