EVALUATION TERMINOLOGY TIP SHEET
|
|
|
- Shawn Powell
- 9 years ago
- Views:
Transcription
1 EVALUATION TERMINOLOGY TIP SHEET Archival Data: Data from existing written or computer records; data that is collected, stored and updated periodically by certain public agencies, such as schools, social services and public health agencies. This data can be crossreferenced in various combinations to identify individuals, groups and geographic areas. Attrition: Loss of subjects from the defined sample during the course of the program; unplanned reduction in the size of the study sample because of participants dropping out of the program, for example because of relocation. Baseline Data: The initial information collected about the condition or performance of subjects prior to the implementation of an intervention, against which progress can be compared at strategic points during and at completion of the program. Benchmark: Reference point or standard against which performance or achievements can be assessed. A benchmark may refer to the performance that has been achieved in the recent past by other comparable organizations. Bias: A point of view that inhibits objectivity. Case studies: An intensive, detailed description and analysis of a single participant, intervention site or program in the context of a particular situation or condition. A case study can be the story of one person s experience with the program. A case study does not attempt to represent every participant s exact experience. Coding: To translate a given set of data or items into descriptive or analytic categories to be used for data labeling and retrieval. Continuous Quality Improvement (CQI): The systematic assessment and feedback of information about planning, implementation and outcomes; may include client satisfaction and utility of the program for participants. Also known as formative evaluation. EVALUATION TERMINOLOGY 5-17
2 CSAP s Core Measures: A compendium of data collection instruments that measure those underlying conditions risks, assets, attitudes and behaviors of different populations related to the prevention and/or reduction of substance abuse. (CSAP is the Center for Substance Abuse Prevention.) Data Driven: A process whereby decisions are informed by and tested against systematically gathered and analyzed information. Data Collection Instruments (evaluation instruments): Specially designed data collection tools (e.g., questionnaires, surveys, tests, observation guides) to obtain measurably reliable responses from individuals or groups pertaining to their attitudes, abilities, beliefs or behaviors. The tools used to collect information. Feasibility: Refers to practical and logistic concerns related to administering the evaluation instrument. (e.g. Is it reasonably suited to your time constraints, staffing, and/or budgetary resources?) Focus group: A group selected for its relevance to an evaluation that is engaged by a trained facilitator in a series of guided discussions designed for sharing insights, ideas, and observations on a topic of concern or interest. Formative evaluation: Information collected for a specific period of time, often during the start-up or pilot phase of a project, to refine and improve implementation and to assist in solving unanticipated problems. This can include monitoring efforts to provide ongoing feedback about the quality of the program. Also known as Continuous Quality Improvement. Goal: A measurable statement of desired longer-term, global impact of the prevention program that supports the vision and mission statements. Immediate outcome: The initial change in a sequence of changes expected to occur as a result of implementation of a science-based program. Impact evaluation: A type of outcome evaluation that focuses on the broad, long-term impacts or results of program activities; effects on the conditions described in baseline data. Incidence: The number of new cases/individuals within a specified population who have initiated a behavior in this case drug, alcohol, or tobacco use during a specific period of time. This identifies new users to be compared to the number of new users historically, over comparable periods of time, usually expressed as a rate. EVALUATION TERMINOLOGY 5-18
3 Indicator: A measure selected as a marker of whether the program was successful in achieving its desired results. Can be a substitute measure for a concept that is not directly observable or measurable, e.g. prejudice, selfefficacy, substance abuse. Identifying indicators to measure helps a program more clearly define its outcomes. Informed consent: The written permission obtained from research participants (or their parents if participants are minors), giving their consent to participate in an evaluation after having been informed of the nature of the research. Intermediate Outcome: The changes that are measured subsequent to immediate change, but prior to the final changes that are measured at program completion in a sequence of changes expected to occur in a science-based program, Interviews: Guided conversations between a skilled interviewer and an interviewee that seek to maximize opportunities for the expression of a respondent s feelings and ideas through the use of open-ended questions and a loosely structured interview guide. Instrument: Device that assists evaluators in collecting data in an organized fashion, such as a standardized survey or interview protocol; a data collection device. Key informant: Person with background, knowledge, or special skills relevant to topics examined by the evaluation; sometimes an informal leader or spokesperson in the targeted population. Logic Model: A graphic depiction of how the components of a program link together and lead to the achievement of program goals/outcomes. A logic model is a succinct, logical series of statements linking the needs and resources of your community to strategies and activities that address the issues and define the expected results/outcomes. Longitudinal study: A study in which a particular individual or group of individuals is followed over a substantial period of time to discover changes that may be attributable to the influences of an intervention, maturation, or the environment. EVALUATION TERMINOLOGY 5-19
4 Mean: The measure of central tendency the sum of all the scores added and then divided by the total number of scores; the average. The mean is sensitive to extreme scores, which can cause the mean to be less representative of the typical average. Median: A measure of central tendency referring to the point exactly midway between the top and bottom halves of a distribution of values; the number or score that divides a list of numbers in half. Mixed-method evaluation: An evaluation design that includes the use of both quantitative and qualitative methods for data collection and data analysis. Mode: A measure of central tendency referring to the value most often given by respondents; the number or score that occurs most often. Objective: A specific, measurable description of an intended outcome. Objectivity: The expectation that data collection, analysis and interpretation will adhere to standards that eliminate or minimize bias; objectivity insures that outcome or evaluation results are free from the influence of personal preferences or loyalties. Observation: Data collection method involving unobtrusive examination of behaviors and/or occurrences, often in a natural setting, and characterized by no interaction between participants and observers. Data is obtained by simply watching what people do and documenting it in a systematic manner. Outcomes: Changes in targeted attitudes, values, behaviors or conditions between baseline measurement and subsequent points of measurement. Changes can be immediate, intermediate or long-term; the results/effects expected by implementing the program s strategies. Outcome evaluation: Examines the results of a program s efforts at various points in time during and after implementation of the program s activities. It seeks to answer the question, what difference did the program make? Pre/post tests: Standardized methods used to assess change in subjects knowledge and capacity to apply this knowledge to new situations. The tests are administered prior to implementation of the program and after completion of the program s activities. Determines performance prior to and after the delivery of an activity or strategy. EVALUATION TERMINOLOGY 5-20
5 Participatory evaluation: The process of engaging stakeholders in an evaluation effort. Stakeholders are those people most invested in the success of a program, including staff, board members, volunteers, other agencies, funding agencies and program participants. Getting input from the stakeholders at all stages of the evaluation effort from deciding what questions to ask, to collecting data, to analyzing data and presenting results. Prevalence: The number of all new and old cases of a disease or occurrences of an event during a specified time, in relation to the size of the population from which they are drawn. (Numbers of people using or abusing substances during a specified period.) Prevalence is usually expressed as a rate, such as the number of cases per 100,000 people. Process evaluation: Addresses questions related to how a program is implemented. Compares what was supposed to happen with what actually happened. Includes such information as dosage and frequency, staffing, participation and other factors related to implementation. Documents what was actually done, how much, when, for whom, and by whom during the course of the project inputs. Qualitative data: Non-numerical data rich in detail and description that are usually presented in a textual or narrative format, such as data from case studies, focus groups, interviews or document reviews. Used with open-ended questions, this data has the ability to illuminate evaluation findings derived from quantitative methods. Quantitative data: Numeric information, focusing on things that can be counted, scored and categorized; used with close-ended questions, where participants have a limited set of possible answers to a question. Quantitative data analysis utilizes statistical methods. Questionnaire: Highly structured series of written questions that is administered to a large number of people; questions have a limited set of possible responses. Random assignment: A process by which the people in a sample to be tested are chose at random from a larger population; a pool of eligible evaluation participants are selected on a random basis. EVALUATION TERMINOLOGY 5-21
6 Reliability: The consistency of a measurement or measurement instrument over time. Consistent results over time with similar populations and under similar conditions confirm the reliability of the measure. (Can the test produce reliable results each time it is used and in different locations or with different populations?) Representative sample: A segment or group taken from a larger body or population that mirrors in composition the characteristics of the larger body or population. Sample: A segment of a larger body or population. Standardized tests: Tests or surveys that have standard printed forms and content with standardized instructions for administration, use, scoring, and interpretation. Subjectivity: Subjectivity exists when the phenomena of interest are described or interpreted in personal terms related to one s attitudes, beliefs or opinions. Surveys: Uses structured questions from specially designed instruments to collect data about the feelings, attitudes and/or behaviors of individuals. Targets: Defines who or what and where you expect to change as a result of your efforts. Theory of change: A set of assumptions about how and why desired change is most likely to occur as a result of your program, based on past research or existing theories of behavior and development. Defines the evidenced-based strategies or approaches proven to address a particular problem. Forms the basis for logic model planning. Threshold: Defines how much change you can realistically expect to see over a specific period of time, as a result of your program strategies. Validity: The extent to which a measure of a particular construct/concept actually measures what it purports to measure; how well a test actually measures what it is supposed to measure. EVALUATION TERMINOLOGY 5-22
EVALUATION METHODS TIP SHEET
EVALUATION METHODS TIP SHEET QUANTITATIVE METHODS: Quantitative data collection methods consist of counts or frequencies, rates or percentages, or other statistics that document the actual existence or
Evaluation: Designs and Approaches
Evaluation: Designs and Approaches Publication Year: 2004 The choice of a design for an outcome evaluation is often influenced by the need to compromise between cost and certainty. Generally, the more
Analyzing and interpreting data Evaluation resources from Wilder Research
Wilder Research Analyzing and interpreting data Evaluation resources from Wilder Research Once data are collected, the next step is to analyze the data. A plan for analyzing your data should be developed
IPDET Module 6: Descriptive, Normative, and Impact Evaluation Designs
IPDET Module 6: Descriptive, Normative, and Impact Evaluation Designs Intervention or Policy Evaluation Questions Design Questions Elements Types Key Points Introduction What Is Evaluation Design? Connecting
Developing an implementation research proposal. Session 2: Research design
Developing an implementation research proposal Session 2: Research design Learning objectives After completing this session, you will be able to: Develop a research design outlining your data collection
PROJECT MANAGEMENT TRAINING MODULES
PROJECT MANAGEMENT TRAINING MODULES KENYA PROJECTS ORGANIZATION < Projects Solutions Provider > Macjo Arcade, 4 th Flr., Suite 15E P.O Box, 3029 00200, Nairobi - Kenya Tel: 254 (0)202319748 Mob: 0725 788
Comparison of Research Designs Template
Comparison of Comparison of The following seven tables provide an annotated template to guide you through the comparison of research designs assignment in this course. These tables help you organize your
Evaluation Plan: Process Evaluation for Hypothetical AmeriCorps Program
Evaluation Plan: Process Evaluation for Hypothetical AmeriCorps Program Introduction: This evaluation plan describes a process evaluation for Financial Empowerment Corps (FEC), an AmeriCorps State and
Prepared by the Policy, Performance and Quality Assurance Unit (Adults) Tamsin White
Principles of Good Research & Research Proposal Guidee Prepared by the Policy, Performance and Quality Assurance Unit (Adults) Tamsin White March 2006 Principles of Good Research All research is different
TOOL D14 Monitoring and evaluation: a framework
TOOL D14 Monitoring and evaluation: a framework 159 TOOL D14 Monitoring and evaluation: a framework TOOL D14 For: About: Purpose: Use: Resource: Commissioners in primary care trusts (PCTs) and local authorities
GLOSSARY OF EVALUATION TERMS
Planning and Performance Management Unit Office of the Director of U.S. Foreign Assistance Final Version: March 25, 2009 INTRODUCTION This Glossary of Evaluation and Related Terms was jointly prepared
Measurement and measures. Professor Brian Oldenburg
Measurement and measures Professor Brian Oldenburg Learning objectives 1. To identify similarities/differences between qualitative & quantitative measures 2. To identify steps involved in choosing and/or
Single and Multiple-Case Study Designs IS493
1 2 Research Strategies Basic oppositions Survey research versus Case study quantitative versus qualitative The whole gamut Experiment Survey Archival analysis Historical research Case study 3 Basic Conditions
Job Grade: Band 5. Job Reference Number:
Job Title: Business Analyst Job Grade: Band 5 Directorate: Job Reference Number: People and Transformation P01216 The Role: This is a challenging role working with and across all services to support the
Chapter 3 Data Interpretation and Reporting Evaluation
Chapter 3 Data Interpretation and Reporting Evaluation Results This chapter explains how to interpret data and how to bring together evaluation results. Tips! - Evaluation does not end with data collection
HIGH SCHOOL MASS MEDIA AND MEDIA LITERACY STANDARDS
Guidelines for Syllabus Development of Mass Media Course (1084) DRAFT 1 of 7 HIGH SCHOOL MASS MEDIA AND MEDIA LITERACY STANDARDS Students study the importance of mass media as pervasive in modern life
FOREIGN AFFAIRS PROGRAM EVALUATION GLOSSARY CORE TERMS
Activity: A specific action or process undertaken over a specific period of time by an organization to convert resources to products or services to achieve results. Related term: Project. Appraisal: An
COM 365: INTRODUCTION TO COMMUNICATION RESEARCH METHODS Unit Test 3 Study Guide
COM 365: INTRODUCTION TO COMMUNICATION RESEARCH METHODS Unit Test 3 Study Guide The test will cover the introductory materials from Chapters 7, 8, 9, & 10 in the Frey, Botan, & Kreps text as well as any
Qualitative methods for effectiveness evaluation: When numbers are not enough
Chapter 7 Qualitative methods for effectiveness evaluation: When numbers are not enough 7.1 Introduction 7.2 Methods of collecting qualitative information 7.2.1 Interviews and focus groups 7.2.2 Questionnaires
Vision Statement for Innovative Software Development in a Large. Corporation
Vision Statement for Innovative Software Development in a Large Corporation Joseph C. Thomas Regent University Center for Leadership Studies LEAD606 Strategic Vision and Organizational Effectiveness 4-Nov-2002
2015 Michigan Department of Health and Human Services Adult Medicaid Health Plan CAHPS Report
2015 State of Michigan Department of Health and Human Services 2015 Michigan Department of Health and Human Services Adult Medicaid Health Plan CAHPS Report September 2015 Draft Draft 3133 East Camelback
Organization Effectiveness Based on Performance Measurement and Organizational Competencies Getting to Perfect!
Organization Effectiveness Based on Performance Measurement and Organizational Competencies Getting to Perfect! John McLaughlin, Human Capital Management Consultant Abstract This article describes a model
Writing Your PG Research Project Proposal
Writing Your PG Research Project Proposal Typically, most research project proposals will contain the following elements: The proposed title of your research project An outline of the scope and rationale
Market Research Methodology
Market Research Methodology JANUARY 12, 2008 MARKET RESEARCH ANALYST Market Research Basics Market research is the process of systematic gathering, recording and analyzing of data about customers, competitors
MARKETING RESEARCH AND MARKET INTELLIGENCE (MRM711S) FEEDBACK TUTORIAL LETTER SEMESTER `1 OF 2016. Dear Student
MARKETING RESEARCH AND MARKET INTELLIGENCE (MRM711S) FEEDBACK TUTORIAL LETTER SEMESTER `1 OF 2016 Dear Student Assignment 1 has been marked and this serves as feedback on the assignment. I have included
6. MEASURING EFFECTS OVERVIEW CHOOSE APPROPRIATE METRICS
45 6. MEASURING EFFECTS OVERVIEW In Section 4, we provided an overview of how to select metrics for monitoring implementation progress. This section provides additional detail on metric selection and offers
Running surveys and consultations
Running surveys and consultations Susannah Wintersgill, Public Affairs Annette Cunningham & Rob Markham, Estates Tuesday 9 June 2015 Contents Five key stages of planning a survey Questionnaire design:
National Disability Authority Resource Allocation Feasibility Study Final Report January 2013
National Disability Authority Resource Allocation Feasibility Study January 2013 The National Disability Authority (NDA) has commissioned and funded this evaluation. Responsibility for the evaluation (including
PROJECT MANAGEMENT PLAN CHECKLIST
PROJECT MANAGEMENT PLAN CHECKLIST The project management plan is a comprehensive document that defines each area of your project. The final document will contain all the required plans you need to manage,
Job Description Head of CRM
Date 10/15 Job Title Head of CRM Solutions Job Grade Management Location London/Solihull Department Client Services Reports to job title Managing Director Job purpose (A concise statement of why the job
Northeast Behavioral Health Partnership, LLC. Cultural Competency Program Description and Annual Plan
Cultural Competency Program Description and Annual Plan July 1, 2010 through June 30, 2011 Table of Contents Mission Statement... 1 Cultural Competency Program Description Introduction... 2 What is Cultural
Best Practices. Modifying NPS. When to Bend the Rules
Best Practices Modifying NPS When to Bend the Rules O ver the past decade, NPS (Net Promoter Score) has become an increasingly popular method for measuring and acting on customer feedback. Although there
Center for Effective Organizations
Center for Effective Organizations HR METRICS AND ANALYTICS USES AND IMPACTS CEO PUBLICATION G 04-8 (460) EDWARD E. LAWLER III ALEC LEVENSON JOHN BOUDREAU Center for Effective Organizations Marshall School
Grant Writing Dictionary
Grant Writing Dictionary Abstract: A one-page description of your project: the purpose, the number of participants, the amount requested, and how the project is unique. Administrative Costs: Grant funds
2 Business, Performance, and Gap Analysis
2 Business, Performance, and Gap Analysis The performance consulting process generally includes identifying business needs, performance needs, and work environment and capability needs. All strategic performance
Practice Overview. REQUIREMENTS DEFINITION Issue Date: <mm/dd/yyyy> Revision Date: <mm/dd/yyyy>
DEPARTMENT OF HEALTH AND HUMAN SERVICES ENTERPRISE PERFORMANCE LIFE CYCLE FRAMEWORK PRACTIICES GUIIDE REQUIREMENTS DEFINITION Issue Date: Revision Date: Document
Section 4: Key Informant Interviews
UCLA CENTER FOR HEALTH POLICY RESEARCH Section 4: Key Informant Interviews Purpose Key informant interviews are qualitative in-depth interviews with people who know what is going on in the community. The
DESIGNING A QUALITATIVE RESEARCH CONCEPTUAL FRAMEWORK AND RESEARCH QUESTIONS. Nicholas J. Sitko PROJECT: INDABA AGRICULTURAL POLICY RESEARCH INSTITUTE
DESIGNING A QUALITATIVE RESEARCH PROJECT: CONCEPTUAL FRAMEWORK AND RESEARCH QUESTIONS Nicholas J. Sitko June 7 th, 2013 INDABA AGRICULTURAL POLICY RESEARCH INSTITUTE 1 Interactive Model of Research Design
Non-Researcher s Guide to Evidence-Based Program Evaluation
Non-Researcher s Guide to Evidence-Based Program Evaluation July 2012 Table of Contents Table of Contents... 2 Course Overview... 4 About This Course... 4 Intended Audience... 4 Course Topics... 4 Learning
Workflow and Process Analysis for CCC
Section 3.6 Design Workflow and Process Analysis for CCC This tool introduces the importance of workflow and process improvement in a community-based care coordination (CCC) program, describes the value
Choosing The Right Evaluation Questions
PHASE II Choosing The Right Evaluation Questions Writing well thought-out, relevant evaluation questions is the secret to a meaningful evaluation. Keep in mind that the more specific the focus of the evaluation,
Promoting hygiene. 9.1 Assessing hygiene practices CHAPTER 9
74 CHAPTER 9 Promoting hygiene The goal of hygiene promotion is to help people to understand and develop good hygiene practices, so as to prevent disease and promote positive attitudes towards cleanliness.
Article Four Different Types of Evidence / Literature Reviews
Article Four Different Types of Evidence / Literature Reviews The rapid growth in the number of reviews undertaken can partly be explained by the current emphasis on evidence-based practice. Healthcare
Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS)
Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS) April 30, 2008 Abstract A randomized Mode Experiment of 27,229 discharges from 45 hospitals was used to develop adjustments for the
Program: Speech Pathology and Audiology B.S. Department: Speech Pathology and Audiology. Number of students enrolled in the program in Fall, 2011: 220
Program: Speech Pathology and Audiology B.S Department: Speech Pathology and Audiology Number of students enrolled in the program in Fall, 2011: 220 Faculty member completing template: Laureen O Hanlon
Written Example for Research Question: How is caffeine consumption associated with memory?
Guide to Writing Your Primary Research Paper Your Research Report should be divided into sections with these headings: Abstract, Introduction, Methods, Results, Discussion, and References. Introduction:
Differences between qualitative and quantitative research methods
Differences between qualitative and quantitative research methods Source: SEEP-AIMS 2000 Learning from Clients: Assessment Tools for microfinance practitioners. Washington, DC: AIMS/Management Systems
20 Best Practices for Customer Feedback Programs Building a Customer-Centric Company
Business growth through customer insight 20 Best Practices for Customer Feedback Programs Building a Customer-Centric Company The content of this paper is based on the book, Beyond the Ultimate Question,
Quality Management Plan Vision Purpose II. Mission III. Principles of Quality Management (QM)
J:HCC/Collaborative//QI Plan RWIII 1 FL DOH Monroe County Health Department (MCHD) Ryan White CARE Act Title III Early Intervention Services/Primary Care Vision Empowered people living healthy lives, self-managing
THE ADMINISTRATIVE UNIT ASSESSMENT HANDBOOK
THE ADMINISTRATIVE UNIT ASSESSMENT HANDBOOK GUIDELINES FOR PLANNING AND IMPLEMENTING MEASURING STUDENT SUPPORT SERVICES & ADMINISTRATIVE OUTCOMES Within institutional effectiveness, departments (academic
The fact is that 90% of business strategies are not implemented through operations as intended. Overview
Overview It is important to recognize that a company s network determines its supply chain efficiency and customer satisfaction. Designing an optimal supply chain network means the network must be able
Data Collection Strategies 1. Bonus Table 1 Strengths and Weaknesses of Tests
Data Collection Strategies 1 Bonus Table 1 Strengths and Weaknesses of Tests Strengths of tests (especially standardized tests) Can provide measures of many characteristics of people. Often standardized
A social marketing approach to behaviour change
A social marketing approach to behaviour change An e-learning course in using social marketing to change or sustain behaviour About The NSMC Established by Government in 2006, we are the centre of excellence
Investors in People Assessment Report. Presented by Alli Gibbons Investors in People Specialist On behalf of Inspiring Business Performance Limited
Investors in People Assessment Report for Bradstow School Presented by Alli Gibbons Investors in People Specialist On behalf of Inspiring Business Performance Limited 30 August 2013 Project Reference Number
Monitoring and Evaluation Plan Primer for DRL Grantees
Monitoring and Evaluation Plan Primer for DRL Grantees I. What is a monitoring and evaluation plan? A monitoring and evaluation plan (M&E plan), sometimes also referred to as a performance monitoring or
Performance Measures for Internal Auditing
Performance Measures for Internal Auditing A simple question someone may ask is Why measure performance? An even simpler response would be that what gets measured gets done. McMaster University s discussion
GUIDELINES FOR PROPOSALS: QUALITATIVE RESEARCH Human Development and Family Studies
Drafted by Lynet Uttal using the Quantitative Research Proposal Guidelines and in consultation with GPC (5/99) GUIDELINES FOR PROPOSALS: QUALITATIVE RESEARCH Human Development and Family Studies Overview:
Stages of Instructional Design V. Professional Development
Stages of Instructional Design V. Professional Development Derived from Gagné, R. M., Briggs, L. J., & Wager, W. W. (1992). Principles of Instructional Design (4th ed.). Fort Worth, TX: Harcourt Brace
SURVEY DESIGN: GETTING THE RESULTS YOU NEED
SURVEY DESIGN: GETTING THE RESULTS YOU NEED Office of Process Simplification May 26, 2009 Sarah L. Collie P. Jesse Rine Why Survey? Efficient way to collect information about a large group of people Flexible
Terms of Reference: Baseline Survey for Global Multi-tier Measurement for Access to Energy
Background Terms of Reference: Baseline Survey for Global Multi-tier Measurement for Access to Energy In order to realize the vision of the Government of Bangladesh (GOB) for universal access to electricity
6 Essential Characteristics of a PLC (adapted from Learning by Doing)
6 Essential Characteristics of a PLC (adapted from Learning by Doing) 1. Shared mission, vision, values, goals Educators in a PLC benefit from clarity regarding their shared purpose, a common understanding
Tips & Tools #17: Analyzing Qualitative Data
Tips & Tools #17: Analyzing Qualitative Data Introduction Qualitative evaluation methods yield narrative data often describing experiences, perceptions, or opinions from key informant interviews, open
Fourth generation techniques (4GT)
Fourth generation techniques (4GT) The term fourth generation techniques (4GT) encompasses a broad array of software tools that have one thing in common. Each enables the software engineer to specify some
9 The Difficulties Of Secondary Students In Written English
9 The Difficulties Of Secondary Students In Written English Abdullah Mohammed Al-Abri Senior English Teacher, Dakhiliya Region 1 INTRODUCTION Writing is frequently accepted as being the last language skill
Step 1: Analyze Data. 1.1 Organize
A private sector assessment combines quantitative and qualitative methods to increase knowledge about the private health sector. In the analytic phase, the team organizes and examines information amassed
Certified in Public Health (CPH) Exam CONTENT OUTLINE
NATIONAL BOARD OF PUBLIC HEALTH EXAMINERS Certified in Public Health (CPH) Exam CONTENT OUTLINE April 2014 INTRODUCTION This document was prepared by the National Board of Public Health Examiners for the
Charting our outcomes
Charting our outcomes 2010-2016 January 2011 P r e p a r e d b y : Evangeline Danseco, PhD Director, Evaluation and Research Sherry McGee, MA Research associate 1 TABLE OF CONTENTS INTRODUCTION... 3 LOGIC
Capacity Assessment Indicator. Means of Measurement. Instructions. Score As an As a training. As a research institution organisation (0-5) (0-5) (0-5)
Assessing an Organization s in Health Communication: A Six Cs Approach [11/12 Version] Name & location of organization: Date: Scoring: 0 = no capacity, 5 = full capacity Category Assessment Indicator As
Measuring the Impact of Volunteering
Measuring the Impact of Volunteering Why is measuring the impact of volunteering important? It is increasingly important for organisations or groups to describe the difference that volunteering makes to,
Evaluating Assessment Management Systems: Using Evidence from Practice
Evaluating Assessment Management Systems: Using Evidence from Practice Lisa Plantefaber, Ph.D. Associate Dean, Institutional Research and Assessment Westfield State University Westfield, Massachusetts,
Software Development Methodologies in Industry. By: Ahmad Deeb
Software Development Methodologies in Industry By: Ahmad Deeb Methodologies Software Development Methodologies in Industry Presentation outline SDM definition Project and analysis approach Research methods
TOOL. Project Progress Report
Purpose TOOL SUMMARY: LOGFRAME CHEAT SHEET The purpose of the is to compile information from the analysis done by project participants, partners and LWR country staff about the progress or advances the
FFIEC Cybersecurity Assessment Tool
Overview In light of the increasing volume and sophistication of cyber threats, the Federal Financial Institutions Examination Council 1 (FFIEC) developed the Cybersecurity Tool (), on behalf of its members,
Glossary Monitoring and Evaluation Terms
Glossary Monitoring and Evaluation Terms This glossary includes terms typically used in the area of monitoring and evaluation (M&E) and provides the basis for facilitating a common understanding of M&E.
RIPE NCC Activity Plan 2005
RIPE NCC Activity Plan 2005 Table of Contents 1) Vision and Strategy 2) New Activities 2005 3) Established Activities 4) Unforeseen Activities 1) Vision and Strategy The RIPE NCC will strengthen the open,
Safe and Healthy Minnesota Students Planning and Evaluation Toolkit
Safe and Healthy Minnesota Students Planning and Evaluation Toolkit AUGUST 2008 Safe and Healthy Minnesota Students Planning and Evaluation Toolkit Tips, techniques, and resources for planning and evaluating
Using qualitative research to explore women s responses
Using qualitative research to explore women s responses Towards meaningful assistance - how evidence from qualitative studies can help to meet survivors needs Possible questions Why do survivors of SV
This alignment chart was designed specifically for the use of Red River College. These alignments have not been verified or endorsed by the IIBA.
Red River College Course Learning Outcome Alignment with BABOK Version 2 This alignment chart was designed specifically for the use of Red River College. These alignments have not been verified or endorsed
SURVEY RESEARCH RESEARCH METHODOLOGY CLASS. Lecturer : RIRI SATRIA Date : November 10, 2009
SURVEY RESEARCH RESEARCH METHODOLOGY CLASS Lecturer : RIRI SATRIA Date : November 10, 2009 DEFINITION OF SURVEY RESEARCH Survey: A method of primary data collection based on communication with a representative
Quality Risk Management The Pharmaceutical Experience Ann O Mahony Quality Assurance Specialist Pfizer Biotech Grange Castle
Quality Risk Management 11 November 2011 Galway, Ireland Quality Risk Management The Pharmaceutical Experience Ann O Mahony Quality Assurance Specialist Pfizer Biotech Grange Castle Overview Regulatory
A framework to plan monitoring and evaluation
6. A framework to plan monitoring and evaluation Key ideas Monitoring and evaluation should be an integral part of (and therefore affect) all stages of the project cycle. As mentioned earlier, you need
MaineCare Value Based Purchasing Initiative
MaineCare Value Based Purchasing Initiative The Accountable Communities Strategy Jim Leonard, Deputy Director, MaineCare Peter Kraut, Acting Accountable Communities Program Manager Why Value-Based Purchasing
Minnesota Co-occurring Mental Health & Substance Disorders Competencies:
Minnesota Co-occurring Mental Health & Substance Disorders Competencies: This document was developed by the Minnesota Department of Human Services over the course of a series of public input meetings held
CALCULATIONS & STATISTICS
CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents
An Introduction to Secondary Data Analysis
1 An Introduction to Secondary Data Analysis What Are Secondary Data? In the fields of epidemiology and public health, the distinction between primary and secondary data depends on the relationship between
Universiteit Leiden. ICT in Business. Leiden Institute of Advanced Computer Science (LIACS) Capability Maturity Model for Software Usage
Universiteit Leiden ICT in Business Capability Maturity Model for Software Usage Name: Yunwei Huang Student-no: s1101005 Date: 16/06/2014 1st supervisor: Dr. Luuk Groenewegen 2nd supervisor: Dr. Nelleke
Logic Models, Human Service Programs, and Performance Measurement
Three Logic Models, Human Service Programs, and Performance Measurement Introduction Although the literature on ment has been around for over two decades now, scholars and practitioners still continue
Guided Reading 9 th Edition. informed consent, protection from harm, deception, confidentiality, and anonymity.
Guided Reading Educational Research: Competencies for Analysis and Applications 9th Edition EDFS 635: Educational Research Chapter 1: Introduction to Educational Research 1. List and briefly describe the
Insight for Informed Decisions
Insight for Informed Decisions NORC at the University of Chicago is an independent research institution that delivers reliable data and rigorous analysis to guide critical programmatic, business, and policy
An Introduction to Organizational Maturity Assessment: Measuring Organizational Capabilities
International Public Management Association Assessment Council An Introduction to Organizational Maturity Assessment: Measuring Organizational Capabilities Selena Rezvani, M.S.W. Objectives Define and
