Download presentation
Presentation is loading. Please wait.
Published bySarina Cove Modified over 9 years ago
2
Class Meeting #4 Qualitative Methods Measurement Tools & Strategies
3
Vocabulary Program drift Triangulation Temporal sequencing Units of analysis Gatekeepers Key informants Participant observers Saturation Case studies
4
Key Characteristics Focus on naturalistic inquiry in situ Reliance on the researcher as the instrument of data collection Reports emphasize narrative over numbers
5
Uses Process evaluations –The “hows” of a program –The “whys” of a program Formative evaluations
6
Advantages Allows examination of complex phenomena without relying on structured data collection Increases chances of uncovering unanticipated but meaningful insights
7
Disadvantages Depth rather than breadth Not suited to large numbers of subjects Time-intensive Labor-intensive Less useful when precision is needed
8
Design Issues Units of analysis (settings) Comparisons (plausibility) Flexible design (serendipity) Access (informants) Sampling (purposeful) Data collection (saturation) Data analysis(meaning units)
9
Methods of Sampling Deviant case Typical case Maximum variation Snowball Convenience
10
Data Collection On-site observation In-depth interviewing Use of documents
11
Measurement Tools Collecting and Compiling Information
12
MethodOverall PurposeAdvantagesChallenges questionnaires, surveys, checklists when need to quickly and/or easily get lots of information from people in a non threatening way -can complete anonymously -inexpensive to administer -easy to compare and analyze -administer to many people -can get lots of data -many samples already exist -might not get careful feedback -wording can bias client's responses -are impersonal -in surveys, may need sampling expert - doesn't get full story interviewswhen want to fully understand someone's impressions or experiences, or learn more about their answers to questionnaires -get full range and depth of information -develops relationship with client -can be flexible with client -can take much time -can be hard to analyze and compare -can be costly -interviewer can bias responses documentation reviewwhen want impression of how program operates without interrupting the program; is from review of applications, finances, memos, minutes, etc. -get comprehensive and historical information -doesn't interrupt program or client's routine in program -information already exists -few biases about information -often takes much time -info may be incomplete -need to be quite clear about what looking for -not flexible means to get data; data restricted to what already exists observationto gather accurate information about how a program actually operates, particularly about processes -view operations of a program as they are actually occurring -can adapt to events as they occur -difficult to interpret seen behaviors -can be complex to categorize observations -can influence behaviors of program participants -can be expensive focus groupsexplore a topic in depth through group discussion, e.g., about reactions to an experience or suggestion, understanding common complaints, etc.; useful in evaluation and marketing -quickly and reliably get common impressions -can be efficient way to get much range and depth of information in short time - can convey key information -can be hard to analyze responses -need good facilitator for safety and closure -difficult to schedule 6-8 people together case studiesto fully understand or depict client's experiences in a program, and conduct comprehensive examination through cross comparison of cases -fully depicts client's experience in program input, process and results -powerful means to portray program to outsiders -usually quite time consuming to collect, organize and describe -represents depth of information, rather than breadth
13
Selecting Methods What information is needed? How much can be collected and analyzed in a low-cost and practical manner using questionnaires, surveys, and checklists? How accurate will the informatin be? Will the method get all of the needed information?
14
Selecting Methods What additional methods should and could be used? Will the information appear as credible to decision makers? Will the nature of the audience conform to the methods? Who can administer the methods? How can the information be analyzed?
15
Citation McNamara, C. (1998). “Basic Guide to Program Evaluation.” Downloaded 1/5/06 from http://www.managementhelp.org/evaluatn/fnl_ eval.htm. http://www.managementhelp.org/evaluatn/fnl_ eval.htm
16
Measurement Tools Behavioral outcomes Scale vs. index Scale measures a solitary concept Index involves creating a new variable that is the sum of other variables thought to measure a single construct Reliability & validity Instrument construction Levels of measurement
17
Data Analysis “The purpose of data analysis is to answer the questions that were the catalyst for the investigation.” Look for patterns, trends, or relationships
18
Univariate Analysis Frequency distributions Central tendency Variability
19
Bivariate Analysis Paired samples t-test t-test for independent samples One-way analysis of variance Chi-square Correlations
20
Multi-Variate Analysis Multivariate analysis of variance Multiple regression Discriminant analysis
21
Types of Errors Type I – you conclude that there is a relationship between two variables when, in fact, there is none. Type II – you fail to find a relationship that actually does exist.
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.