click below
click below
Normal Size Small Size show me how
WGU RFC 1 ch 6
Summary of Chapter 6
| Question | Answer |
|---|---|
| All types of research | require collecting data |
| data | are pieces of evidence used to examine a research topic or hypothesis |
| Constructs | mental abstractions such as personality, creativity, and intelligence that cannot be observed or measured directly |
| Constructs | become variables when they are stated in terms of operational definitions |
| Variables | placeholders that can assume any one of a range of values |
| Categorical variables | assume non-numerical (nominal) values |
| quantitative variables | assume numerical values and are measured on an ordinal, interval, or ratio scale |
| Independent variable | the treatment or cause |
| Dependent variable | outcome or effect of the independent variable |
| Three main ways to collect data for research studies | 1. administering an existing instrument 2. constructing one's own instrument 3. recording naturally occurring events (i.e., observations) |
| Standardized test | administered, scored, and interpreted in the same way no matter when and where it is administered |
| Most quantitative tests are | paper-and-pencil ones |
| Most qualitative researchers | collect data by observation or oral questioning |
| Raw scores | indicate the number of items or points a person got correct |
| Norm-referenced scoring | compares a student's test performance to the performance of other test takers |
| Criterion-referenced scoring | compares a student's test performance to predetermined standards of performance |
| Cognitive tests | measures intellectual processes |
| achievement tests | measure the current status of individuals on school-taught subjects |
| aptitude tests | used to predict how well a test taker is likely to perform in the future |
| general aptitude tests | typically ask the test taker to perform a variety of verbal and nonverbal tasks |
| Affective tests | are assessments designed to measure characteristics related to emotion |
| most affective tests | are non-projective, self report measures in which the individual responds to a series of questions about him or her self |
| Five basic types of scales are used to measure attitudes | 1. Likert scales 2. semantic differential scales 3. rating scales 4. Thurstone scales 5. Guttman scales |
| Attitude scales | ask respondents to state their feelings about various objects, persons, and activities |
| Likert scale | indicate feeling a long a scale such as strongly agree, agree, undecided, disagree, strongly disagree |
| Semantic differential scales | present a continuum of attitudes on which a respondent selects a position to indicate the strength of attitude. On a scale of 1-5, 5 being the best and 1 being the worst |
| rating scales present | statements that respondents must rate on a continuum from high to low. Rank your feelings from 1-5, 1 being the most important and 5 being the least important |
| Interest inventories | ask individuals to indicate personal likes and dislikes |
| Personality describes | characteristics that represent a persons typical behavior |
| Personality inventories include | lists of statements describing human behaviors, and participants must indicate whether each statement pertains to them. |
| Personality inventories | may be specific to a single trait (introversion-extroversion) or maybe general and measure a number of traits. |
| Use of self-report measures | create a concern whether an individual is expressing his or her true attitude, values, interests, or personality |
| Test bias in both cognitive and affective measures | can distort the data obtained |
| Bias is present | when one's ethnicity, race, gender, language, or religious orientation influences test performances |
| Projective tests | presents ambiguous situation and require the test taker to "project"her or his true feelings on the ambiguous situation |
| Association | most commonly used projective technique and exemplified by the inkblot test |
| Validity | the degree to which a test measure what is supposed to measure |
| Test is | not valid per se, it is valid for a particular interpretation and for a particular group |
| Validity | is measured on a continuum, tests are highly valid, moderately valid, or generally valid |
| Content validity | assesses the degree to which a test measures an intended content area |
| Content validity | is of prime importance for achievement tests |
| Content validity | determined by expert judgment of item and sample validity, not by statistical means |
| Criterion related validity | determined by relating performance on a test to a performance on a second test or other measure |
| Criterion validity has two forms | concurrent and predictive |
| Concurrent validity | degree to which the scores on a test are related to scores on another test administered at the same time or to another measure available at the same time |
| Predictive validity | degree to which scores on a test are related to scores on another test administered in the future |
| Construct validity | measure of whether the construct underlying a variable is actually being measured |
| Construct validity | determined by a series of validation studies that can include content and criterion related approaches |
| Used to determine construct validity | confirmatory and disconfirmatory evidence |
| Consequential validity | concerned with the potential of tests to create harmful effects for test takers |
| Validity of any test or measure | can be diminished by such factors as unclear test directions, ambiguous or difficult test items, subjective scoring, and non-standardized administration procedures |
| Reliability | degree to which a test consistently measure whatever it measures |
| Reliability | is expressed numerically from 0.0 to 1.0; a high coefficient indicates high reliability |
| Measurement error | refers to the inevitable fluctuations in scores due to person and test factors |
| No test is perfectly reliable | but the smaller measurement error the more reliable the test |
| The five general types of reliability | 1. Stability 2. Equivalence 3. Equivalence and stability 4. Internal consistency 5. Scorer/rater |
| Stability | also called test-retest reliability; the degree to which test scores are consistent over time |
| Test-retest | determined by correlating scores from the same test, administered more than once |
| Equivalence | Equivalent-forms reliability; the degree to which two similar forms of a test produce similar scores from a single group of test takers |
| Equivalence and stability reliability | the degree to which two forms of a test given at two different times produce similar scores as measured by correlations |
| Internal consistency | deals with the reliability of a single test taken at one time |
| Internal consistency | measures the extent to which the items in the test are consistent among themselves with the test as a whole |
| Internal consistency | Split half, Kuder-Richardson 20 and 21, and Cronbach's alpha main approaches to obtaining internal consistency |
| Split half reliability | determined by dividing a test into two equivalent halves (e.g., odd items vs. even items), correlating the two halves, and using the Spearman Brown formula to determine the reliability of the whole test |
| Kuder-Richardson reliability | deals with internal consistency of tests that are scored dichotomously (i.e., right, wrong)(multiple choice items, true or false items) |
| Cronbach's alpha | deals with internal consistency of tests that are scored with more than two choices (How many previous research classes have you taken? select among the following 0,1,2,3) |
| Scorer/rater reliability | important when scoring tests that are potentially subjective |
| interjudge reliability | refers to the reliability of two or more independent scorers |
| Intrajudge reliability | refers to the reliability of a single individual's rating over time |
| Standardized achievement tests | high reliabilities |
| Porjective tests | considerably lower reliabilities |
| standard of error or measurement | is an estimate of how often one can expect score errors of a given size |
| small standard error of measurement | indicates high reliability |
| large standard error of measurement | indicates low reliability |
| standard error of measurement | used to estimate the difference between a person's obtained and true scores |
| Big differences | indicate low reliability |
| Metal Measurement Yearbooks (MMYs) | most comprehensive sources of test information available |
| Metal Measurement Yearbooks (MMYs) | provide factual information on all known or revised tests, test reviews, and comprehensive bibliographies and indexes |
| Tests in Print (TIP) | comprehensive bibliography of all tests that have appeared in preceding MMYs. |
| Pro-Ed Publications' Tests | describe more than 2,000 tests in education, psychology, and business |
| Test Critiques | reviews of many of these tests |
| ETS Test Collection Database | describes more than 20,000 tests, published and unpublished |
| Other sources of test information | professional journals and test publishers or distributors |
| Three most important factors to consider in selecting a test | 1. validity 2. reliability 3. ease of use |
| Self constructed tests | should be pilot tested before use to determine validity, reliability, and feasibility |