click below
click below
Normal Size Small Size show me how
Adv. Motion
Adv. Motion Analysis
Question | Answer |
---|---|
The observed score is composed of what two entities? | True quantity (X) Error of measurement (em) |
What is equal to the difference between the observed (x) and true quantity (X) | Error of measurement |
What is the equation for Error measurement? | em = x (observed) - X (true quantity) OR X = x + em |
Lack of random error | Reliability |
Validity = ________ | Reliability + Corrections (If no errors then corrections = 0) |
You can't have validity without? | Reliability |
Reliability is quantified by the degree to which measurements are ____ and ____ | consistent (constant/stable) reproducible (repeatable) |
____ concerns the extent to which the instrument yields the same measurement on repeated uses, either by the same operator or by different operators | Reliability |
T/F. Reliability presupposes validity | FALSE...Reliability does not presuppose validity |
Reliability or Validity? Stability, Consistency, Reproducibility | Reliability |
Standard deviation as a percentage of mean | Coefficient of Variation (CV) |
You use ___ to measure if things/trials are different | ANOVA |
You use ___ to compare if things are similar | Intraclass Correlation Coefficient (ICC) |
This is a measure of association/correlation | Pearson Product-Moment Correlation Coefficient (r) |
Pearson Product-Moment Correlation Coefficient (r) | is a measure of the correlation (linear dependence) between two variables X and Y, giving a value between +1 and −1 inclusive Greater than .75 is good Less than .75 is poor to moderate |
Intraclass Correlation Coefficient (ICC) | is a measure of the reliability of measurements or ratings. Greater than .75 is good Less than .75 is poor to moderate |
Name the types of reliability | Intra-tester Inter-tester Intra-subject Instrument Internal consistency Parallel |
Repeatability of measurements taken by the same tester at different times | Intra-rater reliability |
Reproducibility of measurements taken by different testers | Inter-rater reliability |
Describe a p-value of 0.01 | Basically that there is a 1% chance it is wrong and a 99% chance it is correct |
Different testers able to get the same value | inter-rater |
Tester is able to repeat their own measure | Intra-rater |
Which concept is being described. If you don't do well in one questionnaire you should do poorly in another similar questionnaire | Internal consistency |
Why are reliable measures important? | If the reliability of the measurements aren't reported in a study, the consistency of the measurements performed in unknown. The results of the study are questionable b/c the diff. could have been caused by random measurement errors and not real changes |
A calibration issue is an example of what type of reliability? | instrument |
Reproducibility of the subjects' performance (measure taken by the same tester) | Intra-subject reliability |
What compares the measures of the same subject taken at 2 or more times when the same tester performs the measure | Intra-subject |
Ability of the measuring instrument or test to give the same consistent measurements or results | Instrument reliability |
The degree to which different items or set of questions in a test, scale, questionnaire, or interview protocol are associated to each other | Internal consistency |
Internal consistency involves the calculation of the correlation b/w | different items of the intstrument |
Evaluates whether measurements or results from different devices or tests are similar or the same | Parallel reliability |
Lack of systematic error | Validity |
Relates to the extent to which an instrument measures what it is intended to measure | Validity |
Valid/Reliable A ___ measure allows one to rely on the study/test results and interpretations when making clinical decision | Valid |
Validity is quantified by | Accuracy/Correctness Ability to make inferences (conclusions reached based on the basis of evidence) Lack of systematic error |
Name the types of validity | Face Construct (convergent, discriminate) Content Criterion-related (predictive, concurrent, prescriptive) |
Which type of validity relates to questionnaires | Content |
What is the subdivisions of construct validity? | Convergent Discriminate |
What are the subdivisions of Criterion-related validity | Predictive Concurrent Prescriptive |
Which validity? Lowest level of validity is? | Face |
Which validity? Related to the intuitive assumption that a measurement seems to be valid | Face validity |
Which validity? "Feeling" that its valid...It makes sense | Face validity |
Which validity? Does it makes sense and can I provide a rationale? | Construct |
Which validity? Theoretical framework and reasoning to support the measurement is valid | Construct |
Which type of construct validity? Measurements that are believed to reflect the same variables that yield similar results and are comparable | Convergent |
Which type of construct validity? Indicated that measurements that are supposed to assess different characteristics yield different results and are not comparable | Discriminate |
Which validity? Applied to study questionnaires; How well information gathered fully reflects the variable under study | Content |
Which type of validity? Determined by comparing a measurement with a particular factor | Criterion-related |
Which type of validity? Determined whether the predictions originating from the measurements come true | Predictive validity |
Which type of validity? Can I predict an outcome based on this measure | Predictive |
Which type of validity? Used to test whether a new instrument is interchangeable with an establish "golden standard" | Concurrent |
Which type of validity? Relates to how appropriate it is to use a measurement to recommend a treatment | Prescriptive |
The ability of the instrument, device, tool, test, or scale to accurately detect meaningful changes | Responsiveness |
Instrument responsiveness includes the concepts of? | Sensitivity Specificity |
True/False A gonio with 20 degree increments is reliable for responsiveness | False. It is not due to the increments being too big...It is not sensitive to change |
Ability of an instrument to detect changes in the variables when they occur | Sensitivity |
Ability of an instrument's measure not to change when no changes in the variable occur | Specificity |
Sensitivity/Specificity ___ is related to the actual testing of the discriminate validity of a measure | Specificity |
T/F Reporting the measurement characteristics can improve the quality of rehab research and clinical evaluation processes which are fundamental for ideal evidence based practice | True...It's the "Take home message" |
Characteristics of a measurement include | Reliability Validity Responsiveness |
The characteristics of a measurement, along with the objectives of measuring the event in question, should guide the decision concerning which measurements properties or its type, are the most important for a specific study | The characteristics of a measurement, along with the objectives of measuring the event in question, should guide the decision concerning which measurements properties or its type, are the most important for a specific study |
Validity/Reliability/Responsiveness Content? Sensitivity? Inter-tester? Face? Parallel? Content? Construct? Specificity? | Content - Validity Sensitivity - Responsiveness Inter-tester - Reliability Face - Validity Parallel - Reliability Content - Validity Construct - Validity Specificity - Responsiveness |
Physical exertion assessment methods include: | Qualitative methods Pen/Paper-based observation Questionnaires Video-recording Computer-based analysis Direct measurement (EMG,EGM, etc) Combined assessment - mixed methods |
Common examples of qualitative methods is | Interview, focus groups, diaries, photo as visual narratives |
Advantages of Interviews (qualitative methods)? | An interview is common one Adv: -Allows more elaborate answers as opposed to questionnaire - More indepth info |
Disadvantages of Interviews (qualitative methods)? | Time consuming Person could be shy Person can choose what to say Person may be uncomfortable (face to face) |
Advantages of focus groups (Qualitative methods) | Different ideas/perspectives Elicitation of memory by hearing others story Interaction |
Disdvantages of focus groups (Qualitative methods) | Shyness/uncomfortable with group speaking Age/Gender can cause decrease participation Composition of group must be well thought so person isn't scared to speak b/c boss is around Multiple ppl talking at the same time Organization |
What are some visual estimation limits (looking at a picture) | its 2 dimensional It is qualitative...there is no measure (no quantitative) Can't really quantify it |
Name of a survey/questionnaire (we use commonly) that evaluates effort/perceived exertion | Borg RPE |
Visual Analog scale evaluates? | Pain and/or effort exerted |
Which is the most physically demanding task of a outpatient nurse? Inpatient nurse? | ON --> Patient transfers IN--> Patient repositioning |
The statistical analysis of measurement reliability is frequently performed using the? | Pearson correlation coefficient and the intraclass correlation coefficient (ICC) |
What is a measure of rater agreement on a discrete outcome | Kappa test |
____ is determined by comparing the values or results taken by different instruments or tests at the same time. | Parallel reliability |
____ validity depends on the fact that measurements that lack a relationship should not be related | Discriminate |
____ validity supports those measurements that are believed to reflect the same variables that yield similar results and are comparable | Convergent |
An example of a question that deals with ____ validity is “Do the questions in this independence for daily activities’ assessment questionnaire fully cover the problems that a disable person will face in his/hers daily life activities?” | Content |
Evaluates whether measurements taken using different instruments agree with each other. It is used to test whether a new instrument is interchangeable with an established “gold standard” | Concurrent validity |
T/F Concurrent validity and parallel reliability are similar. | True |
What is the difference between concurrent validity and parallel reliability? | The difference is that concurrent validity requires the reference measurements to come from a valid instrument, while parallel reliability only requires reliable measurements from the criterion instrument |
The ____ is the probability that the researcher is willing to accept that he or she might be wrong in rejecting the null hypothesis, or the extent to which the researcher could be wrong in saying that there are differences. | α-level (statistical significance level) |
Lowering the α-level _____ the chances of a type I error | reduces/lowers |
The higher the α-level (i.e. higher than 0.05), the _____ the power, but also the _____ the chances of making a type I error | Greater power Higher chances |