Save
Busy. Please wait.
Log in with Clever
or

show password
Forgot Password?

Don't have an account?  Sign up 
Sign up using Clever
or

Username is available taken
show password


Make sure to remember your password. If you forget it there is no way for StudyStack to send you a reset link. You would need to create a new account.
Your email address is only used to allow you to reset your password. See our Privacy Policy and Terms of Service.


Already a StudyStack user? Log In

Reset Password
Enter the associated with your account, and we'll email you a link to reset your password.
focusNode
Didn't know it?
click below
 
Knew it?
click below
Don't Know
Remaining cards (0)
Know
0:00
Embed Code - If you would like this activity on your web page, copy the script below and paste it into your web page.

  Normal Size     Small Size show me how

MEC1

Foundations of Measurement and Evaluation

QuestionAnswer
Examples of Subjective Assessment Essay Answers
Define Objective Assessment Objective assessment is a form of questioning which has a single correct answer.
Validity Evidence There are different types of validity. Face, content, construct, and criterion related.
Criterion Related Validity Validity that is determined by gauging a test resently created to a standard (which can be in the form of another similar test). The standard could be in existance presently or in the future.
Examples of Assesment Achievement tests, aptitude test, and effective tests.
Evaluation Evaluations focus on Internal situations. An ex. is collecting data on a specific program
Threats to validity Unclear test items, unknown vocabulary, items not covered in instruction, inconsistent testing procedures, confusing test directions.
Types of Formative Eval Questions Questions should not be answered with yes or no. Answers should be towards making the instruction better. The answers should give informative info for improvements.
Define Reliability A degree to which a test is consistently measures whatever it is measuring. A test maybe reliable without being valid but it cannot be valid without being reliable. A test must have stability, equivalence, and internal consistency to be reliable.
Validity Assessment validity is when a test measures what it is intended to measure.
Define Formative Evaluation Is used to improve instruction or informs the person of revisions to instruction. These evals are performed during the development of instruction. Ex. You want to know if the instructional goal was accomplished and any revisions that the product needs.
Split-Half reliability Scores are seperated into two groups and the group results are compared to determine the simularities.
Two types of Criterion Related Validity Concurrent and predictive. Concurrent means the standard used to gauge validity is something in existance in the present (another test concurrently being used). Predictive uses future performance as the standard to gauge validity.Ex. SAT
Examples of Formative Evaluation SME, One on One interviews, small group or focus group, field trial in the location where the instruction is intended to take place.
Reliability Coefficients Statistical means of estimating and interpreting the reliability of assessments.
Examples of summ eval Did it work, was it successfull, should we countinue funding.
Constructed Response Tests Performance tasks such as essay, and or profolios.
Content Validity The test only includes items that are from a specific content area.
Examples of Objective Assessment Multiple choice, true false,multiple-response and matching questions
Types of Reliability Internal reliability, Rater reliability, Equivalence reliability, and Stability reliability.
Examples of Authentic Assesment performance of the skills, or demonstrating use of a particular knowledge simulations and role plays studio portfolios, strategically selecting items exhibitions and displays
Define Authentic Assessment Measurement of "intellectual accomplishments
Face Validity Does the face of the test appear to be valid.
Program eval or summative eval The program itself is being evaluated. An evaluation can be a product, project, or process. The purpose is to make a decision from the results of the eval. Decisions can include discountinuing a program, starting a new one, or comparing programs.
Define Subjective Assessment Subjective assessment is a form of questioning which may have more than one correct answer (or more than one way of expressing the correct answer
Types of Assesments Tests are the only type of Assesment.
Definition of Assesment Assesment means collecting information on an individual student's competence, ability, or improvement. Basicaly gathering data on a student ability to learn or what they have already learned.
Selected Response Tests Mutiple choice, true or false, matching. Students have to choose answers provided by the test maker.
Two Phases of Prog or summ eval Expert Judgement Phase and Field Trial Phase. Judge if it works and fiels is actually putting it to the test.
Inter-Rater Reliability Determine the simularities of the graders who score tests.There needs to be a high degree of consistancy.
Diffences and simularities in eval and research Both use same tools and methods Research generalizes while evals particularizes The difference is in the purpose.
Who are stakeholders Stakeholders are people that will be impacted by the results of the evaluation. Sometimes it may be the developer of the program and other times it may be groups. Ex. parents, teachers, students.
Construct Valdility The experimental demonstration that a test is measuring the construct it claims to be measuring. Construct vadility is not observable. Ex. honesty and intelligence.
Test and Retest Test and Retests Reliability determines if the same test will give simular results when administered to the same people two different times. A relationship between the scores must be detremined.
Differences between summ and forma eval Formative eval is used to make improvements to the exsisting program. Summ is used to make a decision on the program NOT to make improvements.
Research Research is conducted to generalize your findings from a sample to a larger population.
Evaluation tip With evaluations you do not want to generaluze the data for a larger population. The point of the evaluation is to determine if the program was successful.
Reliability tests Test Retest (Tests stability), split half test (tests internal reliability), and Equivalent form (tests equivalence reliability.
Importance of reliability Reliability is important because it indicates that a test will consistently measure what it says it will measure.
Importance of validity Validity is important becaue it dicates that a test will consistently measure what it purports to measure over time.
Created by: kasuao1
Popular Miscellaneous sets

 

 



Voices

Use these flashcards to help memorize information. Look at the large card and try to recall what is on the other side. Then click the card to flip it. If you knew the answer, click the green Know box. Otherwise, click the red Don't know box.

When you've placed seven or more cards in the Don't know box, click "retry" to try those cards again.

If you've accidentally put the card in the wrong box, just click on the card to take it out of the box.

You can also use your keyboard to move the cards as follows:

If you are logged in to your account, this website will remember which cards you know and don't know so that they are in the same box the next time you log in.

When you need a break, try one of the other activities listed below the flashcards like Matching, Snowman, or Hungry Bug. Although it may feel like you're playing a game, your brain is still making more connections with the information to help you out.

To see how well you know the information, try the Quiz or Test activity.

Pass complete!
"Know" box contains:
Time elapsed:
Retries:
restart all cards