Developed by Jennifer Docktor and Ken Heller
To assess written solutions to problems given in undergraduate introductory physics courses.
|Focus||Problem-solving (Useful problem description, physics approach, specific application of physics, mathematical procedures, logical progression)|
|Level||Intro college, High school|
Everything you need to know about implementing the MAPS in your class.
The developers website contains more information about the MAPS rubric and training materials to help you learn how to use the rubric for general purposes or research purposes: http://groups.physics.umn.edu/physed/rubric.html
This is the second highest level of research validation, corresponding to at least 5 of the validation categories below.
Research Validation Summary
Based on Research Into:
- Student thinking
- Student interviews
- Expert review
- Appropriate statistical analysis
- At multiple institutions
- By multiple research groups
- Peer-reviewed publication
The MAPS rubric is based on research on student problem solving at the University of Minnesota over many years. The MAPS rubric builds on previous work by attempting to simplify the rubric and adding more extensive tests of validity, reliability, and utility. The five problem solving processes covered in the rubric are consistent with prior research on problem-solving in physics (Docktor 2009). The validity, reliability, and utility of the rubric scores were studied in a number of different ways. Expert reviewers used the rubric to understand how rubric scores reflect the solvers process, the generalizability of the rubric as well as inter-rater agreement. Subsequent studies looked at the content relevance & representativeness, how the training materials influenced the inter-rater agreement and the reliability and utility of the rubric. Based on these studies, both the rubric and training materials were modified. The rubric was also studied using student interviews. Overall, the validity, reliability and utility of the MAPS rubric were demonstrated with these studies. Research on the MAPS rubric is published in one dissertation, and one peer-reviewed publication.
- J. Docktor, Development and Validation of a Physics Problem-Solving Assessment Rubric, Dissertation, University of Minnesota, 2009.
- J. Docktor, J. Dornfeld, E. Frodermann, K. Heller, L. Hsu, K. Jackson, A. Mason, Q. Ryan, and J. Yang, Assessing student written problem solutions: A problem-solving rubric with application to introductory physics, Phys. Rev. Phys. Educ. Res. 12 (1), 010130 (2016).
- J. Docktor and K. Heller, Assessment of Student Problem Solving Processes, presented at the Physics Education Research Conference 2009, Ann Arbor, Michigan, 2009.
We don't have any translations of this assessment yet.
If you know of a translation that we don't have yet, or if you would like to translate this assessment, please contact us!
Download the MAPS scoring tool.
Because the MAPS is a rubric used to score physics problem solutions, there are no typical scores for this assessment. Here is an example of what the scores on this assessment look like (Table 24) for a specific physics problem (Test 1 Problem 2) for a specific group of students as reported in Docktor 2009.
Coming soon: The PhysPort Assessment Data Explorer
Start learning more from your tests.
- Get 1-click statistics
- Compare to students like yours
- Get practical, personalized recommendations
The latest version of the MAPS, version 4.4, was released in 2008.