Measure Progress
The CMNS Teaching and Learning Center lays out how they use this five-layer model for evaluating their professional development programs to improve teaching and learning.
Participation
What is it?
One of the most straightforward metrics to track is the number and demographics of the people who attend a given event.
How do you measure it?
Here are some ways participation data can be collected.- Track applications and attendance of events, including demographics and characteristics (e.g. department, professional title, etc.). You may use this to determine the characteristics of the audience that is attending, so that efforts can be made to recruit a more diverse audience for future events.
- Create and track membership of an email discussion list.
- Use surveys to gauge how participants learned about your program and reason for participation.
Data from these and other sources can be accumulated in a database for easier analysis.
Satisfaction
What is it?
This level refers to the extent to which attendees and facilitators were satisfied with their experience in the activity provided by the center. These data can be used to increase participation in future events.
How do you measure it?
Here are some ways satisfaction data can be collected and measured:- Conduct a survey among activity participants that asks about their program satisfaction. Ideally, distribute the survey as part of the program in order to maximize sample size, rather than waiting until a later date. Ask questions like:
- Would you recommend this program to a colleague (Yes/No) Why?
- Which components did you find most helpful/least helpful?
- Conduct individual interviews or focus groups with participants after the event.
- Facilitators can discuss among themselves whether the program met the desired goals.
Learning
What is it?
Beyond satisfaction with the experience, determine if participants achieved the goals and objectives of the program (e.g., "Participant will be able to identify evidence-based practices").
How can you measure it?
Learning can be measured by conducting pre/post-surveys, interviews, and focus groups regarding the content that was covered. You can probe to:
- Participant beliefs about importance of material (e.g., "Do you think a flipped classroom is a valuable model?")
- Knowledge, skills, habits of mind (e.g., Explain what you know about a flipped classroom).
- Expertise, values, agency
Application
What is it?
Application refers to the degree to which participants apply the knowledge gained from the program in their classrooms or workshops.
How can you measure it?
- Observe a class taught by program attendees. The persistence of application could be assessed through a series of longitudinal observations. Here are recommended protocols for classroom observation:
- Classroom Observation Protocol for Undergraduate STEM (COPUS)
- Reformed Teaching Observation Protocol (RTOP)
Impact
What is it?
Has the center delivered on its large-scale goals and objectives? For example, do students have better learning outcomes due to interventions that targeted faculty?
See Also: What is Student Success?
How can you measure it?
Impact can be measured in a variety of ways, including the following:
- Student pre- and post-evaluation on:
- Content knowledge (e.g., concept inventories)
- Attitudes (e.g., CLASS survey).
- Classroom climate (e.g., The Learning Climate Questionnaire (LCQ) developed at Purdue University)
- Motivation to stay in the major
- Student retention and attrition (e.g., D/F/W, persistence in the major, career aspirations, time to graduation).
- One tool to track groups and shifts in student pathways is the Ribbon Tool from Tools for Evidence-based Action (TEA).
- Track shifts in faculty or departmental culture (e.g., Are there shifts in faculty awareness, satisfaction and attitudes toward evidence-based teaching?)
Other Metrics
Other ways of evaluating a center include documentation of output or production (e.g., funding, publications, presentations, etc.).
References
Resources, Instruments, and Tools »
Marbach-Ad et. al. (2015), A discipline-based teaching and learning center: A model for professional development. Springer International Publishing. 226 p.
Kirkpatrick, D. L. (1994). Evaluating Training Programs: The Four Levels. San Francisco: Berrett-Koehler.
Guskey, T. R. (2000). Evaluating professional development. Thousand Oaks, CA: Corwin.
Colbeck, C. L. (2003). Measures of success: An evaluator's perspective. Paper presented at the CIRTL Forum, Madison, WI.
Connolly, Mark R., and Susan B. Millar.(2006). Using workshops to improve instruction in STEM courses. Metropolitan Universities 17.4 : 53-65.