All enterprises require measurement in order to enable their management. Testing, grading and evaluation of education programs are the metrics by which we measure students, teachers and schools.
I have a confession to make. I was wrong. You see, I once thought that teaching was lecturing, and I thought that because that is how my graduate mentors taught me to teach.
But I was wrong. Studies have shown that lecturing has little to do with teaching. A University of Maryland study found that right after a physics lecture, almost none of the students could answer the question: “What was the lecture you just heard about?” Another physics professor simply asked students about the material that he had presented only 15 minutes earlier, and he found that only ten percent showed any sign of remembering it (Freedman, 2012).
Measuring student success is a top priority to ensure the best possible student outcomes. Through the years instructors have implemented new and creative strategies to assess student learning in both traditional and online higher education classrooms. Assessments can range from formative assessments, which monitor student learning with quick, efficient, and frequent checks on learning; to summative assessments, which evaluate student learning with “high stakes” exams, projects, and papers at the end of a unit or term.
April 4 - A New Way to Assess Student Learning
I’m “reflecting” a lot these days. My tenure review is a few months away, and it’s time for me to prove (in one fell swoop) that my students are learning. The complexity of this testimonial overwhelms me because in the context of the classroom experience, there are multiple sources of data and no clear-cut formula for truth.
December 19 - A Quiz Design that Motivates Students
Many faculty members use quizzes to keep students prepared and present in class. The approach often tends to be punitive, however, motivating students by extrinsic means. Karen Braun and Drew Sellers, who teach beginning accounting courses, wanted to use quizzes in the usual ways—to get students coming to class having done the reading, to arrive in class on time, and to participate in class discussion, but they wanted their quizzes to be more about intrinsic motivation and less about assessment. How did they achieve that objective? They incorporated a number of “motivational” design features into their use of quizzes.
November 22 - Students, Studying, and Multiple-Choice Questions
Multiple-choice questions are not the pariah of all test questions. They can make students think and measure their mastery of material. But they can also do little more than measure mastery of memorization. Memorizing is usually an easier option than thinking and truly understanding.
When you are a math teacher you are often faced with the dilemma of whether to assign partial credit to a problem that is incorrect, but that demonstrates some knowledge of the topic. Should I give half-credit? Three points out of five? My answer has typically been to give no credit…at first. However, taking a page from my colleagues in the English department (and grad school), I do allow for revisions, which ends up being a much better solution.
October 15 - A Different Kind of Final
Last semester I implemented a different kind of final exam. In the past I have used the standard multiple-choice and short-answer exams. I was thinking about making a change when I discovered Beyond Tests and Quizzes: Creative Assessment in the College Classroom, edited by Richard J. Mezeske and Barbara A. Mezeske. The second chapter, “Concept Mapping: Assessing Pre-Service Teachers’ Understanding and Knowledge,” describes an assessment method that tests higher-level thinking. The author shared his experience using concept maps as a final exam, included an example of the final exam project, offered rubrics for grading, and discussed the advantages and disadvantages of the strategy. I decided this was the change I was going to make.
There are two main forms of assessment often used within the online classroom. Both formative and summative assessments evaluate student learning and assist instructors in guiding instructional planning and delivery. While the purpose of a summative assessment is to check for mastery following the instruction, formative assessment focuses on informing teachers in ways to improve student learning during lesson delivery (Gualden, 2010). Each type of assessment has a specific place and role within education, both traditional and online.
We give students exams for two reasons: First, we have a professional responsibility to verify their mastery of the material. Second, we give exams because they promote learning. Unfortunately, too often the first reason overshadows the second. We tend to take learning outcomes for granted. We assume the learning happens, almost automatically, provided the student studies. But what if we considered how, as designers of exam experiences, we might maximize their inherent potential? Would any of these possibilities make for more and better learning from the exams your students take?
The guidelines suggested below propose how critical thinking skills can be assessed “scientifically” in psychology courses and programs. The authors begin by noting something about psychology faculty that is true of faculty in many other disciplines, which makes this article relevant to a much larger audience. “The reluctance of psychologists to assess the critical thinking (CT) of their students seems particularly ironic given that so many endorse CT as an outcome…” (p. 5) Their goal then is to offer “practical guidelines for collecting high-quality LOA (learning outcome assessment) data that can provide a scientific basis for improving CT instruction.” (p. 5) The guidelines are relevant to individual courses as well as collections of courses that comprise degree programs. Most are relevant to courses or programs in many disciplines; others are easily made so.