It’s good to regularly review the advantages and disadvantages of the most commonly used test questions and the test banks that now frequently provide them.
HIGHER ED TEACHING STRATEGIES FROM MAGNA PUBLICATIONS
Students are stressed. A recent survey revealed that mental health issues, including severe stress, are on the rise. In 2016, 65% of students reported experiencing overwhelming anxiety during the previous 12 months, which is an increase of more than 7% from the 2013 data (National College Health Assessment, 2016). We also know from decades of research that arousal levels are strongly related to performance: not enough arousal and you don’t perform well, but too much arousal (which becomes stress/anxiety) and your performance is negatively impacted (Colman, 2001). Therefore, anything we can do as instructors to reduce students’ stress should have a positive impact on their mental health and academic performance.
Here’s an interesting way to incorporate collaboration in a quizzing strategy, with some pretty impressive results.
Beginning with the mechanics: students took three quizzes in an introductory pharmaceutical science course. First, they completed the quiz individually. After answering each question, they indicated how confident they were that their answer was correct—5 for absolutely certain and 1 for not knowing and guessing. Then for a period of time (length not specified in the article), they were allowed to collaborate with others seated near them on quiz answers. After that discussion, they could change their quiz answers, if they desired. At that point, they again rated their confidence in the correctness of the answers. Quiz answer sheets and confidence levels were then turned in. Immediately, correct quiz answers were revealed and once again students had the opportunity to discuss answers with each other.
If there’s a downside to another academic year coming to a successful close, it’s reading course evaluations. This post explores how we respond to those one or two low evaluations and the occasional negative comments found in answers to the open-ended questions. Do we have a tendency to over-react? I know I did.
The relatively new Scholarship of Teaching and Learning in Psychology journal has a great feature called a “Teacher-Ready Research Review.” The examples I’ve read so far are well organized, clearly written, full of practical implications, and well referenced. This one on multiple-choice tests (mostly the questions on those tests) is no exception. Given our strong reliance on this test type, a regular review of common practices in light of research is warranted.
Given class sizes, teaching loads, and a host of other academic responsibilities, many teachers feel as though multiple-choice tests are the only viable option. Their widespread use justifies a regular review of those features that make these tests an effective way to assess learning and ongoing consideration of those features that compromise how much learning they promote.
Recent pedagogical interests have me wading through research on multi-tasking and revisiting what’s happening with cheating. In both cases, most of us have policies that prohibit, or in the case of electronic devices, curtail the activity. Evidence of the ineffectiveness of policies in both areas is pretty overwhelming. Lots of students are cheating and using phones in class. Thinking about it, I’m not sure other common policies such as those on attendance, deadlines, and participation are all that stunningly successful either. I’m wondering why and guessing there’s a whole constellation of reasons.
For many professors, student assessment is one of the most labor-intensive components of teaching a class. Items must be prepared, rubrics created, and instructions written. The work continues as the tests are scored, papers read, and comments shared. Performing authentic and meaningful student assessment takes time. Consequently, some professors construct relatively few assessments for their courses.
While most faculty stick with the tried-and-true quiz and paper assessment strategies for their online courses, the wide range of technologies available today offers a variety of assessment options beyond the traditional forms. But what do students think of these different forms?
Scott Bailey, Stacy Hendricks, and Stephanie Applewhite of Stephen F. Austin State University experimented with different assessment strategies in two online courses in educational leadership, and surveyed students afterward on their impressions of each one. The students were asked to score the strategies using three criteria: 1) enjoyment, 2) engagement with the material, and 3) transferability of knowledge gained to practice. The resulting votes allowed investigators to rank the various strategies from least to most preferred by students.