How and Why Colleges Should Reform Student Evaluations

BY DAVID A. VAREL

When I was an undergraduate at a liberal arts college in the early 2000s, our student evaluations were all qualitative. We were asked to write short essays describing our experience in our courses—what worked, what didn’t, what could be improved. It was clear from this design that the chief audience was our professors, who would use them to create better courses.

colorful emoticons showing happy and sad expressionsThese days, evaluations have become increasingly caught up in the neoliberal model of education. They are more quantitative in nature, less substantive, and more clearly intended for administrators to justify actions towards faculty members regarding retention, promotion, and tenure.

This model fails instructors, and it fails our students. Luckily, some simple changes to the questions could better measure effective teaching, promote better instruction, and even undermine neoliberal education. Here’s how.

The worst evaluation questions are the most subjective ones: How would you rate the difficulty of this course? How would you rate the instructor’s effectiveness? How would you rate your overall experience in this course? Because no objective criteria exist for such questions, students create their own subjective criteria and responses. This is all wrong. These types of questions, furthermore, encourage students to view themselves as consumers assessing the “product” they received—indeed, they are eerily similar to surveys sent out to customers of major corporations.

Consider how more objective questions can reframe an evaluation around tangible qualities and teaching practices that are hallmarks of effective teaching. All of these can retain the quantitative component by having students rank-order their responses on a scale (like 1–4 or 1–6) regarding how much a professor practiced that particular technique. For instance, start with fundamental questions at the heart of any well-organized course:

  • Alignment: To what degree were assignments rooted in material covered in class? (1. not at all, 2. a little, 3. a significant amount, 4. a great deal)
  • Diverse Assessments: To what degree did the instructor incorporate both formative assessments (more frequent, lower-stakes assignments meant to build student skills) and summative assessments (less frequent, higher-stakes assignments meant to test students’ mastery of cumulative knowledge and skills) (1. not at all, 2. a little, 3. a significant amount, 4. a great deal)
  • Differentiated Instruction: To what degree did the instructor incorporate different types of instruction during class—that is, lecture, whole-class discussion, small-group discussion, debate, media presentations? (1. not at all, 2. a little, 3. a significant amount, 4. a great deal)

Evaluations should then probe the details of the instruction, which are also essential to creating a successful course:

  • When did the instructor learn students’ names? (1. early on, 2. mid-semester, 3. end of semester, 4. never)
  • How long, on average, did it take for the instructor to grade assignments? (1. less than a week, 2. 1-2 weeks, 3. 2-3 weeks, 4. a month or more)
  • How much individualized feedback did the instructor provide on course assignments? (1. none, 2. a little, 3. a significant amount, 4. a great deal)

Similar questions could be created for other important qualities related to the timeliness of responses to emails, the instructor’s punctuality and use of the entire class period, the clarity of course objectives, the level of respect and empathy demonstrated toward students, the willingness to accommodate students’ special circumstances, the clarity of assignment requirements and expectations, the provision of resources to help students improve, and students’ development of specific knowledge and skills, to name a few.

There is, of course, much room for debate over the specifics of each question and which responses may best reflect quality teaching or even be appropriate. The nature and size of courses must certainly be taken into account, and differences will naturally emerge among disciplines.

Yet the bigger point remains: We should create evaluations that measure the concrete practices of effective teaching rather than the subjective feelings of students—which we know are biased along racial, gender, and other lines.

The result could be quietly revolutionary. Students would be forced to consider what the hallmarks of good teaching are, and perhaps even recognize that a course they did not like was administered competently. They would be discouraged from seeing themselves as consumers and encouraged, rather, to identify as pupils owed not satisfaction and ease but competency, clarity, and professionalism.

The impact could be no less transformative for instructors, especially because the education-as-commodity model has inspired some deeply problematic behaviors. Professors, responding to the serious stakes that evaluations can have on their often-precarious careers, are incentivized to try to game the system.

Anecdotes abound. I know of professors who give out cookies (which apparently works!) or announce serious changes in course expectations—like an optional final exam—right before handing out evaluations. Such unprofessional attempts at placation send all the wrong messages to students, and they undermine whatever utility the evaluations could hold in comparing the effectiveness of different instructors. More troubling, however, are the subtler ways in which professors adapt their teaching—by scaling back the rigor, inflating grades, and generally trying to please students rather than challenging them.

Reformed evaluations, conversely, would redirect instructors’ energies into becoming better teachers. Now they would be rewarded for organizing better courses and practicing the behaviors that foster student engagement and growth. The evaluations would in fact have educational value because many professors have received precious little training on how to plan or teach a course well. Knowing they will be evaluated for it, professors would be encouraged to learn students’ names, to provide timely and individualized feedback, to align their curriculums, and to adopt many of the other best practices that could improve instruction.

Although such evaluative changes are obviously no panacea, they are easy-to-implement reforms that would go a long way toward counteracting some of the problems characterizing student evaluations today.

David A. Varel is an adjunct professor of history at Metropolitan State University of Denver, a licensed public high-school teacher, and the author of two books: The Scholar and the Struggle: Lawrence Reddick’s Crusade for Black History and Black Power, and The Lost Black Scholar: Resurrecting Allison Davis in American Social Thought.

4 thoughts on “How and Why Colleges Should Reform Student Evaluations

  1. Prof. Varel offers some excellent suggestions for improving student evaluations of courses and instructors, and I would add one, more informal suggestion based on my own experience over nearly 40 years. I pass around note cards once a week and ask students to give me feedback. They can make suggestions about anything, from the volume of my lectures to assigned readings, subjects we cover (or don’t), the temperature in the classroom, etc. I try to incorporate their suggestions and feedback as we go through the semester: there’s no point in telling me once the term is over that I didn’t speak loudly enough…. It’s all anonymous, and my teaching assistants pass out and collect and pass along student feedback so I have no idea who said what or complained about something. Students have always appreciated the fact that I care enough about their assessment of me and my course, and they highlight this in comments on the ‘formal’ evaluations at the end of the term. It’s a bit of work in a large lecture course, but not everyone has something to write every week, so it turns out to be manageable. I have also noticed over the years that students will raise complaints each week — usually minor, always polite — but little or no praise, but at the end of the term they are more likely to praise the course on the official evaluation.

    • Faculty unfriendly largely quantitative evaluations have been the norm for my almost 50 year teaching career. The problem is not new. Very seldom are teaching evaluations–written, form surveys, or inperson–used constructively. This is the part of what I now call “the myth of collegiality and the contradictions of academic life.”

      Matters worsened with the advent of the national corporatized RateMyProfessor. The worst of the worst!

      I did discover that retired profs like me and others who have left teaching can have their records removed from free access by asking.

      • Thank you Harvey + J + Graff.

        “the myth of collegiality and the contradictions of academic life.”

        I am a retired (35 years) admin staff worker in an academic department in a large mid-western public university. I am reminded of a dicey tenure promotion where one of the already tenured faculty in the department insisted that all, over 500 pages,of a junior assistant professor’s evaluations be put in their tenure file to be considered by the department executive committee, This was unprecedented, not one other tenured faculty member in the department had been required to do the same.The protocol was that the evaluations were tallied by number and comments and were compiled in a 6 year summary format.The evaluation form in and of itself was as noted here, for the most part useless in evaluating the faculty member, but was being used to go against one faculty member in an otherwise before that, collegial department. This was the beginning of the end of a department that served students and resisted administration.

        Evaluations need to be changed, some good ideas here, but these ideas won’t solve the problem. Ask questions students can answer. Don’t ask questions another faculty member can answer and administrators can use in opposition to their colleagues.

    • Why is it that faculty and administrations are afraid to acknowledge that student/responders are a diverse group with varied educational backgrounds, experiences, motivations and interests. We pretend that student/responders are a monolithic and objective group. One student’s response to any question is not the shared experience of another. Is the meaning of a 4 out of 7 (on the likert scale), the same for all students? I believe students have a right to express their ideas and thoughts about the instructor, but not through a forced questionnaire (questions decided upon by administrations). Instead, students should be compelled to write their thoughts down and show the reasons for their complaints, problems, what they liked, disliked. Moreover, this is one of the reasons why student common rooms were created, so that they could express themselves with their student colleagues. Let students write critical reports and let their unions and organizations provide the platform. Student evaluations should never be used to decide promotion and tenure. Trying to appease every student and their values is a road to high marks, lower standards, and simply focusing on the popularity of being “nice and accommodating.”

Comments are closed.