Racism and Sexism in Teaching Evaluations

In the world of academia, where the pursuit of knowledge and excellence in teaching are paramount, one might assume that evaluation methods would be impartial and objective. However, a thought-provoking article by David Delgado Shorter, a UCLA Professor of World Arts and Cultures, sheds light on the problematic nature of student evaluations. In his article titled “Teaching Evaluations Are Racist, Sexist, And Often Useless: It’s Time To Put These Flawed Measures In Their Place,” Shorter questions the validity and fairness of using student evaluations as a basis for academic merit and promotion decisions.

Shorter’s journey into this subject began when he reviewed his own teaching evaluations from the previous years, aiming to compile them for promotion purposes. What he found was a mixture of bizarre comments and personal narratives that had little to do with the actual course content. He realized that this was not an isolated incident; many of his Black and Asian colleagues, especially women, faced even more problematic evaluations.

These concerns prompted Shorter to delve into the research surrounding teaching evaluations. He discovered a wealth of peer-reviewed papers spanning decades, all pointing to the same disturbing trend: gender and racial biases in student evaluations. Women consistently received lower ratings than men, and younger women were often judged less professionally than their older counterparts. Women of color faced additional challenges, being rated as less effective than white women. These biases, based on gender, race, and even seemingly unrelated factors like the time of day a course was taught, raised serious questions about the validity of using student evaluations as a sole measure of teaching effectiveness.

The American Sociological Association (ASA) recognized these issues and recommended in 2019 that student evaluations should not be used as the sole basis for merit and promotion decisions unless part of a broader, more holistic assessment. Some universities, such as the University of Southern California, the University of Oregon, and the University of Nebraska at Lincoln, have already taken steps to combine student evaluations with other forms of assessment in personnel decisions. The ASA’s stance has garnered support from nearly two dozen professional organizations.

The legal implications of relying solely on student evaluations are also a cause for concern. In a case at Ryerson University (now Toronto Metropolitan University) in 2009, an arbitrator, William Kaplan, acknowledged “serious and inherent limitations” of student evaluations, describing them as “imperfect at best and downright biased and unreliable at worst.” This raises the possibility of legal challenges if colleges continue to use these evaluations as the primary criterion for decision-making.

In response to these issues, Shorter’s own department at UCLA decided to prioritize fairness and reliability. They chose not to rely on student evaluations for job security and instead implemented a system that allowed faculty members to use peer-assessment and self-evaluation, with documented revisions to pedagogical statements. This approach aligns with the principle that academics should be assessed by their peers and experts in their respective fields rather than relying solely on student evaluations.

The Problem with Rigor

“It’s Time to Cancel the Word Rigor,” read a recent headline in education press.[1]  The article detailed growing concerns about hidden bias within what many see as conventional teaching practices. Here, “rigor” was taken to task for throwing roadblocks up for some students more than others, even as its exact meaning remains vague. Webster’s Dictionary defines rigor as “severity, strictness or austerity,” which educators often translate into difficult courses and large amounts of work, rationalized in the interest of excellence and high standards.[2]

While there is nothing wrong with challenging coursework, per se, this interpretation of rigor often becomes a recipe for failure for otherwise intelligent and hardworking students.  Such failures can result when rigor is used to incentivize or stratify students, as in gateway or “weed out” courses with prescribed grading targets, or situations where faculty overuse tests as motivation. Rigor discussions I have witnessed rarely consider instructional quality, teaching effectiveness, or principles of learning. Instead faculty complain about poor student attention, comprehension, or commitment. As the Chronicle explains, “all credit or blame falls on individual students, when often it is the academic system that creates the constructs, and it’s the system we should be questioning when it erects barriers for students to surmount or make them feel that they don’t belong.”[3] Continue reading “The Problem with Rigor”

The Algorithm Rejected Me

School is where most kids first become aware of what I call the  “update imperative.”  After all, education is a process continual improvement in a step-by-step process of knowledge acquisition and socialization. In this sense schooling represents much more than the beginning of education. For many kids it’s a time of moving from the familiarity of home into the larger world of other people, comparative judgement, and a system of tasks and rewards. Along the way, a package of attitudes and beliefs is silently conditioned: conformity to norms, obedience to authority, and the cost of failure. All of this is presented with a gradually intensifying pressure to succeed, rationalized as a rehearsal for adult life. Rarely are the ideological parameters of this “hidden curriculum” ever challenged, or even recognized. Much like work, American K-12 schools are driven largely by mandates of individual achievement and material accumulation.

By the time college applications are due, levels of anxiety can run out of control, given the role of degrees in long term earnings.  Many students start the admissions Hunger Games as early as middle school, plotting their chances, polishing their transcripts, and doing anything they can to get good grades. Everyone knows how admissions data now flows in an age in which students apply to an average of 10 schools each. Unsurprisingly perhaps, overall applications have increased by 22% in the past year alone.[i] And while the applicant side of this equation has been much publicized, what happens in the admissions office remains shrouded in mystery. Largely unknown are secret criteria driven by algorithms to determine things like likelihood to enroll or willingness to pay. Even less known are kinds of AI analytics used to monitor and grade students, sometimes making prejudicial judgements along the way. Continue reading “The Algorithm Rejected Me”