If you are using a tablet, please flip to landscape mode. If you are using a smartphone, please use desktop mode.
If you are using a laptop or a desktop PC, please maximize your window.
Thanks for your patience as I continue to build this website.
4/8/18
At this point, talk like: “The
internet is revolutionizing the way that people communicate across the world!”
has become dreadfully trite. Society as a whole is so
accustomed to the internet as a way of life that people rarely stop to think
about how far-reaching the web has truly become. Almost every decision we make
nowadays, ranging from the life-changing (what career path to go down, where to
live) to the mundane (where to eat tonight, how much is too much for a pack of
staples), we make using information that we find online. Often, this
information comes in the form of reviews.
Review websites have been a staple of the
internet almost since the beginning, with the launch of ConsumerAffairs.com in
1998. Websites like ConsumerAffairs
saw great success, quickly becoming invaluable to people making purchase
decisions. Seeing an opportunity, John Swapceinski, a
student at San Jose State University, created TeacherRatings.com in 1999 after
having a bad experience with a professor1. This simple website extended the
internet-review concept to university professors, creating a similarly
invaluable tool for students deciding what classes to take and which professors
to avoid. According to the mission
statement, “The site does what students have been doing forever - checking
in with each other – their friends, their brothers, their sisters, and their
classmates.”
In 2001 the website was renamed RateMyProfessors.com
and has become a fixture at college campuses in the time since. The rapid
growth of this website has seen a backlash from many critics, from professors
to journalists to even some students themselves. They cite “untrustworthy
reviews,” and they say that “Students should not base decisions about their
education on it … and professors should not get ideas from it.” They say that the
reviews do not “represent the wisdom of crowds
2” and that “only students with the strongest
opinions – either for or against – post on [the] website.3” After investigating these claims, I have found
the general consensus that “RateMyProfessors is inaccurate and does not matter in the
real world” to be far from sound. Before I debunk some of these common myths,
however, it is important to know why professor reviews are so important in the
first place.
HOW DOES RATEMYPROFESSORS AFFECT THE STUDENT EXPERIENCE?
“It’s actually influenced my class choices,” a
USC student said. “I definitely prefer [knowing how professors are], because
then I know what I have to do to do better [in the course]. So
if I know that one professor is a really hard teacher, then I know from the
beginning to stay on top of the readings4.” This student is not the only one who is
strongly influenced by professor ratings. In a Texas A&M University study, 104
students were given two hypothetical professors (A and X): one with positive
reviews, and one with negative reviews. They measured how attracted the student
was to the professor (note: attraction does not mean physical attraction in
this context, but a measure of how much students favor the professor based on
reviews) using the following questions: “How likely are you to choose a class taught
by Professor (A/X), how likely are you to recommend this professor to another
student, how likely are you to be satisfied with the numerical rating of
Professor (A/X), and how likely are you to put this class as your first
priority?5” The results showed, unsurprisingly, that “students
[who] received a positive feedback review were more attracted to the
professor.”
Clearly, RateMyProfessors has done a lot
to alter whether students decide to take a professor. What’s more, if they do
decide to take a professor’s course, studies show a marked change in how students
who have looked at reviews experience the course. In a Western Michigan
University study, undergraduates were presented “with ratings and comments
about a professor, [shown] a 10-20 minute lecture by
him, and assessed [on] their perceptions of the educational experience6.” As in the Texas A&M study, neither the
professor nor the reviews were real, although in this study participants were not
told that the professor in question was fictitious. Importantly, this study had
students watch a lecture after reading the reviews to see the professor for
themselves. Interestingly, even though all students watched the same taped
lecture, students who were shown positive reviews “reported higher levels
of affective learning (i.e., “liking” of the
lecture material) and motivation to learn” then their counterparts who had
viewed negative reviews, mixed positive/negative reviews, or no reviews at all.
Students who were shown negative reviews reported the lowest levels of
“affective learning.” Even more surprisingly, the study also found that “students
who read positive ratings before viewing the lecture performed significantly
better on a quiz [of] the content (a letter grade higher, on average) and
reported a greater likelihood of actually engaging in the behaviors the
professor recommended.” This suggests that reviews hold much more sway over the
experiences of students than most people would expect.
If the results of these two studies are looked
at together, the trends show that positive reviews cause students to be more
“attracted” to professors, and that this attraction creates a positive
perception of the professor. According to the Western Michigan study, this
positive perception actually translates into tangible
benefits like increased interest in the content, more motivation to learn, and
even significantly better grades when tested on the material. The psychology
involved here is fascinating and it shows just how crucial reviews from other
students can be in a student’s experience of a teacher.
With this in mind, it makes sense that bad teacher ratings
seriously damage both a professor's reputation among students and the academic
performance of students in those classes. It is also very interesting to note
that these reviews can create a positive feedback loop, in which positive
reviews cause a student to have positive perceptions of a professor, thereby
creating more positive reviews7.
DOES RATEMYPROFESSORS AFFECT A TEACHER’S STANDING WITH THE UNIVERSITY?
Although RateMyProfessors reviews have a tectonic
influence on how students see their professors, there is little evidence
showing that these reviews affect how universities evaluate their professors.
Andrew Coile, former Computer Science professor at
California State University and current Apple Inc. employee, says on an online
forum that “For faculty RTP (Retention, Tenure, and Promotion) process, the
official student evaluations were the only ones considered8.” For context, most universities conduct
official student evaluations at the end of every semester. These are used by
the university to determine things like faculty tenure, and by professors
themselves as constructive feedback. Most universities keep the results of
these evaluations private, although there has been a recent movement to
publicize them9 (more on this later). Other professors on the
forum echo the assertion that universities do not look at RateMyProfessors when discussing
hiring, promotion, or tenure. It is worth noting, however, that official course
evaluations may be indirectly affected by RateMyProfessors reviews given
the positive feedback loop effect discussed above, and if this is true, then RateMyProfessors could
have a subtle effect on a professor’s standing with his university.
ARE RATEMYPROFESSORS REVIEWS ACCURATE?
Although universities themselves do not use
them, ratings posted on RateMyProfessors.com clearly have a strong bearing on
professors’ reputations among students. Given that this is the case, it is
important to ask whether it is fair for online reviews to impact professors’ perceived
teaching ability in such a significant way. Whether or not it is reasonable for
online reviews to sway professor reputations depends of the overall accuracy of
these reviews. While accuracy of subjective comments can never be perfectly
measured, analysts have been able to gain access to official university course
evaluations as a benchmark to compare against.
The results of these benchmarks are
surprising. One study published in the Journal
of Education for Business includes data from 5 different universities and
1,167 faculty members on many different facets of RateMyProfessors ratings10. The study finds, among other things, that RateMyProfessors
summary ratings actually “correlate highly” with official course evaluation
summaries. A Lander University study was able to replicate these results:
“Easiness website ratings were significantly positively correlated with actual assigned
grades. Further, clarity and helpfulness website ratings were significantly
positively correlated with … the institutionally administered … forms[11].” This same correlation was found in other
studies as well12.
Although accuracy relies on many factors, these studies at least provide evidence
that RateMyProfessors
rankings generally mirror the results of official course evaluations, which university
administrations accept as accurate. This of course will vary from professor to
professor depending on how many reviews have been left. Some researchers have
found that 10 reviews is a good number to get a rough consensus13,
and some have found accuracy with only a single review14, although more is always better.
UNIVERSITIES SHOULD RELEASE
COURSE EVALUATIONS
Although RateMyProfessors already has
general accuracy when compared to university course evaluations, one way to
improve accuracy further would be for universities to release their own course
evaluations publicly. The way I see it, universities have no excuse not to disclose
this data. Most universities do not give reasons for withholding course
evaluation results, but the few arguments that I could find hardly hold water. At
USC, Vice Provost for Academic and Faculty Affairs Elizabeth Graddy cites “University privacy concerns” and “incivility”
in online responses15. To be fair, this is a good argument for not
releasing specific student responses to open-ended questions, but course
evaluations typically include few open-ended questions if any at all. In my experience,
course evaluations are mostly done on bubble sheets with 10-20 multiple choice questions
about the class and professor. The privacy argument makes little sense against
releasing multiple choice data.
Another argument Graddy
gives against disclosing course evaluations is that her department is “trying
to figure out how we can improve our instrument” because the evaluations found
a “a bias against women and faculty of color.” Interestingly, the Faculty
Senate at the University of Michigan cites the exact same reason as an argument
against releasing course evaluations16. The fact that two universities found the
same race/gender bias in their course evaluations likely means that these
skewed results are not a problem with the “instrument,” but rather an
indication of an overarching issue that needs to be addressed. Trends like this
are actually a good reason why the data should be
released because it can facilitate academic discussion about the issue. Keeping
data under lock and key simply because the results are unpleasant is not
something that the scientific community tolerates, and prestigious universities
like UMichigan and USC need to set an example. To UMichigan’s credit, some of their course data was
eventually released to students a year later17. If you are interested, there are many more bulleted
arguments against releasing course evaluations on this
page. I will not address all of them here to for the sake of brevity and
because I feel that the remaining arguments are inferior to the ones already
discussed.
CONCLUSION
Universities should absolutely
release course evaluations, but for universities that don’t, RateMyProfessors is
the only place for students to get intel on a professor that they are
unfamiliar with. Because of this, I say that students should go ahead and look
at the reviews if they want to. Despite common belief, these students are actually relatively safe looking at RateMyProfessors reviews, because
they are fairly accurate. They should be aware, however, of the effects (positive
and negative) that reviews can have on student psyche and performance. RateMyProfessors
is certainly not
perfect. Some potential problems include the fact that deceitful professors
can underhandedly rate
themselves and that students can submit more than one rating. Because
analysts have found reviews to be generally accurate however, these are likely
only problems in rare, isolated cases. Most professor pages with at least a few
reviews have a high chance of being trustworthy. I hope that all
of this knowledge is helpful to you in making an educated decision about
courses. Good luck in your professor-hunting and may the enrollment Gods ever
be in your favor!
[2] https://www.nytimes.com/2010/03/14/magazine/14FOB-medium-t.html
[3] https://sites.sju.edu/blogs/2015/11/05/7-reasons-why-i-hate-rate-my-professor/
[4] https://dailytrojan.com/2018/02/20/rate-professors-influences-course-choices/
[5]https://static1.squarespace.com/static/5443d7c7e4b06e8b47de9a55/t/59b49ca6f7e0abec5d24cd2b/1505008807272/Who-Needs-Rate-My-Professor3.pdf
[6] https://www.natcom.org/communication-currents/instructors-corner-1-think-RateMyProfessorscom-doesnt-matter-think-again
[8] https://academia.stackexchange.com/questions/101009/do-professors-care-about-their-rating-on-websites-such-as-RateMyProfessors-or-koo
[9] https://dailytrojan.com/2018/02/20/rate-professors-influences-course-choices/
[13] http://pareonline.net/pdf/v15n5.pdf
[14] https://www.tandfonline.com/doi/abs/10.3200/JOEB.84.1.55-61?src=recsys
[15] https://dailytrojan.com/2018/02/20/rate-professors-influences-course-choices/
[16] https://record.umich.edu/articles/faculty-senate-passes-resolution-course-evaluation-data
[17] https://www.michigandaily.com/section/news/art-20-gives-students-more-information-about-courses