Teachers, Cheating, and Incentives
In recent years there seems to have been a surge in academic dishonesty in high schools. No doubt this can be explained in part by 1) increased vigilance and reporting, 2) greater pressure on students to succeed, and 3) the communicable nature of dishonest behavior (when people see others do something, whether it’s enhancement of a resume or parking illegally, they’re more likely to do the same). But, I also think that a fourth, and significant, cause in this worrisome trend has to do with the way we measure and reward teachers.
To think about the effects of these measurements, let’s first think about corporate America, where measurement of performance has a much longer history. Recently I met with one of the CEOs I most respect, and he told me a story about when he himself mismanaged the incentives for his employees, by over-measurement. A few years earlier he had tried to create a specific performance evaluation matrix for each of his top employees, and he asked them to focus on optimizing that particular measure; for some it was selection of algorithms, for others it was return on investment for advertising, and so on. He also changed their compensation structure so that 10 percent of their bonus depended on their performance relative to that measure.
What he quickly found was that his top employees did not focus 10 percent of their time and efforts on maximizing that measure, they gave almost all of their attention to it. This was not such good news, because they began to do anything that would improve their performance on that measure even by a tiny bit—even if they caused problems with other employees’ work in the process. Ultimately they were consumed with maximizing what they knew they would be measured on, regardless of the fact that this was only part of their overall responsibility. This kind of behavior falls in line with the phrase “you are what you measure,” which is the idea that once we measure something, we make it salient and motivational. In these situations people start over-focusing on the measurable thing and neglect other aspects of their job or life.
So how does this story of mis-measurements in corporate America relate to teaching? I suspect that any teachers reading this see the parallels. The mission of teaching, and its evaluation, is incredibly intricate and complex. In addition to being able to read, write, and do some math and science, we want students to be knowledgeable, broad-minded, creative, lifelong learners, etc etc etc. On top of that, we can all readily agree that education is a long-term process that sometimes takes many years to come to fruition. With all of the complexity and difficulty of figuring out what makes good teaching, it is also incredibly difficult to accurately and comprehensively evaluate how well teachers are doing.
Now, imagine that in this very complex system we introduce a measurement of just one, relatively simple, criteria: the success of their students on standardized tests. And say, on top of that, we make this particular measurement the focal point of evaluation and compensation. Under such conditions we should expect teachers to over-emphasize the activity that is being measured and neglect other aspects of teaching, and we have evidence from the No Child Left Behind program that this has been the case. For example, we find that teachers teach to the test, which improves the results for that test but allows other areas of education and instruction (that is, those areas not represented on the tests) to fall by the wayside.
And how is this related to dishonesty in the school system? I don’t think that teachers are cheating this way (by themselves changing answers, or by allowing students to cheat) simply to increase their salaries. After all, if they were truly performing a cost-benefit analysis, they would probably choose another profession—one where the returns for cheating were much higher. But having this single measure for performance placed so saliently in front of them, and knowing it’s just as important for their school and their students as it is for their own reputation and career, most likely motivates some teachers to look the other way when they have a chance to artificially improve those numbers.
So what do we do? The notion that we take something as broad as education and reduce it to a simple measurement, and then base teacher pay primarily on it, has a lot of negative consequences. And, sadly, I suspect that fudging test scores is relatively minor compared with the damage that this emphasis on tests scores has had on the educational system as a whole.
Interestingly, the outrage over teachers cheating seems to be much greater than the outrage over the damage of mis-measurement in the educational system and over the No Child Left Behind Act more generally. So maybe there is some good news in all of this: Perhaps we now have a reason to rethink our reliance on these inaccurate and distracting measurements, and stop paying teachers for their students’ performance. Maybe it is time to think more carefully about how we want to educate in the first place, and stop worrying so much about tests.
(This post also appeared as part of a leadership roundtable on the right way to approach teacher incentives in the Washington Post. The Washington Post will post more opinions about this topic here. )
Conflicts of Interest in Dentistry
According to an article in SmartMoney, as many as 48% of U.S. dentists have seen their profits plummet thanks to the recession.
In and of itself, this isn’t a particularly remarkable statistic – after all, most of our wallets have taken a hit this past year – but what follows is an interesting discussion: how are dentists coping with this drop in income? Angie C. Marek reports a variety of tactics in her article (including lowered rates, freebies, eliminated IOUs, etc.), most of which benefit the patient – but they don’t all. Some dentists are softening the financial blow by upselling and overtreating patients.
One example came from a woman who, upon switching cities and dentists, was surprised to learn that her hitherto problem-free mouth was suddenly rife with problems: several cavities required coatings and two veneers needed replacement. Or so her dentist told her. However, this turned out to be just another case of overtreatment.
The problem here is conflicts of interests (COIs), which are instances when professionals are pulled in two directions, torn between personal gain and the good of the patient. And the sad news is that when faced with COIs dentists (or physicians) sometimes end up going the self-interested route, and this can have undesirable consequences for the patient.
Conflicts of interest are nothing new, they have been a problem for as long as there have been professions, and they are very pervasive. For instance, there’s the doctor who at accepts consulting fees from a drug company and studies their drug, the one who prescribes the treatment a drug rep pushed on him the week before over a free lunch, and even the doctor who urges a treatment on a patient in part so that he can use his costly new medical equipment.
This isn’t to say that these are dishonorable people who only see dollar signs and say to hell with the patient. Rather, COIs can deeply color the person’s perception, and thereby end up leading even the most upstanding citizens astray, and this happens often.
The long and short of it is, next time you are at the dentist’s office – think about your dentist’s conflicts of interests.
Classroom Ethics 101
On the first day of one of my classes, I asked my undergraduate students whether they had enough self-control to avoid using their computers during class for non-class-related activities. They promised that if they used their laptops, it would only be for course-related activities like taking notes. However, as the semester drew on, I noticed more and more students checking Facebook, surfing the web, and emailing. And I noticed that as these behaviors increased, so did their cheating on weekly quizzes. In a class of 500 students, it was difficult to manage this deterioration. As my students’ attention and respect continued to degrade, I became increasingly frustrated.
Finally, we got to the point in the semester where we covered my research on dishonesty and cheating. After discussing the importance of ethical standards and honor code reminders, two of my students took it upon themselves to run something of an experiment on the rest of the students. They sent an email to everyone in the class from a fabricated (but conceivably real) classmate, and included a link to a website that was supposed to contain the answers to a past year’s final exam. Half the students received this email:
———- Forwarded message ———-
From: Richard Zhang <richardzhang44@gmail.com>
Subject: Ariely Final Exam Answers
To:
Hey guys,
Thought you might find this useful. See link below.
——————————————-From: Ira Onal<ira.onal@gmail.com>
To: Richard Zhang < >
Subject: Re:Hello!
Hey Richard,
Good to hear from you again. Yes, I was the TA for Ariely’s class. Here’s a link with the answers from the test when I was TA, and I don’t think he changes the questions/answers every semester. Hope this is helpful and let me know if you have any questions:
http://dl.dropbox.com/u/26171004/iraonal.html
Best of luck,
Ira
Ira T. Onal
Duke University Trinity School ’09
ira.onal@gmail.com | (410) 627-0299
Richard Zhang < > wrote:
> Hey Ira,
> I hope all is going well. I’m in Ariely’s class and saw your name on the syllabus – are you/were you the TA? I also heard there is an exam in the class, and was wondering if you had any guidance/tips for it. He just has a bunch of short quizzes this year, so should I use those to study from?
>Best,
Richard
–
Richard Zhang
Duke University ’12
(315) 477-1603
——————————————-
The other half got the same email but also included the following message:
~ ~ ~
P.S. I don’t know if this is cheating or not, but here’s a section of the University’s Honor Code that might be pertinent. Use your own judgment:
“Obtaining documents that grant an unfair advantage to an individual is not allowed.”
~ ~ ~
Using Google Analytics, the students tracked how many people from each group visited the website. The disparaging news is that without the honor code reminder, about 69% of the class accessed the website with the answers. However, when the message included the reminder about the honor code, 41% accessed the website. As it turns out, students who were reminded of the honor code were significantly less likely to cheat. Now, 41% is still a lot, but it is much less than 69%.
The presence of the honor code, as well as the ambiguity of the moral norm, may have had a role in the students’ behavior. When the question of morality becomes salient, students are forced to decide whether they consider their behavior to be cheating – and presumably most of them decided that it is.
Moreover, a qualitative look at the email responses from students (to the fictitious student who sent them the link to the test answers) showed that while those who did not see the code were generally thankful, those given the honor code were often upset and offended.
***
The issue of cheating arose again with the approach of finals. I received several emails from students who were concerned about their classmates cheating, so I decided to look into the situation with a post-exam survey. The day after the exam, I asked all the students to report (anonymously) their own cheating and the cheating they suspected of their peers.
The results showed that while the students estimated that ~30-45% of their peers had cheated on the final exam, very few of them admitted that they themselves had cheated. Now, you might be thinking that we should take these self-reports with a grain of salt – after all, even on an anonymous survey, students will most likely underreport their own cheating. But we can also look at the grades on the exam, and because less than 1% of students got a 90% or better (and the average got 70% correct), I am relatively confident that the students’ perception of cheating was much more exaggerated than the actual level (or they could just be very bad at cheating).
While it may seem like good news that fewer of their peers cheat than they suspect, in fact such an overestimation of the real amount of cheating can become an incredibly damaging social norm. The trouble with this kind of inflated perception is that when students think that all of their peers are cheating, they feel that it is socially acceptable to cheat and feel pressured to cheat in order to keep up. In fact, a few students have come to my office complaining that they were penalized because they decided not to cheat — and what was amazing to me was that in being honest, they truly felt that there was some injustice done to them.
The bottom line is that if people perceive that cheating is running rampant, what are the chances that next year’s students will adopt even more lenient moral standards and live up to the perception of cheating among their peers?
An Extreme Take on The Ten Commandments Experiment
We once ran a study on cheating where we asked students to try to recall the Ten Commandments before an exam, and found that this moral reminder deterred them from cheating. Well, a professor at Middle Tennessee State University recently made practical use of the study – but in an extreme way.
Fed up with the low ethical standards among his MBA students, Professor Michael Tang passed out an honor pledge that not only listed the Ten Commandments, but also included a concluding flourish indicating that those who cheated would “be sorry for the rest of [their] life and go to Hell.” In response, several students called the department chair to complain and a good deal of controversy ensued.
But what the news coverage didn’t address (perhaps because no one at the school had) were the merits of this extreme pledge. Might this be an effective way of curbing dishonesty? I think yes, very much so. I also suspect that even those who don’t believe in God would take this pledge seriously.
Still, though I don’t doubt its effectiveness, the question remains whether we want to invoke such stringent punishments (stringent for those who believe, that is) on an MBA exam. Judging from the reactions in this case, I’m guessing that for most people, the answer is “no.” But it also makes me wonder about the people who didn’t want to sign this pledge….
A Dinner with Drug Reps
Over the years I’ve written all sorts of blog posts on dishonesty, and because the new book release, I want to repost an updated version of them to accompany. For the next few days I’ll post one every other day. Enjoy!
Janet Schwartz of Tulane University and I once spent an evening with a few former pharmaceutical reps, men who used to be in the business of selling a wide range of drugs to treat all kinds of diseases and conditions, from fibromyalgia to depression to restless leg syndrome. As drug representatives, they would go from doctor to doctor attempting to convince physicians to prescribe their company’s drugs. How? Typically they would start by passing on informative pamphlets and giving out products like pens, clipboards, and notepads advertising their drugs.
But we knew there was more to the story, so we tried the pharmaceutical reps at their own game – we took them to a nice dinner and kept the wine flowing. Once we got them a bit sauced, they were ready to tell all. And what we learned was fairly shocking.
Picture these guys: attractive, charming young men. Not the kind of guys who would have trouble finding a date. One of them told us a story about how he was once trying to persuade a reluctant female physician to attend a seminar about a medication he was promoting. After a bit of persuading, she finally decided to attend – but only after he agreed to escort her to a ballroom dancing class. This, according to our new friends, was a typical kind of quid pro quo where the rep does a personal favor for the doctor and the doctor promotes the rep’s product in return.
Another common practice was to bring meals to the doctor’s office (one of the perks of being a receptionist), and one office even required alternating days of steak or lobster for lunch in exchange for access to the (well-fed) doctors.
Even more shocking was that when the reps were in the physician’s office, they were sometimes called into the examination room (as “experts”) to inform the patients about the drug directly. And the device reps experienced a surprisingly intimate level of involvement in patient care, often selling medical devices in the operating room, while the surgery was going on.
Aside from learning about their profession, we also realized how well these pharmaceutical reps understood classic psychological persuasion strategies, and how they employed them in a sophisticated and intuitive manner. One clever tactic they used was to hire physicians to give a brief lecture to other physicians about a drug. Now, they really didn’t care what the audience took from the lecture, but were actually interested in what the act of giving the lecture did to the speaker himself. They found that after giving a short lecture about the benefits of a drug, the speaker would begin to believe his own words and soon prescribe accordingly. Psychological studies show that people quickly start believing whatever comes out of their own mouths, even when they are paid to say it. This is a clear case of cognitive dissonance at play; doctors reason that if they are touting this drug, they must believe in it themselves — and so their beliefs alter to align with their speech.
The reps employed other tricks like switching on and off various accents, personalities, political affiliations, and basically served as persuasion machines (they may have mentioned the word “chameleon”). They were great at putting doctors at ease, relating to them as similar working people who go deep-sea fishing or play baseball together. They used these shared experiences to develop an understanding that the physicians write prescriptions for their “friends.” The physicians, of course, did not think that they were compromising their values when they were out shooting the breeze with the drug reps.
I was recently at a conference for the American Medical Association, where I gave a lecture about conflicts of interest. Interestingly, the lecture just before me was by a representative from a device company that created brain implants. In his lecture he made the case for selling devices in the operating room because doctors could need help learning how to use the device. And in order to fight conflicts of interest, the company no longer takes physicians to Hawaii for their annual conferences — and instead they have their conference in Wisconsin.
So, what do we do? First, we must realize that doctors have conflicts of interest. With this understanding we need to place barriers that prevent this kind of schmoozing, and to keep reps from undue access to physicians or patients. They, of course, have the right to send doctors information, but their interactions should stop there.
I have one more idea: What if we only allow people to be drug reps if they are over 75 and unattractive? Not only would these individuals have more personal experience with the healthcare system, it also could reduce conflicts of interest and open up job opportunities to an undervalued population.
Religion and Research
Direct my steps by Your word, and let no iniquity have dominion over me.
Redeem me from the oppression of man, that I may keep Your precepts.
Make Your face shine upon Your servant, and teach me Your statutes.
Rivers of water run down from my eyes, because men do no keep Your law.
-Psalm 119: 133-136
If you read Predictably Irrational, you may recall that we carried out a study on cheating that assessed the value of moral reminders. In the experiment, we asked participants to complete a test, told them they’d receive cash for every correct answer, and made sure they knew they had ample room to cheat. Now here’s the kicker: prior to starting, we had half the participants list ten books off their high-school reading list, and the other half to recall the Ten Commandments, a manipulation that turned out to have a marked effect on the results: While many in the first group deceitfully reported a higher number of correct answers, no one in the second group cheated.
How do we explain the findings? A tempting conclusion to draw would be to equate religiosity with a higher morality; however, this argument doesn’t hold, since in a follow-up study with atheist participants, recalling the Ten Commandments had the exact same effect. Rather, what was at play here was the power of a moral reminder: Prime a person to think about ethics right before they have an opportunity to cheat, and they’ll avoid immoral behavior.
This experiment also suggests to me that religion can be a good source of ideas for social science research. If you think about religion as a social mechanism that has evolved over time, then you can ask what purpose(s) its many rules serve and how they can help us to better understand human nature.
For example, though religious leaders may not have understood the exact psychology of moral reminders, they’ve certainly had enough of an intuitive sense of their importance to circulate the Ten Commandments and emphasize a whole score of other religious tenets, statutes, and regulations. Whether or not they could cite the causes for it, somewhere along the line they gathered that a good way to keep people in check was to present them with a moral benchmark to keep in mind (e.g. reciting prayers, for instance, before dinner as a continual reminder of the standards).
Given religion’s role in society and the way it evolves over time, I think we could benefit from using its wisdom to direct social science research. The key is to zero in on a religious tenet and ask why it’s there and what it suggests about human behavior, and to then empirically test the hypothesis with the hopes of deriving science from religious texts.
God bless.
Social power and morality
The following is taken from the graduation speech of Michael Lewis at Princeton in 2012. In it, he discusses an experiment that explores the relationship between power and morality.
“…… a pair of researchers in the Cal psychology department staged an experiment. They began by grabbing students, as lab rats. Then they broke the students into teams, segregated by sex. Three men, or three women, per team. Then they put these teams of three into a room, and arbitrarily assigned one of the three to act as leader. Then they gave them some complicated moral problem to solve: say what should be done about academic cheating, or how to regulate drinking on campus.
Exactly 30 minutes into the problem-solving the researchers interrupted each group. They entered the room bearing a plate of cookies. Four cookies. The team consisted of three people, but there were these four cookies. Every team member obviously got one cookie, but that left a fourth cookie, just sitting there. It should have been awkward. But it wasn’t. With incredible consistency the person arbitrarily appointed leader of the group grabbed the fourth cookie, and ate it. Not only ate it, but ate it with gusto: lips smacking, mouth open, drool at the corners of their mouths. In the end all that was left of the extra cookie were crumbs on the leader’s shirt.
This leader had performed no special task. He had no special virtue. He’d been chosen at random, 30 minutes earlier. His status was nothing but luck. But it still left him with the sense that the cookie should be his.”
—–
We’ve probably all heard the saying “absolute power corrupts absolutely.” Well, there is a great deal of research concerning the link between social power and morality, and most of it suggests that absolute power is not required to change people’s morals; sadly it tends to show that more power leads to less care for others, and less moral behavior.
A year in the life of a city bike.
At one point the people who run Hudson Urban Bikes, a bike rental company in the West Village, wondered what would happen to a bike if it was left chained to a post in the city for one year, and they took a picture of it each day to document its progress. The bicycle began its experimental journey equipped with all necessary equipment plus a basket, water bottle, splashguard and a few other goodies.
For quite a while the bike sits quietly chained next to a host of other bikes, retaining all of its accouterment. Then, on day 160 all of a sudden the water bottle goes missing. Then a few weeks later on day 212, both the lock and the basket walk off. From there things really begin to deteriorate, and it’s not long before the seat is missing, followed soon after by the front tire, splashguard, and handle bars.
Finally the forlorn frame itself disappears.
To my mind, this experiment cleverly mimics several aspects of dishonesty. People are basically fairly honest most of the time, but at some point they are tempted to cheat or take one small thing, or they see someone else do so. Over time this works through them, and maybe they take another small thing. After a while, this becomes habit, and people begin cheating at full throttle, and next thing you know, the whole bicycle is missing (figuratively).
That said, I think it bodes well that the bike lasted as long as it did, particularly after the lock was removed. It seems we can rest a little easier knowing that people, for the most part, don’t cheat as much as they could, or as much as we would expect them too, rationally speaking—after all, just think of how many people walked by the apparently free, unlocked bike and ignored it.
Women, Men, and Math Problems.
In the experiments my colleagues and I have run on cheating, we’ve used a task in which pride about personal performance and ability has no part. The matrix test is merely a search task (wherein participants find the two numbers out of 12 that add up to 10) rather than a skill. It’s not something you’re going to brag to your friends about in all likelihood.
Recent graduate Heidi Nicklaus of Rutgers University was interested in the opposite; she wondered how people’s pride about their perceived and imputed abilities would affect their dishonesty. Specifically, she was interested in gender stereotypes. We’ve all heard the stereotype that men tend to excel at math more than women, and that women can talk and write circles around men with their superior verbal skills. So the question was, if men are more proud of their mathematic ability and women of verbal, it might cause them to cheat more.
In her experiment, Heidi first primed her participants with two comical videos that exaggerate gender stereotypes (see below). Then participants were presented with one of two sets of fake data (presented as legitimate); one supported the math versus verbal aptitude stereotypes, the other countered them. Finally, participants took brief 10-question tests measuring both math and verbal aptitude, and were told they would receive $.50 for each correct answer. Similar to the experiments my colleagues and I have run on cheating, half of participants in each condition could cheat while the other half could not.
The results showed that when people could cheat, they generally did, which is what I’ve always found in cheating experiments. On average, people claimed one extra correct answer than when cheating was not possible (an average of 4 instead of 3 correct answers out of 10 on both math and verbal tests). No news here, so what about the effect of gender stereotypes? Did having them reinforced or, alternatively, countered before taking the test have any influence on cheating?
First, the hypothesis. What Heidi expected to find was that men and women would cheat along stereotypical lines, that is, that men would cheat more on math (to show that they did, indeed, excel in mathematics) and women would cheat more on the verbal portion for the same reason. So it was intriguing when Heidi found that men cheated more on math question than expected, but men and women cheated equally on verbal questions (rather than women cheating more as anticipated).
These findings—that people did not cheat more to keep up with perceived higher achievement by others—are similar to what my colleagues and I have found. In one experiment our results showed, similarly, that people cheated by the same amount regardless of whether they thought their peers solved an average of 4 or 8 out of 20 questions in a given amount of time (reporting an average of 6 correct answers). People cheated as much as they could justify, and apparently others’ performance is not of any great concern in this justification.
Oh, and as for the stereotype that kicked off the experiment: there were no differences in performance on math or verbal questions based on gender. So hopefully this harmful stereotype will fall by the wayside sooner rather than later, since nearly all similar studies yield the same conclusion.
Why We Lie (from WSJ)
Why We Lie (from the WSJ)
We like to believe that a few bad apples spoil the virtuous bunch. But research shows that everyone cheats a little—right up to the point where they lose their sense of integrity.
Not too long ago, one of my students, named Peter, told me a story that captures rather nicely our society’s misguided efforts to deal with dishonesty. One day, Peter locked himself out of his house. After a spell, the locksmith pulled up in his truck and picked the lock in about a minute.
“I was amazed at how quickly and easily this guy was able to open the door,” Peter said. The locksmith told him that locks are on doors only to keep honest people honest. One percent of people will always be honest and never steal. Another 1% will always be dishonest and always try to pick your lock and steal your television; locks won’t do much to protect you from the hardened thieves, who can get into your house if they really want to. The purpose of locks, the locksmith said, is to protect you from the 98% of mostly honest people who might be tempted to try your door if it had no lock.
We tend to think that people are either honest or dishonest. In the age of Bernie Madoff and Mark McGwire, James Frey and John Edwards, we like to believe that most people are virtuous, but a few bad apples spoil the bunch. If this were true, society might easily remedy its problems with cheating and dishonesty. Human-resources departments could screen for cheaters when hiring. Dishonest financial advisers or building contractors could be flagged quickly and shunned. Cheaters in sports and other arenas would be easy to spot before they rose to the tops of their professions.
But that is not how dishonesty works. Over the past decade or so, my colleagues and I have taken a close look at why people cheat, using a variety of experiments and looking at a panoply of unique data sets—from insurance claims to employment histories to the treatment records of doctors and dentists. What we have found, in a nutshell: Everybody has the capacity to be dishonest, and almost everybody cheats—just by a little. Except for a few outliers at the top and bottom, the behavior of almost everyone is driven by two opposing motivations. On the one hand, we want to benefit from cheating and get as much money and glory as possible; on the other hand, we want to view ourselves as honest, honorable people. Sadly, it is this kind of small-scale mass cheating, not the high-profile cases, that is most corrosive to society…..
For the rest of the article, please see the WSJ