Stanley Milgram Shock Experiment

Saul McLeod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Learn about our Editorial Process

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

On This Page:

Stanley Milgram, a psychologist at Yale University, carried out one of the most famous studies of obedience in psychology.

He conducted an experiment focusing on the conflict between obedience to authority and personal conscience.

Milgram (1963) examined justifications for acts of genocide offered by those accused at the World War II, Nuremberg War Criminal trials. Their defense often was based on obedience  – that they were just following orders from their superiors.

The experiments began in July 1961, a year after the trial of Adolf Eichmann in Jerusalem. Milgram devised the experiment to answer the question:

Could it be that Eichmann and his million accomplices in the Holocaust were just following orders? Could we call them all accomplices?” (Milgram, 1974).

Milgram (1963) wanted to investigate whether Germans were particularly obedient to authority figures, as this was a common explanation for the Nazi killings in World War II.

Milgram selected participants for his experiment by newspaper advertising for male participants to take part in a study of learning at Yale University.

The procedure was that the participant was paired with another person and they drew lots to find out who would be the ‘learner’ and who would be the ‘teacher.’  The draw was fixed so that the participant was always the teacher, and the learner was one of Milgram’s confederates (pretending to be a real participant).

stanley milgram generator scale

The learner (a confederate called Mr. Wallace) was taken into a room and had electrodes attached to his arms, and the teacher and researcher went into a room next door that contained an electric shock generator and a row of switches marked from 15 volts (Slight Shock) to 375 volts (Danger: Severe Shock) to 450 volts (XXX).

The shocks in Stanley Milgram’s obedience experiments were not real. The “learners” were actors who were part of the experiment and did not actually receive any shocks.

However, the “teachers” (the real participants of the study) believed the shocks were real, which was crucial for the experiment to measure obedience to authority figures even when it involved causing harm to others.

Milgram’s Experiment (1963)

Milgram (1963) was interested in researching how far people would go in obeying an instruction if it involved harming another person.

Stanley Milgram was interested in how easily ordinary people could be influenced into committing atrocities, for example, Germans in WWII.

Volunteers were recruited for a controlled experiment investigating “learning” (re: ethics: deception). 

Participants were 40 males, aged between 20 and 50, whose jobs ranged from unskilled to professional, from the New Haven area. They were paid $4.50 for just turning up.

Milgram

At the beginning of the experiment, they were introduced to another participant, a confederate of the experimenter (Milgram).

They drew straws to determine their roles – learner or teacher – although this was fixed, and the confederate was always the learner. There was also an “experimenter” dressed in a gray lab coat, played by an actor (not Milgram).

Two rooms in the Yale Interaction Laboratory were used – one for the learner (with an electric chair) and another for the teacher and experimenter with an electric shock generator.

Milgram Obedience: Mr Wallace

The “learner” (Mr. Wallace) was strapped to a chair with electrodes.

After he has learned a list of word pairs given to him to learn, the “teacher” tests him by naming a word and asking the learner to recall its partner/pair from a list of four possible choices.

The teacher is told to administer an electric shock every time the learner makes a mistake, increasing the level of shock each time. There were 30 switches on the shock generator marked from 15 volts (slight shock) to 450 (danger – severe shock).

Milgram Obedience IV Variations

The learner gave mainly wrong answers (on purpose), and for each of these, the teacher gave him an electric shock. When the teacher refused to administer a shock, the experimenter was to give a series of orders/prods to ensure they continued.

There were four prods, and if one was not obeyed, then the experimenter (Mr. Williams) read out the next prod, and so on.

Prod 1 : Please continue. Prod 2: The experiment requires you to continue. Prod 3 : It is absolutely essential that you continue. Prod 4 : You have no other choice but to continue.

These prods were to be used in order, and begun afresh for each new attempt at defiance (Milgram, 1974, p. 21). The experimenter also had two special prods available. These could be used as required by the situation:

  • Although the shocks may be painful, there is no permanent tissue damage, so please go on’ (ibid.)
  • ‘Whether the learner likes it or not, you must go on until he has learned all the word pairs correctly. So please go on’ (ibid., p. 22).

65% (two-thirds) of participants (i.e., teachers) continued to the highest level of 450 volts. All the participants continued to 300 volts.

Milgram did more than one experiment – he carried out 18 variations of his study.  All he did was alter the situation (IV) to see how this affected obedience (DV).

Conclusion 

The individual explanation for the behavior of the participants would be that it was something about them as people that caused them to obey, but a more realistic explanation is that the situation they were in influenced them and caused them to behave in the way that they did.

Some aspects of the situation that may have influenced their behavior include the formality of the location, the behavior of the experimenter, and the fact that it was an experiment for which they had volunteered and been paid.

Ordinary people are likely to follow orders given by an authority figure, even to the extent of killing an innocent human being.  Obedience to authority is ingrained in us all from the way we are brought up.

People tend to obey orders from other people if they recognize their authority as morally right and/or legally based. This response to legitimate authority is learned in a variety of situations, for example in the family, school, and workplace.

Milgram summed up in the article “The Perils of Obedience” (Milgram 1974), writing:

“The legal and philosophic aspects of obedience are of enormous import, but they say very little about how most people behave in concrete situations. I set up a simple experiment at Yale University to test how much pain an ordinary citizen would inflict on another person simply because he was ordered to by an experimental scientist. Stark authority was pitted against the subjects’ [participants’] strongest moral imperatives against hurting others, and, with the subjects’ [participants’] ears ringing with the screams of the victims, authority won more often than not. The extreme willingness of adults to go to almost any lengths on the command of an authority constitutes the chief finding of the study and the fact most urgently demanding explanation.”

Milgram’s Agency Theory

Milgram (1974) explained the behavior of his participants by suggesting that people have two states of behavior when they are in a social situation:

  • The autonomous state – people direct their own actions, and they take responsibility for the results of those actions.
  • The agentic state – people allow others to direct their actions and then pass off the responsibility for the consequences to the person giving the orders. In other words, they act as agents for another person’s will.

Milgram suggested that two things must be in place for a person to enter the agentic state:

  • The person giving the orders is perceived as being qualified to direct other people’s behavior. That is, they are seen as legitimate.
  • The person being ordered about is able to believe that the authority will accept responsibility for what happens.
According to Milgram, when in this agentic state, the participant in the obedience studies “defines himself in a social situation in a manner that renders him open to regulation by a person of higher status. In this condition the individual no longer views himself as responsible for his own actions but defines himself as an instrument for carrying out the wishes of others” (Milgram, 1974, p. 134).

Agency theory says that people will obey an authority when they believe that the authority will take responsibility for the consequences of their actions. This is supported by some aspects of Milgram’s evidence.

For example, when participants were reminded that they had responsibility for their own actions, almost none of them were prepared to obey.

In contrast, many participants who were refusing to go on did so if the experimenter said that he would take responsibility.

According to Milgram (1974, p. 188):

“The behavior revealed in the experiments reported here is normal human behavior but revealed under conditions that show with particular clarity the danger to human survival inherent in our make-up.

And what is it we have seen? Not aggression, for there is no anger, vindictiveness, or hatred in those who shocked the victim….

Something far more dangerous is revealed: the capacity for man to abandon his humanity, indeed, the inevitability that he does so, as he merges his unique personality into larger institutional structures.”

Milgram Experiment Variations

The Milgram experiment was carried out many times whereby Milgram (1965) varied the basic procedure (changed the IV).  By doing this Milgram could identify which factors affected obedience (the DV).

Obedience was measured by how many participants shocked to the maximum 450 volts (65% in the original study). Stanley Milgram conducted a total of 23 variations (also called conditions or experiments) of his original obedience study:

In total, 636 participants were tested in 18 variation studies conducted between 1961 and 1962 at Yale University.

In the original baseline study – the experimenter wore a gray lab coat to symbolize his authority (a kind of uniform).

The lab coat worn by the experimenter in the original study served as a crucial symbol of scientific authority that increased obedience. The lab coat conveyed expertise and legitimacy, making participants see the experimenter as more credible and trustworthy.

Milgram carried out a variation in which the experimenter was called away because of a phone call right at the start of the procedure.

The role of the experimenter was then taken over by an ‘ordinary member of the public’ ( a confederate) in everyday clothes rather than a lab coat. The obedience level dropped to 20%.

Change of Location:  The Mountain View Facility Study (1963, unpublished)

Milgram conducted this variation in a set of offices in a rundown building, claiming it was associated with “Research Associates of Bridgeport” rather than Yale.

The lab’s ordinary appearance was designed to test if Yale’s prestige encouraged obedience. Participants were led to believe that a private research firm experimented.

In this non-university setting, obedience rates dropped to 47.5% compared to 65% in the original Yale experiments. This suggests that the status of location affects obedience.

Private research firms are viewed as less prestigious than certain universities, which affects behavior. It is easier under these conditions to abandon the belief in the experimenter’s essential decency.

The impressive university setting reinforced the experimenter’s authority and conveyed an implicit approval of the research.

Milgram filmed this variation for his documentary Obedience , but did not publish the results in his academic papers. The study only came to wider light when archival materials, including his notes, films, and data, were studied by later researchers like Perry (2013) in the decades after Milgram’s death.

Two Teacher Condition

When participants could instruct an assistant (confederate) to press the switches, 92.5% shocked to the maximum of 450 volts.

Allowing the participant to instruct an assistant to press the shock switches diffused personal responsibility and likely reduced perceptions of causing direct harm.

By attributing the actions to the assistant rather than themselves, participants could more easily justify shocking to the maximum 450 volts, reflected in the 92.5% obedience rate.

When there is less personal responsibility, obedience increases. This relates to Milgram’s Agency Theory.

Touch Proximity Condition

The teacher had to force the learner’s hand down onto a shock plate when the learner refused to participate after 150 volts. Obedience fell to 30%.

Forcing the learner’s hand onto the shock plate after 150 volts physically connected the teacher to the consequences of their actions. This direct tactile feedback increased the teacher’s personal responsibility.

No longer shielded from the learner’s reactions, the proximity enabled participants to more clearly perceive the harm they were causing, reducing obedience to 30%. Physical distance and indirect actions in the original setup made it easier to rationalize obeying the experimenter.

The participant is no longer buffered/protected from seeing the consequences of their actions.

Social Support Condition

When the two confederates set an example of defiance by refusing to continue the shocks, especially early on at 150 volts, it permitted the real participant also to resist authority.

Two other participants (confederates) were also teachers but refused to obey. Confederate 1 stopped at 150 volts, and Confederate 2 stopped at 210 volts.

Their disobedience provided social proof that it was acceptable to disobey. This modeling of defiance lowered obedience to only 10% compared to 65% without such social support. It demonstrated that social modeling can validate challenging authority.

The presence of others who are seen to disobey the authority figure reduces the level of obedience to 10%.

Absent Experimenter Condition 

It is easier to resist the orders from an authority figure if they are not close by. When the experimenter instructed and prompted the teacher by telephone from another room, obedience fell to 20.5%.

Many participants cheated and missed out on shocks or gave less voltage than ordered by the experimenter. The proximity of authority figures affects obedience.

The physical absence of the authority figure enabled participants to act more freely on their own moral inclinations rather than the experimenter’s commands. This highlighted the role of an authority’s direct presence in influencing behavior.

A key reason the obedience studies fascinate people is Milgram presented them as a scientific experiment, contrasting himself as an “empirically grounded scientist” compared to philosophers. He claimed he systematically varied factors to alter obedience rates.

However, recent scholarship using archival records shows Milgram’s account of standardizing the procedure was misleading. For example, he published a list of standardized prods the experimenter used when participants questioned continuing. Milgram said these were delivered uniformly in a firm but polite tone.

Analyzing audiotapes, Gibson (2013) found considerable variation from the published protocol – the prods differed across trials. The point is not that Milgram did poor science, but that the archival materials reveal the limitations of the textbook account of his “standardized” procedure.

The qualitative data like participant feedback, Milgram’s notes, and researchers’ actions provide a fuller, messier picture than the obedience studies’ “official” story. For psychology students, this shows how scientific reporting can polish findings in a way that strays from the less tidy reality.

Critical Evaluation

Inaccurate description of the prod methodology:.

A key reason the obedience studies fascinate people is Milgram (1974) presented them as a scientific experiment, contrasting himself as an “empirically grounded scientist” compared to philosophers. He claimed he systematically varied factors to alter obedience rates.

However, recent scholarship using archival records shows Milgram’s account of standardizing the procedure was misleading. For example, he published a list of standardized prods the experimenter used when participants questioned continuing. Milgram said these were delivered uniformly in a firm but polite tone (Gibson, 2013; Perry, 2013; Russell, 2010).

Perry’s (2013) archival research revealed another discrepancy between Milgram’s published account and the actual events. Milgram claimed standardized prods were used when participants resisted, but Perry’s audiotape analysis showed the experimenter often improvised more coercive prods beyond the supposed script.

This off-script prodding varied between experiments and participants, and was especially prevalent with female participants where no gender obedience difference was found – suggesting the improvisation influenced results. Gibson (2013) and Russell (2009) corroborated the experimenter’s departures from the supposed fixed prods. 

Prods were often combined or modified rather than used verbatim as published.

Russell speculated the improvisation aimed to achieve outcomes the experimenter believed Milgram wanted. Milgram seemed to tacitly approve of the deviations by not correcting them when observing.

This raises significant issues around experimenter bias influencing results, lack of standardization compromising validity, and ethical problems with Milgram misrepresenting procedures.

Milgram’s experiment lacked external validity:

The Milgram studies were conducted in laboratory-type conditions, and we must ask if this tells us much about real-life situations.

We obey in a variety of real-life situations that are far more subtle than instructions to give people electric shocks, and it would be interesting to see what factors operate in everyday obedience. The sort of situation Milgram investigated would be more suited to a military context.

Orne and Holland (1968) accused Milgram’s study of lacking ‘experimental realism,”’ i.e.,” participants might not have believed the experimental set-up they found themselves in and knew the learner wasn’t receiving electric shocks.

“It’s more truthful to say that only half of the people who undertook the experiment fully believed it was real, and of those two-thirds disobeyed the experimenter,” observes Perry (p. 139).

Milgram’s sample was biased:

  • The participants in Milgram’s study were all male. Do the findings transfer to females?
  • Milgram’s study cannot be seen as representative of the American population as his sample was self-selected. This is because they became participants only by electing to respond to a newspaper advertisement (selecting themselves).
  • They may also have a typical “volunteer personality” – not all the newspaper readers responded so perhaps it takes this personality type to do so.

Yet a total of 636 participants were tested in 18 separate experiments across the New Haven area, which was seen as being reasonably representative of a typical American town.

Milgram’s findings have been replicated in a variety of cultures and most lead to the same conclusions as Milgram’s original study and in some cases see higher obedience rates.

However, Smith and Bond (1998) point out that with the exception of Jordan (Shanab & Yahya, 1978), the majority of these studies have been conducted in industrialized Western cultures, and we should be cautious before we conclude that a universal trait of social behavior has been identified.

Selective reporting of experimental findings:

Perry (2013) found Milgram omitted findings from some obedience experiments he conducted, reporting only results supporting his conclusions. A key omission was the Relationship condition (conducted in 1962 but unpublished), where participant pairs were relatives or close acquaintances.

When the learner protested being shocked, most teachers disobeyed, contradicting Milgram’s emphasis on obedience to authority.

Perry argued Milgram likely did not publish this 85% disobedience rate because it undermined his narrative and would be difficult to defend ethically since the teacher and learner knew each other closely.

Milgram’s selective reporting biased interpretations of his findings. His failure to publish all his experiments raises issues around researchers’ ethical obligation to completely and responsibly report their results, not just those fitting their expectations.

Unreported analysis of participants’ skepticism and its impact on their behavior:

Perry (2013) found archival evidence that many participants expressed doubt about the experiment’s setup, impacting their behavior. This supports Orne and Holland’s (1968) criticism that Milgram overlooked participants’ perceptions.

Incongruities like apparent danger, but an unconcerned experimenter likely cued participants that no real harm would occur. Trust in Yale’s ethics reinforced this. Yet Milgram did not publish his assistant’s analysis showing participant skepticism correlated with disobedience rates and varied by condition.

Obedient participants were more skeptical that the learner was harmed. This selective reporting biased interpretations. Additional unreported findings further challenge Milgram’s conclusions.

This highlights issues around thoroughly and responsibly reporting all results, not just those fitting expectations. It shows how archival evidence makes Milgram’s study a contentious classic with questionable methods and conclusions.

Ethical Issues

What are the potential ethical concerns associated with Milgram’s research on obedience?

While not a “contribution to psychology” in the traditional sense, Milgram’s obedience experiments sparked significant debate about the ethics of psychological research.

Baumrind (1964) criticized the ethics of Milgram’s research as participants were prevented from giving their informed consent to take part in the study. 

Participants assumed the experiment was benign and expected to be treated with dignity.

As a result of studies like Milgram’s, the APA and BPS now require researchers to give participants more information before they agree to take part in a study.

The participants actually believed they were shocking a real person and were unaware the learner was a confederate of Milgram’s.

However, Milgram argued that “illusion is used when necessary in order to set the stage for the revelation of certain difficult-to-get-at-truths.”

Milgram also interviewed participants afterward to find out the effect of the deception. Apparently, 83.7% said that they were “glad to be in the experiment,” and 1.3% said that they wished they had not been involved.

Protection of participants 

Participants were exposed to extremely stressful situations that may have the potential to cause psychological harm. Many of the participants were visibly distressed (Baumrind, 1964).

Signs of tension included trembling, sweating, stuttering, laughing nervously, biting lips and digging fingernails into palms of hands. Three participants had uncontrollable seizures, and many pleaded to be allowed to stop the experiment.

Milgram described a businessman reduced to a “twitching stuttering wreck” (1963, p. 377),

In his defense, Milgram argued that these effects were only short-term. Once the participants were debriefed (and could see the confederate was OK), their stress levels decreased.

“At no point,” Milgram (1964) stated, “were subjects exposed to danger and at no point did they run the risk of injurious effects resulting from participation” (p. 849).

To defend himself against criticisms about the ethics of his obedience research, Milgram cited follow-up survey data showing that 84% of participants said they were glad they had taken part in the study.

Milgram used this to claim that the study caused no serious or lasting harm, since most participants retrospectively did not regret their involvement.

Yet archival accounts show many participants endured lasting distress, even trauma, refuting Milgram’s insistence the study caused only fleeting “excitement.” By not debriefing all, Milgram misled participants about the true risks involved (Perry, 2013).

However, Milgram did debrief the participants fully after the experiment and also followed up after a period of time to ensure that they came to no harm.

Milgram debriefed all his participants straight after the experiment and disclosed the true nature of the experiment.

Participants were assured that their behavior was common, and Milgram also followed the sample up a year later and found no signs of any long-term psychological harm.

The majority of the participants (83.7%) said that they were pleased that they had participated, and 74% had learned something of personal importance.

Perry’s (2013) archival research found Milgram misrepresented debriefing – around 600 participants were not properly debriefed soon after the study, contrary to his claims. Many only learned no real shocks occurred when reading a mailed study report months later, which some may have not received.

Milgram likely misreported debriefing details to protect his credibility and enable future obedience research. This raises issues around properly informing and debriefing participants that connect to APA ethics codes developed partly in response to Milgram’s study.

Right to Withdrawal 

The BPS states that researchers should make it plain to participants that they are free to withdraw at any time (regardless of payment).

When expressing doubts, the experimenter assured them all was well. Trusting Yale scientists, many took the experimenter at his word that “no permanent tissue damage” would occur, and continued administering shocks despite reservations.

Did Milgram give participants an opportunity to withdraw? The experimenter gave four verbal prods which mostly discouraged withdrawal from the experiment:

  • Please continue.
  • The experiment requires that you continue.
  • It is absolutely essential that you continue.
  • You have no other choice, you must go on.

Milgram argued that they were justified as the study was about obedience, so orders were necessary.

Milgram pointed out that although the right to withdraw was made partially difficult, it was possible as 35% of participants had chosen to withdraw.

Replications

Direct replications have not been possible due to current ethical standards . However, several researchers have conducted partial replications and variations that aim to reproduce some aspects of Milgram’s methods ethically.

One important replication was conducted by Jerry Burger in 2009. Burger’s partial replication included several safeguards to protect participant welfare, such as screening out high-risk individuals, repeatedly reminding participants they could withdraw, and stopping at the 150-volt shock level. This was the point where Milgram’s participants first heard the learner’s protests.

As 79% of Milgram’s participants who went past 150 volts continued to the maximum 450 volts, Burger (2009) argued that 150 volts provided a reasonable estimate for obedience levels. He found 70% of participants continued to 150 volts, compared to 82.5% in Milgram’s comparable condition.

Another replication by Thomas Blass (1999) examined whether obedience rates had declined over time due to greater public awareness of the experiments. Blass correlated obedience rates from replication studies between 1963 and 1985 and found no relationship between year and obedience level. He concluded that obedience rates have not systematically changed, providing evidence against the idea of “enlightenment effects”.

Some variations have explored the role of gender. Milgram found equal rates of obedience for male and female participants. Reviews have found most replications also show no gender difference, with a couple of exceptions (Blass, 1999). For example, Kilham and Mann (1974) found lower obedience in female participants.

Partial replications have also examined situational factors. Having another person model defiance reduced obedience compared to a solo participant in one study, but did not eliminate it (Burger, 2009). The authority figure’s perceived expertise seems to be an influential factor (Blass, 1999). Replications have supported Milgram’s observation that stepwise increases in demands promote obedience.

Personality factors have been studied as well. Traits like high empathy and desire for control correlate with some minor early hesitation, but do not greatly impact eventual obedience levels (Burger, 2009). Authoritarian tendencies may contribute to obedience (Elms, 2009).

In sum, the partial replications confirm Milgram’s degree of obedience. Though ethical constraints prevent full reproductions, the key elements of his procedure seem to consistently elicit high levels of compliance across studies, samples, and eras. The replications continue to highlight the power of situational pressures to yield obedience.

Milgram (1963) Audio Clips

Below you can also hear some of the audio clips taken from the video that was made of the experiment. Just click on the clips below.

Why was the Milgram experiment so controversial?

The Milgram experiment was controversial because it revealed people’s willingness to obey authority figures even when causing harm to others, raising ethical concerns about the psychological distress inflicted upon participants and the deception involved in the study.

Would Milgram’s experiment be allowed today?

Milgram’s experiment would likely not be allowed today in its original form, as it violates modern ethical guidelines for research involving human participants, particularly regarding informed consent, deception, and protection from psychological harm.

Did anyone refuse the Milgram experiment?

Yes, in the Milgram experiment, some participants refused to continue administering shocks, demonstrating individual variation in obedience to authority figures. In the original Milgram experiment, approximately 35% of participants refused to administer the highest shock level of 450 volts, while 65% obeyed and delivered the 450-volt shock.

How can Milgram’s study be applied to real life?

Milgram’s study can be applied to real life by demonstrating the potential for ordinary individuals to obey authority figures even when it involves causing harm, emphasizing the importance of questioning authority, ethical decision-making, and fostering critical thinking in societal contexts.

Were all participants in Milgram’s experiments male?

Yes, in the original Milgram experiment conducted in 1961, all participants were male, limiting the generalizability of the findings to women and diverse populations.

Why was the Milgram experiment unethical?

The Milgram experiment was considered unethical because participants were deceived about the true nature of the study and subjected to severe emotional distress. They believed they were causing harm to another person under the instruction of authority.

Additionally, participants were not given the right to withdraw freely and were subjected to intense pressure to continue. The psychological harm and lack of informed consent violates modern ethical guidelines for research.

Baumrind, D. (1964). Some thoughts on ethics of research: After reading Milgram’s” Behavioral study of obedience.”.  American Psychologist ,  19 (6), 421.

Blass, T. (1999). The Milgram paradigm after 35 years: Some things we now know about obedience to authority 1.  Journal of Applied Social Psychology ,  29 (5), 955-978.

Brannigan, A., Nicholson, I., & Cherry, F. (2015). Introduction to the special issue: Unplugging the Milgram machine.  Theory & Psychology ,  25 (5), 551-563.

Burger, J. M. (2009). Replicating Milgram: Would people still obey today? American Psychologist, 64 , 1–11.

Elms, A. C. (2009). Obedience lite. American Psychologist, 64 (1), 32–36.

Gibson, S. (2013). Milgram’s obedience experiments: A rhetorical analysis. British Journal of Social Psychology, 52, 290–309.

Gibson, S. (2017). Developing psychology’s archival sensibilities: Revisiting Milgram’s obedience’ experiments.  Qualitative Psychology ,  4 (1), 73.

Griggs, R. A., Blyler, J., & Jackson, S. L. (2020). Using research ethics as a springboard for teaching Milgram’s obedience study as a contentious classic.  Scholarship of Teaching and Learning in Psychology ,  6 (4), 350.

Haslam, S. A., & Reicher, S. D. (2018). A truth that does not always speak its name: How Hollander and Turowetz’s findings confirm and extend the engaged followership analysis of harm-doing in the Milgram paradigm. British Journal of Social Psychology, 57, 292–300.

Haslam, S. A., Reicher, S. D., & Birney, M. E. (2016). Questioning authority: New perspectives on Milgram’s ‘obedience’ research and its implications for intergroup relations. Current Opinion in Psychology, 11 , 6–9.

Haslam, S. A., Reicher, S. D., Birney, M. E., Millard, K., & McDonald, R. (2015). ‘Happy to have been of service’: The Yale archive as a window into the engaged followership of participants in Milgram’s ‘obedience’ experiment. British Journal of Social Psychology, 54 , 55–83.

Kaplan, D. E. (1996). The Stanley Milgram papers: A case study on appraisal of and access to confidential data files. American Archivist, 59 , 288–297.

Kaposi, D. (2022). The second wave of critical engagement with Stanley Milgram’s ‘obedience to authority’experiments: What did we learn?.  Social and Personality Psychology Compass ,  16 (6), e12667.

Kilham, W., & Mann, L. (1974). Level of destructive obedience as a function of transmitter and executant roles in the Milgram obedience paradigm. Journal of Personality and Social Psychology, 29 (5), 696–702.

Milgram, S. (1963). Behavioral study of obedience . Journal of Abnormal and Social Psychology , 67, 371-378.

Milgram, S. (1964). Issues in the study of obedience: A reply to Baumrind. American Psychologist, 19 , 848–852.

Milgram, S. (1965). Some conditions of obedience and disobedience to authority . Human Relations, 18(1) , 57-76.

Milgram, S. (1974). Obedience to authority: An experimental view . Harpercollins.

Miller, A. G. (2009). Reflections on” Replicating Milgram”(Burger, 2009), American Psychologis t, 64 (1):20-27

Nicholson, I. (2011). “Torture at Yale”: Experimental subjects, laboratory torment and the “rehabilitation” of Milgram’s “obedience to authority”. Theory & Psychology, 21 , 737–761.

Nicholson, I. (2015). The normalization of torment: Producing and managing anguish in Milgram’s “obedience” laboratory. Theory & Psychology, 25 , 639–656.

Orne, M. T., & Holland, C. H. (1968). On the ecological validity of laboratory deceptions. International Journal of Psychiatry, 6 (4), 282-293.

Orne, M. T., & Holland, C. C. (1968). Some conditions of obedience and disobedience to authority. On the ecological validity of laboratory deceptions. International Journal of Psychiatry, 6 , 282–293.

Perry, G. (2013). Behind the shock machine: The untold story of the notorious Milgram psychology experiments . New York, NY: The New Press.

Reicher, S., Haslam, A., & Miller, A. (Eds.). (2014). Milgram at 50: Exploring the enduring relevance of psychology’s most famous studies [Special issue]. Journal of Social Issues, 70 (3), 393–602

Russell, N. (2014). Stanley Milgram’s obedience to authority “relationship condition”: Some methodological and theoretical implications. Social Sciences, 3, 194–214

Shanab, M. E., & Yahya, K. A. (1978). A cross-cultural study of obedience. Bulletin of the Psychonomic Society .

Smith, P. B., & Bond, M. H. (1998). Social psychology across cultures (2nd Edition) . Prentice Hall.

Further Reading

  • The power of the situation: The impact of Milgram’s obedience studies on personality and social psychology
  • Seeing is believing: The role of the film Obedience in shaping perceptions of Milgram’s Obedience to Authority Experiments
  • Replicating Milgram: Would people still obey today?

Learning Check

Which is true regarding the Milgram obedience study?
  • The aim was to see how obedient people would be in a situation where following orders would mean causing harm to another person.
  • Participants were under the impression they were part of a learning and memory experiment.
  • The “learners” in the study were actual participants who volunteered to be shocked as part of the experiment.
  • The “learner” was an actor who was in on the experiment and never actually received any real shocks.
  • Although the participant could not see the “learner”, he was able to hear him clearly through the wall
  • The study was directly influenced by Milgram’s observations of obedience patterns in post-war Europe.
  • The experiment was designed to understand the psychological mechanisms behind war crimes committed during World War II.
  • The Milgram study was universally accepted in the psychological community, and no ethical concerns were raised about its methodology.
  • When Milgram’s experiment was repeated in a rundown office building in Bridgeport, the percentage of the participants who fully complied with the commands of the experimenter remained unchanged.
  • The experimenter (authority figure) delivered verbal prods to encourage the teacher to continue, such as ‘Please continue’ or ‘Please go on’.
  • Over 80% of participants went on to deliver the maximum level of shock.
  • Milgram sent participants questionnaires after the study to assess the effects and found that most felt no remorse or guilt, so it was ethical.
  • The aftermath of the study led to stricter ethical guidelines in psychological research.
  • The study emphasized the role of situational factors over personality traits in determining obedience.

Answers : Items 3, 8, 9, and 11 are the false statements.

Short Answer Questions
  • Briefly explain the results of the original Milgram experiments. What did these results prove?
  • List one scenario on how an authority figure can abuse obedience principles.
  • List one scenario on how an individual could use these principles to defend their fellow peers.
  • In a hospital, you are very likely to obey a nurse. However, if you meet her outside the hospital, for example in a shop, you are much less likely to obey. Using your knowledge of how people resist pressure to obey, explain why you are less likely to obey the nurse outside the hospital.
  • Describe the shock instructions the participant (teacher) was told to follow when the victim (learner) gave an incorrect answer.
  • State the lowest voltage shock that was labeled on the shock generator.
  • What would likely happen if Milgram’s experiment included a condition in which the participant (teacher) had to give a high-level electric shock for the first wrong answer?
Group Activity

Gather in groups of three or four to discuss answers to the short answer questions above.

For question 2, review the different scenarios you each came up with. Then brainstorm on how these situations could be flipped.

For question 2, discuss how an authority figure could instead empower those below them in the examples your groupmates provide.

For question 3, discuss how a peer could do harm by using the obedience principles in the scenarios your groupmates provide.

Essay Topic
  • What’s the most important lesson of Milgram’s Obedience Experiments? Fully explain and defend your answer.
  • Milgram selectively edited his film of the obedience experiments to emphasize obedient behavior and minimize footage of disobedience. What are the ethical implications of a researcher selectively presenting findings in a way that fits their expected conclusions?

Print Friendly, PDF & Email

The Milgram Experiment: How Far Will You Go to Obey an Order?

Understand the infamous study and its conclusions about human nature

  • Archaeology
  • Ph.D., Psychology, University of California - Santa Barbara
  • B.A., Psychology and Peace & Conflict Studies, University of California - Berkeley

A brief Milgram experiment summary is as follows: In the 1960s, psychologist Stanley Milgram conducted a series of studies on the concepts of obedience and authority. His experiments involved instructing study participants to deliver increasingly high-voltage shocks to an actor in another room, who would scream and eventually go silent as the shocks became stronger. The shocks weren't real, but study participants were made to believe that they were.

Today, the Milgram experiment is widely criticized on both ethical and scientific grounds. However, Milgram's conclusions about humanity's willingness to obey authority figures remain influential and well-known.

Key Takeaways: The Milgram Experiment

  • The goal of the Milgram experiment was to test the extent of humans' willingness to obey orders from an authority figure.
  • Participants were told by an experimenter to administer increasingly powerful electric shocks to another individual. Unbeknownst to the participants, shocks were fake and the individual being shocked was an actor.
  • The majority of participants obeyed, even when the individual being shocked screamed in pain.
  • The experiment has been widely criticized on ethical and scientific grounds.

Detailed Milgram’s Experiment Summary

In the most well-known version of the Milgram experiment, the 40 male participants were told that the experiment focused on the relationship between punishment, learning, and memory. The experimenter then introduced each participant to a second individual, explaining that this second individual was participating in the study as well. Participants were told that they would be randomly assigned to roles of "teacher" and "learner." However, the "second individual" was an actor hired by the research team, and the study was set up so that the true participant would always be assigned to the "teacher" role.

During the Milgram experiment, the learner was located in a separate room from the teacher (the real participant), but the teacher could hear the learner through the wall. The experimenter told the teacher that the learner would memorize word pairs and instructed the teacher to ask the learner questions. If the learner responded incorrectly to a question, the teacher would be asked to administer an electric shock. The shocks started at a relatively mild level (15 volts) but increased in 15-volt increments up to 450 volts. (In actuality, the shocks were fake, but the participant was led to believe they were real.)

Participants were instructed to give a higher shock to the learner with each wrong answer. When the 150-volt shock was administered, the learner would cry out in pain and ask to leave the study. He would then continue crying out with each shock until the 330-volt level, at which point he would stop responding.

During this process, whenever participants expressed hesitation about continuing with the study, the experimenter would urge them to go on with increasingly firm instructions, culminating in the statement, "You have no other choice, you must go on." The study ended when participants refused to obey the experimenter’s demand, or when they gave the learner the highest level of shock on the machine (450 volts).

Milgram found that participants obeyed the experimenter at an unexpectedly high rate: 65% of the participants gave the learner the 450-volt shock.

Critiques of the Milgram Experiment

The Milgram experiment has been widely criticized on ethical grounds. Milgram’s participants were led to believe that they acted in a way that harmed someone else, an experience that could have had long-term consequences. Moreover, an investigation by writer Gina Perry uncovered that some participants appear to not have been fully debriefed after the study —they were told months later, or not at all, that the shocks were fake and the learner wasn’t harmed. Milgram’s studies could not be perfectly recreated today, because researchers today are required to pay much more attention to the safety and well-being of human research subjects.

Researchers have also questioned the scientific validity of Milgram’s results. In her examination of the study, Perry found that Milgram’s experimenter may have gone off script and told participants to obey many more times than the script specified. Additionally, some research suggests that participants may have figured out that the learner was not harmed: in interviews conducted after the Milgram experiment, some participants reported that they didn’t think the learner was in any real danger. This mindset is likely to have affected their behavior in the study.

Variations on the Milgram Experiment

Milgram and other researchers conducted numerous versions of the experiment over time. The participants' levels of compliance with the experimenter’s demands varied greatly from one study to the next. For example, when participants were in closer proximity to the learner (e.g. in the same room), they were less likely to give the learner the highest level of shock.

Another version of the Milgram experiment brought three "teachers" into the experiment room at once. One was a real participant, and the other two were actors hired by the research team. During the experiment, the two non-participant teachers would quit as the level of shocks began to increase. Milgram found that these conditions made the real participant far more likely to "disobey" the experimenter, too: only 10% of participants gave the 450-volt shock to the learner.

In yet another version of the Milgram experiment, two experimenters were present, and during the experiment, they would begin arguing with one another about whether it was right to continue the study. In this version, none of the participants gave the learner the 450-volt shock.

Replicating the Milgram Experiment

Researchers have sought to replicate Milgram's original study with additional safeguards in place to protect participants. In 2009, Jerry Burger replicated Milgram’s famous experiment at Santa Clara University with new safeguards in place: the highest shock level was 150 volts, and participants were told that the shocks were fake immediately after the experiment ended. Additionally, participants were screened by a clinical psychologist before the experiment began, and those found to be at risk of a negative reaction to the study were deemed ineligible to participate.

Burger found that participants obeyed at similar levels as Milgram’s participants: 82.5% of Milgram’s participants gave the learner the 150-volt shock, and 70% of Burger’s participants did the same.

The Legacy of the Milgram Experiment

Milgram’s interpretation of his research was that everyday people are capable of carrying out unthinkable actions in certain circumstances. His research has been used to explain atrocities such as the Holocaust and the Rwandan genocide, though these applications are by no means widely accepted or agreed upon.

Importantly, not all participants obeyed the experimenter’s demands , and Milgram’s studies shed light on the factors that enable people to stand up to authority. In fact, as sociologist Matthew Hollander writes, we may be able to learn from the participants who disobeyed, as their strategies may enable us to respond more effectively to an unethical situation. The Milgram experiment suggested that human beings are susceptible to obeying authority, but it also demonstrated that obedience is not inevitable.

  • Baker, Peter C. “Electric Schlock: Did Stanley Milgram's Famous Obedience Experiments Prove Anything?” Pacific Standard (2013, Sep. 10). https://psmag.com/social-justice/electric-schlock-65377
  • Burger, Jerry M. "Replicating Milgram: Would People Still Obey Today?."  American Psychologist 64.1 (2009): 1-11. http://psycnet.apa.org/buy/2008-19206-001
  • Gilovich, Thomas, Dacher Keltner, and Richard E. Nisbett. Social Psychology . 1st edition, W.W. Norton & Company, 2006.
  • Hollander, Matthew. “How to Be a Hero: Insight From the Milgram Experiment.” HuffPost Contributor Network (2015, Apr. 29). https://www.huffingtonpost.com/entry/how-to-be-a-hero-insight-_b_6566882
  • Jarrett, Christian. “New Analysis Suggests Most Milgram Participants Realised the ‘Obedience Experiments’ Were Not Really Dangerous.” The British Psychological Society: Research Digest (2017, Dec. 12). https://digest.bps.org.uk/2017/12/12/interviews-with-milgram-participants-provide-little-support-for-the-contemporary-theory-of-engaged-followership/
  • Perry, Gina. “The Shocking Truth of the Notorious Milgram Obedience Experiments.” Discover Magazine Blogs (2013, Oct. 2). http://blogs.discovermagazine.com/crux/2013/10/02/the-shocking-truth-of-the-notorious-milgram-obedience-experiments/
  • Romm, Cari. “Rethinking One of Psychology's Most Infamous Experiments.” The Atlantic (2015, Jan. 28) . https://www.theatlantic.com/health/archive/2015/01/rethinking-one-of-psychologys-most-infamous-experiments/384913/
  • Gilligan's Ethics of Care
  • What Was the Robbers Cave Experiment in Psychology?
  • What Is Behaviorism in Psychology?
  • What Is the Zeigarnik Effect? Definition and Examples
  • What Is a Conditioned Response?
  • Psychodynamic Theory: Approaches and Proponents
  • Social Cognitive Theory: How We Learn From the Behavior of Others
  • Kohlberg's Stages of Moral Development
  • What's the Difference Between Eudaimonic and Hedonic Happiness?
  • Genie Wiley, the Feral Child
  • What Is the Law of Effect in Psychology?
  • What Is the Recency Effect in Psychology?
  • Heuristics: The Psychology of Mental Shortcuts
  • What Is Survivor's Guilt? Definition and Examples
  • 5 Psychology Studies That Will Make You Feel Good About Humanity
  • What Is Cognitive Bias? Definition and Examples

cropped Screenshot 2023 08 20 at 23.18.57

Behavioral Study of Obedience: Insights from Milgram’s Groundbreaking Experiments

A haunting question emerges from the annals of psychological research: how far will ordinary people go in obeying orders, even when those orders conflict with their moral conscience? This unsettling inquiry has captivated psychologists, sociologists, and ethicists for decades, sparking intense debates about the nature of human behavior and the power of authority.

The study of obedience in psychology delves deep into the human psyche, exploring the complex interplay between individual morality and societal pressures. At its core, obedience refers to the act of complying with commands or instructions from an authority figure, even when those directives may conflict with personal beliefs or ethical standards. It’s a phenomenon that permeates every aspect of our lives, from childhood interactions with parents and teachers to adult relationships with bosses, government officials, and other authority figures.

Understanding obedience is crucial for unraveling the mysteries of human behavior. It helps us comprehend why seemingly good people can sometimes participate in harmful or unethical acts when instructed to do so by those in positions of power. This knowledge has far-reaching implications for fields as diverse as organizational psychology, military training, and even the study of historical atrocities.

The history of obedience research is rich and complex, with roots stretching back to the aftermath of World War II. As the world grappled with the horrors of the Holocaust, psychologists and social scientists sought to understand how ordinary citizens could have participated in such widespread acts of cruelty and genocide. This quest for answers led to a series of groundbreaking experiments that would forever change our understanding of human behavior.

Milgram’s Obedience Experiment: A Landmark Study

In the early 1960s, a young psychologist named Stanley Milgram embarked on a research project that would become one of the most famous and controversial studies in the history of psychology. Motivated by a desire to understand the mechanisms behind the atrocities committed during the Holocaust, Milgram designed an experiment to test the limits of obedience to authority.

The experimental design was deceptively simple, yet profoundly revealing. Participants were told they were taking part in a study on learning and memory. They were instructed to administer electric shocks to a “learner” (actually an actor) whenever the learner gave an incorrect answer to a series of questions. The voltage of the shocks increased with each wrong answer, eventually reaching levels that would be fatal if real.

What the participants didn’t know was that the true purpose of the experiment was to see how far they would go in following orders to harm another person. The results were shocking, to say the least. A staggering 65% of participants continued to administer shocks up to the maximum voltage, despite hearing the learner’s (fake) cries of pain and pleas to stop.

These findings sent shockwaves through the scientific community and beyond. They suggested that ordinary people could be compelled to commit acts of cruelty simply by following orders from an authority figure. The implications were profound and deeply unsettling.

However, Milgram’s study was not without its critics. Behavioral research design principles were called into question, with many arguing that the experiment was unethical and potentially harmful to participants. The psychological distress experienced by some subjects raised serious concerns about the balance between scientific inquiry and ethical considerations.

Despite these controversies, Milgram’s work remains a cornerstone of obedience research, sparking countless debates and inspiring numerous follow-up studies. Its impact on our understanding of human behavior cannot be overstated.

Factors Influencing Obedience

Milgram’s experiments and subsequent research have identified several key factors that influence obedience. One of the most significant is the presence of authority figures. We’re hardwired to respect and follow those in positions of power, whether they’re wearing a lab coat, a police uniform, or a business suit. This tendency can sometimes override our own moral judgments, leading us to comply with orders we might otherwise question.

Social pressure and conformity also play crucial roles in obedience. Conformity occurs when people change their behavior to fit in with the group, even if it means going against their personal beliefs. In Milgram’s experiments, the presence of other participants (actually confederates of the experimenter) who obeyed without question increased the likelihood that the real subject would also comply.

The proximity of the authority figure and the victim also influences obedience levels. Milgram found that when participants were physically closer to the learner or further from the experimenter, they were less likely to obey harmful commands. This suggests that distance can create a psychological buffer that makes it easier to follow orders that conflict with our moral standards.

Personal responsibility and its diffusion also play significant roles in obedience situations. When individuals feel that the responsibility for their actions is shared with others or rests primarily with the authority figure giving the orders, they’re more likely to comply with unethical demands. This diffusion of responsibility can lead to a dangerous abdication of moral agency.

Replications and Variations of Milgram’s Study

Since Milgram’s groundbreaking work, numerous researchers have attempted to replicate and expand upon his findings. Cross-cultural studies have shown that while there are some variations in obedience levels across different societies, the basic tendency to obey authority is a universal human trait.

Gender differences in obedience behavior have also been explored, with mixed results. Some studies suggest that women may be slightly more likely to disobey unethical orders than men, but these differences are generally small and context-dependent.

Modern adaptations of obedience experiments have sought to address ethical concerns while still probing the limits of human compliance. For example, some researchers have used role-playing scenarios or simulated environments to study obedience without putting participants in potentially distressing situations.

The advent of virtual reality technology has opened up new avenues for obedience research. Human behavior experiments in virtual environments allow researchers to create realistic scenarios that test obedience without the ethical concerns associated with traditional methods. These studies have largely confirmed Milgram’s findings, showing that even in simulated environments, people tend to obey authority figures to a surprising degree.

Implications of Obedience Studies

The insights gained from obedience research have far-reaching implications for various aspects of society. Understanding destructive obedience can help us develop strategies to prevent individuals from blindly following harmful orders. This knowledge is particularly relevant in contexts where authority figures might abuse their power, such as in totalitarian regimes or corrupt organizations.

In the realm of organizational psychology, obedience studies have informed management practices and leadership training. By recognizing the power of authority and the potential for its misuse, companies can create more ethical work environments and foster a culture of responsible decision-making.

The findings from obedience research also have significant implications for military and law enforcement training. These institutions must strike a delicate balance between instilling discipline and obedience while also encouraging ethical behavior and moral courage. Training programs now often include scenarios that challenge trainees to question unethical orders and make independent moral judgments.

Perhaps most poignantly, obedience studies help us understand historical atrocities like the Holocaust. While they don’t excuse the actions of those who participated in such events, they provide insight into the psychological mechanisms that can lead ordinary people to commit extraordinary acts of cruelty when ordered to do so by authority figures.

Critiques and Limitations of Obedience Research

Despite its profound impact, obedience research is not without its critics. One of the main concerns is ecological validity – the extent to which laboratory findings can be generalized to real-world situations. Critics argue that the artificial nature of experiments like Milgram’s may not accurately reflect how people would behave in genuine high-stakes situations.

Ethical constraints in modern obedience studies have limited researchers’ ability to fully replicate or extend Milgram’s work. While this is undoubtedly necessary to protect participants, it also means that our understanding of obedience in extreme situations remains somewhat limited.

Some researchers have proposed alternative explanations for obedience behavior observed in these experiments. For instance, some argue that participants may have been motivated more by a desire to please the experimenter or avoid confrontation than by blind obedience to authority.

The generalizability of findings to real-world situations is another area of debate. While obedience studies provide valuable insights, the complexity of real-life scenarios involving authority and moral decision-making may not be fully captured in experimental settings.

Behavioral theories limitations also come into play when considering obedience research. These theories often struggle to account for individual differences in personality, cultural background, and personal experiences that may influence obedience behavior.

The Ongoing Relevance of Obedience Research

Despite these limitations, the study of obedience remains highly relevant in contemporary society. In an era of increasing polarization and the rise of authoritarian tendencies in various parts of the world, understanding the mechanisms of obedience is more crucial than ever.

The basic principles that govern your behavior , including obedience to authority, continue to shape our social interactions and societal structures. By studying these principles, we can develop strategies to promote ethical behavior and resist harmful compliance.

Future directions for obedience studies in psychology are likely to focus on more nuanced aspects of compliance and resistance. Researchers may explore how individual differences in personality, moral reasoning, and critical thinking skills influence obedience behavior. The role of technology in shaping our responses to authority figures is another promising area of investigation.

Behavioral science projects focusing on obedience could yield valuable insights into how we can foster environments that encourage ethical decision-making and moral courage. For instance, studies might explore how education and training can enhance individuals’ ability to question authority and make independent moral judgments.

The Power of Critical Thinking and Moral Responsibility

As we reflect on the lessons learned from decades of obedience research, one thing becomes clear: the importance of critical thinking and moral responsibility in the face of authority cannot be overstated. While obedience can serve important social functions, blind compliance with unethical orders can lead to disastrous consequences.

Cultivating a society that values independent moral reasoning alongside respect for legitimate authority is a challenging but crucial task. It requires a delicate balance between maintaining social order and encouraging individuals to question and challenge unethical directives.

Education plays a vital role in this process. By teaching critical thinking skills and ethical reasoning from an early age, we can equip individuals with the tools they need to navigate complex moral situations. Behavioral experiments in educational settings can help students understand the dynamics of obedience and develop strategies for resisting harmful compliance.

Moreover, fostering a culture that values moral courage is essential. This means celebrating those who stand up against unethical orders and creating systems that protect whistleblowers and dissenters. By doing so, we can create a society that is more resilient to the dangers of destructive obedience.

Conclusion: The Ongoing Journey of Understanding Human Behavior

The study of obedience has come a long way since Milgram’s groundbreaking experiments, but many questions remain unanswered. As we continue to grapple with the complexities of human behavior, obedience research serves as a powerful reminder of our capacity for both compliance and resistance.

Human behavior and mind study is an ongoing journey, one that requires constant vigilance and self-reflection. By understanding the factors that influence our obedience to authority, we can work towards creating a world where ethical considerations trump blind compliance.

As we move forward, it’s crucial to remember that the power to resist harmful orders lies within each of us. By cultivating critical thinking skills, fostering moral courage, and remaining vigilant against the abuse of authority, we can harness the positive aspects of obedience while guarding against its dangers.

The haunting question that opened this exploration may never be fully answered, but our ongoing efforts to understand and address the complexities of obedience bring us closer to a society where moral conscience prevails over unquestioning compliance. In this pursuit, we not only advance our scientific understanding of human behavior but also contribute to the creation of a more just and ethical world.

References:

1. Milgram, S. (1963). Behavioral Study of Obedience. The Journal of Abnormal and Social Psychology, 67(4), 371-378.

2. Blass, T. (1999). The Milgram Paradigm After 35 Years: Some Things We Now Know About Obedience to Authority. Journal of Applied Social Psychology, 29(5), 955-978.

3. Burger, J. M. (2009). Replicating Milgram: Would People Still Obey Today? American Psychologist, 64(1), 1-11.

4. Haslam, S. A., & Reicher, S. D. (2012). Contesting the “Nature” of Conformity: What Milgram and Zimbardo’s Studies Really Show. PLoS Biology, 10(11), e1001426.

5. Zimbardo, P. G. (2007). The Lucifer Effect: Understanding How Good People Turn Evil. Random House.

6. Slater, M., Antley, A., Davison, A., Swapp, D., Guger, C., Barker, C., … & Sanchez-Vives, M. V. (2006). A Virtual Reprise of the Stanley Milgram Obedience Experiments. PloS one, 1(1), e39.

7. Passini, S., & Morselli, D. (2009). Authority Relationships Between Obedience and Disobedience. New Ideas in Psychology, 27(1), 96-106.

8. Twenge, J. M., & Campbell, W. K. (2009). The Narcissism Epidemic: Living in the Age of Entitlement. Free Press.

9. Cialdini, R. B. (2009). Influence: Science and Practice (5th ed.). Pearson Education.

10. Bandura, A. (1999). Moral Disengagement in the Perpetration of Inhumanities. Personality and Social Psychology Review, 3(3), 193-209.

Similar Posts

Behavior Trends: Unveiling the Patterns Shaping Society

Behavior Trends: Unveiling the Patterns Shaping Society

From TikTok trends to the rise of remote work, the invisible forces shaping our daily decisions and societal norms are more powerful than ever. We’re living in an era where the pace of change is dizzying, and our collective behaviors are evolving at breakneck speed. But what exactly are these forces, and how do they…

Front Stage Behavior: Unveiling the Sociology of Public Personas

Front Stage Behavior: Unveiling the Sociology of Public Personas

Lights, camera, action: the curtain rises on the grand theater of social interaction, where front stage behavior takes center stage, captivating audiences and shaping the very fabric of our everyday lives. We’re all actors in this grand performance, whether we realize it or not. From the moment we step out our front door to the…

Behavior Picture Analysis: Decoding Non-Verbal Communication in Images

Behavior Picture Analysis: Decoding Non-Verbal Communication in Images

A picture may be worth a thousand words, but the unspoken messages hidden within can reveal volumes about human behavior and interaction. In our increasingly visual world, the ability to decipher these silent cues has become an invaluable skill. From a casual snapshot to a carefully composed portrait, every image tells a story that goes…

Behavioral Effects: Understanding Their Impact on Individuals and Society

Behavioral Effects: Understanding Their Impact on Individuals and Society

From subtle influences to life-altering consequences, behavioral effects shape the very fabric of our individual experiences and collective society, making their study a crucial key to unlocking the mysteries of the human mind and the world we inhabit. The intricate dance of human behavior, with its myriad steps and rhythms, has long fascinated scientists, philosophers,…

Behavior Connections: Unraveling the Complex Web of Human Interactions

Behavior Connections: Unraveling the Complex Web of Human Interactions

In the seemingly chaotic dance of human interactions, an intricate web of behavior connections lies waiting to be untangled. Like a tapestry woven with countless threads, our actions and reactions intertwine in ways both subtle and profound. These behavioral links form the very fabric of our social existence, shaping our relationships, decisions, and even our…

Bot Behavior: Unveiling the Patterns and Impacts of Automated Online Actors

Bot Behavior: Unveiling the Patterns and Impacts of Automated Online Actors

In a digital world teeming with activity, an unseen army of automated agents shapes our online experiences, often without our knowledge or consent. These digital entities, known as bots, have become an integral part of our online ecosystem, influencing everything from search results to social media interactions. But what exactly are these bots, and how…

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Save my name, email, and website in this browser for the next time I comment.

American Psychological Association Logo

This page has been archived and is no longer being updated regularly.

Ethics Rounds

A classic study, revisited

By Dr. Stephen Behnke

May 2009, Vol 40, No. 5

Print version: page 76

Electric current

Jerry Burger set out to investigate whether a contemporary study of obedience would yield results comparable to what Stanley Milgram found four decades ago. The January American Psychologist provides a thought-provoking account of Burger's results.

I write this month to encourage "Ethics Rounds" readers to find their January American Psychologist and examine a fascinating series of articles. Jerry Burger, of the Santa Clara University department of psychology, set out to investigate whether he could replicate the findings of Stanley Milgram's studies on obedience. Burger's thoughtful and thought-provoking article, along with five commentaries, should be read by students of psychology at every level of professional development. I hope Burger's work will lead and to a renewal of investigations into this area of human behavior and a thoughtful review of the ethical parameters of such experimental studies. Three of the many themes that emerge from these articles involve attention to ethical issues, the relationship between ethics and methodology, and the role of institutional review boards (IRBs) in regulating social science research.

Burger titles the first section of his article "Ethical Concerns." That Burger chose to begin his article in this manner underscores the centrality of ethics in discussions regarding post-Milgram studies on obedience, and ethics emerges as a major theme in four of the five commentaries on Burger's article. Burger's steps to safeguard participants' welfare began with an initial screening. A research assistant asked questions including whether the prospective participant had ever been diagnosed with a psychiatric disorder, was currently receiving psychotherapy or taking medication for emotional difficulties, had ever had problems with alcohol or substance abuse, or had experienced serious trauma such as child abuse, domestic violence, or combat. A positive response to any of these questions excluded the individual from participating; approximately 30 percent of prospective subjects were precluded on this basis. Prospective subjects were next given a series of scales that included depression and anxiety inventories. A clinical psychologist reviewed the responses and conducted interviews to identify individuals who might have a negative reaction to participating. Of the individuals who made it to this second stage of the process, nearly 40 percent were excluded. A clinical psychologist was instructed to end a participant's involvement immediately if any sign of excessive stress emerged. Burger emphasized repeatedly that participants were free to withdraw at any time and still receive their $50 payment. Debriefing took place immediately following a participant's session, so that participants learned within moments of their session's end that the confederate had not actually been shocked.

Beyond these safeguards, Burger altered Milgram's methodology in a manner that commentators found especially intriguing. Burger terms this alteration "The 150-Volt Solution." Burger observed that nearly four out of five people who passed the 150-volt mark in Milgram's study continued to administer shocks in 15-volt increments to the end of the scale, 450 volts. Burger reasoned that the 150-volt mark, the point at which the "learner" protests loudly and demands to end the study, could therefore serve as something of a watershed, the point which, if crossed, likely signals that the study participant will go the entire range of the generator's value. By assuming (with caution) that individuals who proceed beyond 150 volts would continue obeying instructions to administer up to the maximum shock, Burger concluded there was no need to continue the study after this point. He thus avoids the conditions under which Milgram's subjects experienced their most intense levels of distress. With the 150-volt solution, Burger attempts to accomplish two goals: employ the science of psychology to examine the phenomenon of obedience to authority, and simultaneously remove what many consider the most ethically problematic aspect of Milgram's study, the distress induced in the participants.

A second theme to emerge in the articles involves the relationship between methodology and ethics. Burger begins his discussion section by noting "People learning about Milgram's obedience studies often ask whether similar results would be found today. Ethical concerns prevent researchers from providing a definitive answer to that question." Advancing science and protecting the individuals with whom psychologists work are two of psychology's core values. Because every interaction with another human being raises at least the possibility of harm, avoiding harm entirely would end all research. On the other hand, allowing research to move forward with no restrictions whatsoever would risk harms that psychologists agree could not be justified by the knowledge gained. For this reason, our profession must find a point on that continuum that strikes the correct balance. One of Burger's important contributions is highlighting in an explicit manner the complexities of meeting this challenge.

In their commentary, Ludy Benjamin and Jeffry Simpson, while accepting that "most of the recent rules and regulations imposed by IRBs are reasonable, legitimate, and necessary," suggest it may be time to adjust where psychology has struck this balance for "studies high in experimental realism:"

Every day, distressing, difficult and discomforting events happen to tens of thousands of people from all walks of life. To better understand how bad events can be translated into good or better outcomes, we occasionally need to study how individuals manage and cope with negative or adverse situations in well-controlled experiments. The pendulum may have swung too far away from the use of well-designed, thoughtful, carefully developed, and potentially illuminating high-impact studies in social and personality psychology. One implication of Burger's (2009) article is that it may help reopen this important discussion.

Noting in his commentary that "Research on (behavioral) obedience to authority has been virtually nonexistent in recent years, a peculiar cessation given the unparalleled and nonabating interesting in Milgram's research," Arthur Miller discusses the impact of Burger's methodology on his results. Miller calls the 150 volt level "both imaginative and convincing," yet addresses how Burger's alterations of Milgram's methodology—both the extensive screening and the 150-volt solution—limit comparisons of the two studies by precluding one of Milgram's most striking findings, the intensity of the stress participants experienced as they obeyed even to administer shocks of apparently dangerous voltage:

It seems likely that if Burger's (2009) participants had, in fact, been taken past the 150-volt level, the indication of stress and emotion reported by Milgram would have appeared. But one cannot be certain. Here, Burger's use of intense screening comes into play. Because of that screening, participants who would have shown more emotion or stress could have been precluded from being in the study. So, the "half-full, half-empty glass" metaphor seems relevant. Yes, we have a procedure that seems to fulfill some aspects of "replicating Milgram," but at a rather serious cost.

Methodology can be the primary vehicle through which the ethical tension between competing values is negotiated. "Better" methodologies from this perspective are those that accommodate both goals of advancing science and protecting subjects' welfare. Burger's innovation, while not without useful critiques, is to have found a methodology that suggests a better accommodation between competing values.

A third theme to emerge is that of IRB approval. Alan Elm's commentary puts the question most starkly: "So how did Jerry Burger (2009) get his research approved by the IRB at a respectable American university?" Much attention has been focused on the shortcomings of IRBs in providing reasonable and informed review of social science studies. That IRB approval receives so much attention in these articles is an indication of the level of this concern among researchers, a concern sufficiently great that APA has this year created a new committee under its Board of Scientific Affairs whose primary focus will be IRBs and research-related regulations. Arthur Miller offers an optimistic possibility in his commentary: "it is conceivable that Burger's study could now be cited—for example, for IRB review committees—as empirical evidence for the ethical propriety of his modified procedure."

I have addressed only a few of the many thought-provoking contributions of Burger's study and the five commentaries. I hope psychologists will discuss Burger's methods and results from many perspectives, especially that of ethics. And I will be very interested to see whether Miller's suggestion, that Burger's study may affect how IRBs conduct their work, ultimately proves accurate.

Stephen Behnke, PhD, JD, directs APA's Ethics Office.

"Ethics Rounds" columns have been converted into CE programs. Visit APA's Continuing Education Office.

Letters to the Editor

  • Send us a letter

Psychological research, obedience and ethics

milgram experiment ethical considerations

Introduction

One of the best known studies in the history of psychology is the research on obedience carried out by Stanley Milgram in the 1960s. In his research Milgram demonstrated the lengths to which people are willing to go just because someone in authority tells them to do something. The studies Milgram conducted also raised the issue of ethics in research, as some critics argued that he failed to take sufficient precautions to protect the integrity and wellbeing of his participants. At the same time, more than any other study in psychology, the findings of Milgram’s research demonstrate why ethics are important.

As well as reading about Milgram's work and ethics, you will engage in an online activity to learn about the code of ethics concerning the psychological research that is conducted with human participants. You will also gain an understanding of the guidelines that govern the use of non-human animals in psychological research in a second online activity, and why psychologists conduct such research. You will have the opportunity of viewing two short films which will introduce you to the research of Alex Thornton who studies meerkats, and of Tetsuro Matsuzawa who works with chimpanzees.

This OpenLearn course is an adapted extract from the Open University course DE100 Investigating psychology 1 .

Learning outcomes

After studying this course, you should be able to:

describe the research of Stanley Milgram on obedience

recognise the main ethics principles governing psychological research

understand the ethics issues concerning research involving non-human animals

appreciate the value of conducting research with animals.

1 Milgram’s obedience study

Milgram was one of the most innovative and productive social psychologists of his generation, who undertook a variety of studies that explored social psychological aspects of everyday life. However, he is largely remembered for one dramatic piece of work – the obedience studies. The best way to get inside this study is to imagine that you are one of the participants taking part in Milgram’s experiment. So read on with that in mind.

1.1 The set-up

It’s 1961 and you are arriving at the doors of the Psychology Department of the prestigious Yale University in the USA. The reason you are here is that you replied to an advert in the local paper asking for volunteers to take part in a study on memory. The advert (see Figure 1) offered a fee plus expenses and said that you would be paid on arrival at the laboratory.

Described image

This is an image of the advertisement Milgram placed in a local paper to recruit members of the public to participate in his study. The only information provided about the study is that it involves a scientific investigation of memory and learning. The advert highlighted that participants would be paid four dollars for one hour of their time, and also that they don't need any special training, education or experience. Wanted were five hundred New Haven men: factory workers, city employees, labourers, barbers, businessmen, clerks, profession people, telephone workers, construction workers, salespeople, white-collar workers and others, aged between twenty and fifty. The announcement also specifically states that high school and college students could not be used.

As you walk through the doors you are met by a serious-looking man in a laboratory coat who turns out to be the experimenter. He introduces you to a genial middle-aged man who is described as a fellow volunteer. The experimenter explains that the study will involve one of the volunteers taking on the role of a ‘teacher’ and the other taking on the role of a ‘learner’. As part of the experiment, the ‘teacher’ will engage the ‘learner’ in a simple memory task. The ‘learner’ and the ‘teacher’ will be in different rooms and will communicate through microphones (see Figure 2). The experimenter reveals that the study is designed to investigate the effect of punishment on learning. The ‘teacher’ will be asked to administer an electric shock to the ‘learner’ every time the latter makes an incorrect response on the memory task.

Described image

This shows a diagram of the experimental set-up. The experimenter is shown as seated in the same room, but behind the 'teacher'. The 'learner' is shown as seated in an adjacent room on his own. The wiring attached to the 'learner' passes through the wall and into the apparatus, the shock generator, placed on a table in front of the participant.

To select who will be the ‘teacher’ and who will be the ‘learner’, you draw slips of paper. You pick out the ‘teacher’ slip. You then watch as the ‘learner’ is strapped into a chair, and you hear the experimenter tell him that ‘although the shocks can be extremely painful, they cause no permanent tissue damage’. The experimenter now gives you a sample shock of 45 volts to show you what the ‘learner’ will experience during the study. The shock is unpleasant, but short of being painful.

The experimenter then takes you into the adjacent room and sits you down in front of an impressive-looking apparatus that will be used to administer the shocks (see Figure 3).

Described image

This figure shows the apparatus - the shock generator - used in the study which is described next in the text.

The shock generator consists of a row of switches that run in 15 volt increments from 15 volts through to 450 volts. Under the label for each switch are some descriptive words, such as ‘slight shock’ (15 volts), ‘moderate shock’ (75 volts), ‘strong shock’ (135 volts), ‘very strong shock’ (195 volts), ‘intense shock’ (225 volts), ‘extremely intense shock’ (315 volts), ‘danger: severe shock’ (375 volts) and finally ‘XXX’ (435 volts). Suddenly, this looks quite serious and you probably hope that you don’t have to go very far up the scale. This is especially so given that you received the 45-volt shock, and you know that this was unpleasant enough. The last switch on the shock generator administers an electric impulse ten times as strong!

In the first phase of the experiment, the experimenter asks you, the ‘teacher’, to read a series of word pairs to the ‘learner’ who is expected to memorise them (for instance, ‘green-grass’ , ‘blue-sky’, ‘nice-day’). In the second phase, the test phase, you are asked to read out the first word of the pairs (e.g. ‘green’), followed by four possible responses (‘grass, hat, ink, apple’). If the ‘learner’ identifies the paired word correctly, you are to move on to the next word pair on the list. If the answer is wrong you have to tell the ‘learner’ the correct answer, indicate the level of punishment you are going to give them (starting with 15 volts), and flick the appropriate switch on the shock generator. For every subsequent incorrect answer, you are told to move one switch up the scale of shocks.

The experiment starts. To begin with everything is fine and the ‘learner’ gets most of the answers right. You have only used the shock generator a couple of times, and at this stage the shocks are mild. Then the ‘learner’ starts to get the answers wrong and you are moving up the shock scale into the ‘strong shock’ range. Although you cannot see the ‘learner’ you can hear him and as the shocks increase he starts to shout out. You have heard him grunt at the low voltage but now he is starting to ask to be let out. At 120 volts you hear him shout out in an agitated tone, complaining that he is in pain, and at 150 volts he asks to be released.

Suddenly, you feel uncomfortable and you decide to stop. The experimenter, the man in the grey coat, objects and asks you to carry on, in spite of the ‘learner’s’ protestations.

What do you think you would do in this situation? At what point would you stop? 200 volts? 150 volts? Would you respond to the cries of your fellow volunteer or would you complete the job you agreed to do and carry out the instructions of the experimenter?

How many people do you think would continue to follow the orders? At what point do you think people would stop?

Before Milgram carried out the study, he posed the same questions as in Question 1 to different groups of people, including ordinary members of the public, college students, psychologists and psychiatrists. He asked them to speculate on how far they thought most people would go if asked to administer shocks. Most ordinary people said that participants would generally refuse to administer shock, or at least not go very far beyond the point where the ‘learner’ experienced pain. Also, most said that participants should rebel, and that they should not continue beyond around 150 volts. Among the professional groups, there was widespread agreement that nobody taking part in the study go all the way.

You will be relieved to know that in the actual study carried out by Milgram, no person was hurt during the procedure, and the only actual shock administered was the 45-volt ‘tester’ given to the ‘teacher’. In fact, the whole situation was staged. The role of ‘experimenter’ was played by a 31-year-old biology teacher. The ‘learner’, presented as a ‘fellow volunteer’, was in on the deception and was merely playing the part. In reality, he was a 47-year-old accountant, who was chosen for the role because he appeared mild-mannered and likeable. He was not the sort of person one would want to see hurt. The drawing of slips of paper was fixed to ensure that the ‘naive participant’ was always cast in the role of the ‘teacher’, and the ‘shock generator’ was simply a simulator. The sounds (the moans and cries) that the participants heard were a recording played from the adjacent room. Importantly, however, the deception was so good that participants believed that they were actually administering shocks. So the study presented an ingenious way of discovering how far people would be willing to go, just because a psychological experiment on ‘the effects of punishment on learning’ demanded it. Most people like to think that they (and people around them) would not go very far. But what happened when Milgram actually placed people in that position?

1.2 The results

In the first instance Milgram conducted the study on a sample of forty participants, all of them male. Each played the role of ‘teacher’ in the situation described in Section 1.1. Each participant went through the identical experimental procedure: all forty heard the same instructions, encountered the same ‘experimenter’ and ‘learner’, heard identical (pre-recorded) cries from the next room. The ‘experimenter’ in the grey lab coat offered the same words of encouragement. The sessions were filmed (Figure 4) and notes were taken by observers looking through an observation mirror.

Described image

This figure comprises three photos taken during the Milgram study. The first is of the experimenter instructing the participant on how to use the shock generator. The second is of the participant getting seated at the table in front of the shock generator. The third is of the ‘learner’ being strapped to the apparatus.

Milgram found that, of the forty participants who took part in the study, all obeyed up to 300 volts, the twentieth switch on the shock generator. This is the point at which the ‘learner’ was heard screaming: ‘I absolutely refuse to answer any more. Get me out of here. You can’t hold me here. Get me out. Get me out of here.’ However, only five of the forty participants refused to continue beyond this point. Four gave only one more shock before breaking off, with an additional five stopping between 315 volts and 435 volts. But as many as twenty-six continued to the end of the scale and administered the maximum 450 volts. This is despite the fact that, at 330 volts, they had already heard intense and prolonged screaming: ‘Let me out of here. Let me out of here. … Let me out of here. You have no right to hold me here. Let me out! Let me out!’ Shocks beyond 330 volts were accompanied by eerie silence. Nevertheless, twenty-six ordinary members of the public from Connecticut administered the maximum shock and continued to do so until the experimenter called a halt to the proceedings.

As well as counting the number of participants who went all the way on the shock generator, Milgram also observed their reactions. Participants who took part in the study generally displayed signs of nervousness and tension. Many were visibly uncomfortable and probably would not have continued had they not heard the experimenter say things like ‘Please continue’, ‘Please carry on’, ‘It is absolutely essential that you continue’ or ‘You have no choice; you must go on’. At the end of the study, many of the obedient participants heaved sighs of relief or shook their heads in apparent regret. Some even had laughing fits during the experiment, probably brought on by anxiety. Milgram (1963, p. 375) wrote that ‘full-blown, uncontrollable seizures were observed for 3 subjects. On one occasion we observed a seizure so violently convulsive that it was necessary to call a halt to the experiment’. (You may have noticed that in this quote Milgram refers to people who took part in his study as ‘subjects’. This was common practice in psychology in the 1960s. Today the word ‘participant’ is used instead as the word ‘subject’ is considered demeaning, and lacking in respect towards volunteers on whose participation much of psychological research ultimately depends.)

Do Milgram’s findings seem plausible to you? Ordinary members of the public were prepared to administer electric shocks to another person on the mere (albeit persistent) request of a man in a laboratory coat. They did so despite the protests from the ‘victim’ and continued even after the supposed recipient of the shocks went quiet. Before the study, when Milgram asked his fellow professionals to predict how many participants would refuse to go all the way, they said that all of them would do so. In reality only 35 per cent did. In Milgram’s study, the average voltage at which participants stopped shocking the ‘learner’ was 368 volts. Members of the public predicted that people would stop at around 140 volts. This is a remarkable discrepancy. It is therefore not surprising that Milgram’s research went on to provoke considerable debate.

Box 1 Why do it this way?

Milgram’s obedience work is remarkable, not only because of the important questions it sought to explore, but also because it is a fine example of good experimental procedure in social psychology.

The most important feature of any laboratory experiment is its controlled nature. Note that every person who took part in Milgram’s research underwent an identical experience. All participants received the same instructions, encountered the same individuals (the ‘experimenter’ and the fellow ‘volunteer’) and heard identical cries and protestations from the ‘learner’. To ensure consistency in the experimental procedure, Milgram even recorded the anguished cries in advance, and played them to participants from a tape.

This equivalence of experience across the forty participants was essential if meaningful comparisons were to be made. It ensured that any difference in behaviour observed in the study could not be attributed, for instance, to the fact that some participants heard louder or more desperate cries than others. For similar reasons, Milgram used the same ‘learner’ and ‘experimenter’ with each participant. He wanted to ensure that none of the results could be accounted for by differences in the personality or the demeanour of the confederates.

Another interesting aspect of Milgram’s research is that he recruited participants from the general public, using a newspaper advert. At the time (and still now in many psychology departments) participants tended to be recruited mainly from among the student population. However, Milgram was interested in exploring the level of obedience to scientific authority among people with no direct link to the university or research environment, so he recruited from the general public.

Finally, in Milgram’s original study, all forty participants were male. Why do you think this was the case? This was not because Milgram wanted to exclude women from his research. He later conducted further studies in which he explored gender differences in obedience. In the initial study, however, he decided to control for the potential effects of gender on the findings by limiting the sample to men.

1.3 The variations

The findings of Milgram’s original study highlighted the phenomenon of obedience, but it could not reveal what it is about the situation that made participants administer potentially lethal shocks to a fellow human being. To address this question Milgram carried out further research in which he introduced subtle variations to the original procedure. By examining the effects of these variations on levels of obedience, he was able to isolate specific aspects of the situation that might influence whether participants obey or not.

By the time Milgram completed his research in 1962 he had processed 800 people through nineteen variations of the original design. For instance, in one variation, Milgram introduced into the proceedings a dialogue about a heart attack. He wanted to see whether alerting the participants to the impact of the shocks on the ‘learner’s’ health might reduce obedience. Note that all other aspects of the original study were preserved. Interestingly, the conversation about the heart attack made no real difference. Twenty six out of the forty participants still continued to 450 volts, although those who stopped did so at a lower voltage with five stopping as soon as the ‘learner’ asked to be let out. So, the reference to the heart attack made those who disobeyed do so earlier, but it did not prevent the more obedient participants from going all the way.

Milgram also varied the proximity of the ‘learner’ and ‘teacher’. In one variation he put them in the same room, while in another he required the ‘teacher’ to hold the ‘learner’s’ arm down on a plate to receive the electric shock. This manipulation had a clear effect. Milgram found that the closer you place the ‘teacher’ to the ‘learner’, the fewer shocks the ‘teacher’ is likely to administer. Equally, the further you place the ‘learner’ away from the ‘teacher’, the less the impact their pleas are likely to have.

Equally crucial was the presence of the authority figure. In one variation, the ‘experimenter’ in the grey coat pretended to have to leave the experiment owing to some emergency and was replaced by a person in plain clothes, who was not a scientist. Only 20 per cent of participants went all the way and gave the ‘learner’ 450-volt shocks. Similar results were obtained when orders were given by phone. The physical presence of an authority figure was therefore crucial.

In another variation Milgram placed two ‘experimenters’ in the room. One told the participants to continue (as in the original study), while the other told them to stop. In this variation, all the participants stopped giving the shocks very early on. This showed that an absence of a clear authority figure reduces obedience.

Milgram also conducted a version of the experiment in which he placed a second ‘teacher’ in the room, although this one was a stooge instructed to obey until the end. In this variation all the participants went along with the confederate and shocked up to 450 volts! So the mere presence of another obedient ‘volunteer’ made all the participants go all the way.

One of the main conclusions of Milgram’s work was that under certain conditions involving the presence of authority, people suspend their capacity to make informed moral judgments and defer responsibility for their actions to those in authority. When people are in this particular frame of mind, the nature of the task that they are asked to perform becomes largely irrelevant, and the main determinant of their actions is the commands of the authority figure.

1.4 Summary

  • Milgram found that most people would administer potentially lethal levels of shock to another human being, just because they were told to do so by an authority figure.
  • The use of a controlled experimental procedure enabled Milgram to explore different aspects of the situation that influence the extent to which people will obey authority.
  • Two key factors in obedience are the presence of a clear authority figure, and the distance between the person administering the shock and the ‘victim’.

2 Milgram’s study and ethics

At the beginning of Section 1 you were asked to put yourself in the shoes of one of the participants in Milgram’s research. How do you think being a participant in the study felt? As you already read, many of the participants were visibly uncomfortable during the procedure. This is one of the reasons why the study created a storm, starting with a hostile review of the research in a newspaper, the St . Louis Post-Dispatch . The newspaper criticised Milgram and Yale University for putting the participants in such a stressful situation. It claimed that Milgram violated the rules of ethics which guide psychological research. The charge was repeated in academic circles, and led to Milgram’s application to join the American Psychological Association being put on hold for a year. Milgram made a robust rebuttal of the charges and the debate about the issues led to the introduction of new codes of good practice for psychologists.

Before we look at the arguments that swirled around the obedience study we need to consider what we mean by ethics . It all starts with morals , which are rules to guide our behaviour. These rules are based on a number of socially agreed principles which are used to develop clear and logical guidelines to direct behaviour. They also contain ideas about what is good and desirable in human behaviour. Ethics , in the context of psychological research, refers to a moral framework that governs what psychologists can and cannot do.

The first generally accepted code of ethics for research on humans was devised in 1947 as a response to the very events that provoked Milgram’s research. During the Second World War (1939–45), under the Nazi regime, research was carried out on human beings that led to many deaths, deformities and long-term injuries. Revelations about this research were as great a shock for the post-war world as the death camps, because these acts of brutality and murder were conducted by doctors and scientists.

After the war the victors held a series of trials, in the German city of Nuremberg, of people who had taken part in the worst excesses of the horrors that had swept across Europe. Among them were twenty-three doctors involved in the brutal experiments. Sixteen of them were found guilty, of whom seven were sentenced to death. Significantly, the judgement included a statement about how scientists should behave when experimenting on other humans. This is referred to as the Nuremberg Code (see Table 1) and it became the basis for future ethical codes in medicine and psychology.

Table 1 The Nuremberg Code (1946)
1 The voluntary consent of the human subject is absolutely essential
2 The experiment should yield fruitful results for the good of society, that cannot be obtained by other means
3 The experiment should be based on previous research so that the anticipated results can justify the research
4 All unnecessary physical and mental suffering should be avoided
5 No experiment should be conducted where there is reason to believe that death or disabling injury may be the result
6 The degree of risk should also be less than the potential humanitarian importance of the research
7 Adequate precautions should be in place to protect the subjects against any possible injury
8 Experiments should only be conducted by qualified persons
9 The human subject should always be at liberty to end the experiment
10 The scientist in charge should be prepared to terminate any experiment if there is probable cause to believe that continuation is likely to result in injury or death

Four key principles emerged from the Nuremberg Code. First, participants must be able to give informed consent to the procedure. Second, they must retain the right to withdraw from the study whenever they want. Third, the welfare of the participant must be protected wherever possible. The fourth principle is the most difficult to interpret because it concerns the costs and benefits of the study. It says that any risks to the participants must be greatly outweighed by the possible benefits for the greater good.

2.2 The case against Milgram

Before you go on to read about the criticism of Milgram’s obedience studies, try to think through all the issues relating to ethics that are raised by this work.

In what way were the participants deceived, or harmed? Did they have the right to withdraw? Do you think that in Milgram’s case the ends justify the means? Do the benefits of the study justify the costs? Do you think that the results of the study are worth the pain and discomfort caused to the participants?

Among those who were highly critical of Milgram’s study was fellow psychologist Diana Baumrind. She started her critique by noting the dilemma that all research psychologists face: ‘Certain problems in psychological research require the experimenter to balance his career and scientific interests against the interests of his prospective subjects’ (Baumrind, 1964, p. 421).

Baumrind challenged Milgram on whether he had properly protected the welfare of the participants . She used direct quotes from Milgram’s original report to illustrate the lack of regard she said was shown to the participants. In particular, she noted the detached manner in which Milgram described the emotional turmoil experienced by the volunteers. For example:

In a large number of cases the degree of tension [in the participants] reached extremes that are rarely seen in sociopsychological laboratory studies. Subjects were observed to sweat, tremble, stutter, bite their lips, groan, and dig their fingernails into their flesh. These were characteristic rather than exceptional responses to the experiment.

In Baumrind’s view, and in the view of numerous others, the levels of anxiety experienced by participants were enough to warrant halting the experiment. What is more, just because someone volunteers to take part in the study (i.e. gives informed consent at the start of the study), it does not mean that the researcher no longer has responsibilities towards them and their wellbeing. On the principle of cost – benefit , Baumrind challenged the view that the scientific worth of the study balanced out the distress caused to the participants. She acknowledged that some harm to participants might be a necessary part of some research – for example, when testing out new medical procedures – as in those cases results cannot be achieved in any other way. Social psychology, however, is not in the same game as medicine and is unlikely to produce life-saving results. The strength of the conclusions does not, therefore, justify harming participants. Milgram related his study to the behaviour of people who worked in the Nazi death camps and suggested that his study illuminated the way that ordinary people living ordinary lives are capable of playing a part in destructive and cruel acts. Baumrind dismissed this justification for the study and suggested there are few, if any, parallels between the behaviour in the study and the behaviour in the death camps.

Baumrind went on to make a further criticism by considering the effect of this work on the public image of psychology, and suggested that it would be damaged because the general public would judge that the participants were not protected or respected.

A further potential problem with Milgram’s experiment concerns the participants’ right to withdraw . Do you think that this principle, embedded in the Nuremberg Code, was sufficiently observed in Milgram’s research? Recall that one of the key aspects of the experimental procedure was that whenever a participant demonstrated a reluctance to carry on with administering the shocks, they were told by the ‘experimenter’ in the grey coat ‘you must go on’, or ‘you have no choice; you must go on’. It might be argued that telling a participant that they ‘have no choice’ but to continue with the experiment contravenes the right to withdraw, which is enshrined in the ethics code. To be fair, fourteen of the forty participants in the original study did withdraw, in spite of being told that they had no choice, so it could be argued that, ultimately, the participants did have a choice. It is just that making that choice was made more difficult by the presence of the ‘experimenter’ and by his prods. After all, the study was about obedience, and the instructions from the ‘experimenter’ were essential to the investigation. Exercising or not exercising the right to withdraw is what the study was about.

2.3 The case for the defence

Milgram made a series of robust defences for the study, starting with a response to the newspaper article that first raised concerns. He dismissed the accusation that participants were severely traumatised by the experience. He argued that ‘relatively few subjects experienced greater tension than a nail-biting patron at a good Hitchcock thriller’ (quoted in Blass, 2007). This was rather disingenuous, given his other descriptions of their reactions (see above). However, Milgram made a more measured response to the academic arguments. He pointed out, for instance, that he could not have known the outcome of the research before he started. As you already read, before embarking on the study he asked fellow professionals how they expected people to behave, and they predicted that participants would not continue to obey and administer severe shocks to the ‘learner’.

More importantly, Milgram was not oblivious to the psychological needs of his participants and was aware of the potential harm caused by the study. Immediately after the study, its true purpose was revealed to the participants. They were interviewed and given questionnaires to check they were all right. A friendly reconciliation was also arranged with the ‘victim’ whom they thought they had shocked. This procedure, known as debriefing , is commonplace today, but this was not the case in the 1960s. So, in this respect at least, Milgram was ahead of the game in terms of ethics procedures (Blass, 2004).

Milgram also conducted a follow-up survey of the participants one year after the study, to ensure that there was no long-term harm (Colman, 1987). The results showed that 84 per cent said they were ‘glad to have been in the experiment’, and only 1.3 per cent said they were very sorry to have taken part. Milgram also described how the participants had been examined by a psychiatrist who was unable to find a single participant who showed signs of long-term harm. Morris Braverman, a 39-year-old social worker, was one of the participants in Milgram’s experiment who continued to give shocks until the maximum was reached. He claimed, when interviewed a year after the experiment, that he had learned something of personal importance as a result of being in the experiment. His wife said, with reference to his willingness to obey orders, ‘You can call yourself an Eichmann’ (Milgram, 1974, p. 54).

Milgram’s basic defence was that the harm to the participants was not as great as it might appear, and for some of them the change in their understanding of their own behaviour and the behaviour of others was a positive event. He makes a further defence that we have to treat all people with respect and that this involves allowing them to make choices even if those choices are not always for the best. In direct response to Baumrind’s criticisms he wrote:

I started with the belief that every person who came to the laboratory was free to accept or to reject the dictates of authority. This view sustains a conception of human dignity insofar as it sees in each man a capacity for choosing his own behavior.

2.4 The judgement

So what do you think should be the final judgement on the ethics of Milgram’s study? As you can see from the debate between Milgram and Baumrind, ethics is something that psychologists debate and often disagree on. Ethics principles, like all rules, are subject to interpretation and disagreement.

And yet, while individuals might have their personal view about whether a piece of research is ethical or not, what really matters is the judgment of institutions that regulate the profession. In the USA the regulatory body is the American Psychological Association. Its equivalent in the UK is the British Psychological Society. These institutions have ethics committees which issue guidelines and codes of conduct related to ethics in research and can reprimand researchers who can be shown to have violated the rules. At the time of Milgram’s study, his research was investigated by the ethics committee of the American Psychological Association, who eventually came to the conclusion that it was ethically acceptable. Notably, however, Milgram’s studies could not be carried out today, as the ethics guidelines have become more restrictive since the 1960s.

Finally, one further issue regarding Milgram’s study is worth pointing out. Although the ethics of Milgram’s research have been questioned, it could be argued that the obedience study, more than any other study in psychology, demonstrated why ethics are important. Recall that what Milgram’s study showed was that ordinary people were willing to harm another human being just because they were told to do so by a person they believed was a psychologist, and because doing so was supposedly ‘required by the experiment’. This shows that people generally are ready to give scientists the benefit of the doubt and go along with what they are doing, even when it involves harming individuals. This in itself illustrates how important it is to have some moderation of scientific activity, and have limits imposed on what scientists can and cannot do.

2.5 Summary

  • Psychologists have a duty of care towards participants and must ensure that their wellbeing is preserved throughout a study.
  • Participants must be asked to give informed consent before taking part in research and have a right to withdraw at any point.
  • Milgram’s obedience studies kick-started an ethics debate in psychology and highlighted the need for the development of more stringent guidelines for the conduct of research psychologists.
  • Although Milgram’s obedience study was judged to be ethical at the time of publication, it would be in violation of the strict ethics guidelines in place today.

Activity 1: Ethics in psychological research

This activity introduces you to the idea of ethics in psychological research. When conducting research, psychologists cannot do what they like – they must abide by a code of conduct. Above all, the aim of this is to protect the welfare of the people participating in the research. It is important that you grasp the main ethics principles and their importance, as you will be returning to them throughout this course. At the end of this activity, you will find a handout summarising the main teaching points. You should download this and add it to your files.

Research ethics

When a psychologist is working out how to undertake a particular study, it is important that they consider whether what they are doing and how they are doing it is going to be ethical. All universities and research organisations have panels that judge whether research is ethical or not, and in the UK psychological research also needs to meet the requirements of the British Psychological Society (BPS), who specify a code of ethics and conduct, which includes:

  • Research should not include risks to the psychological wellbeing, physical health, personal values or dignity of participants.
  • Participants should give informed consent before taking part in research.
  • Participants should be able to stop participating in the research at any point.

The principles cover a number of other very important points, such as confidentiality, debriefing and protection. This activity is going to focus on the three principles above and you are going to have a go at applying them to a specific research project.

Task: Is it ethical?

On the following pages you will find a brief description of a psychological research study. Your task is to imagine you are on a research ethics panel that has been asked to consider the research being proposed. In each case, read the description and decide whether you think the study described is ethical in terms of the three principles described in the introduction:

  • Research should not threaten the psychological wellbeing, health, values or dignity of participants.

The ethics panel received the following proposal:

milgram experiment ethical considerations

This study is designed to discover what effect peer pressure might have on people’s voting behaviour. Phase one of the study will involve adding a question stating ‘which party did you vote for at the last election?’ to the end of an in-class exam being taken by twenty undergraduate students. Phase two will take place a week later in a scheduled seminar, and consist of asking the same students, one after another, to tell the group as a whole whom they had voted for.

Think about the three ethics principles and how you might expect a psychological study to meet each one, and then decide whether you think this study meets each ethics principle.

Ethics principle 1 - study 1.

Wellbeing, health, values and dignity

The correct answer is b.

This study threatened both the values and the dignity of the participants. UK law protects the right for voting to be secret for a good reason, so asking students how they voted in an exam and also to state this publicly contravenes the values attached to keeping how you voted a confidential matter. Putting students in a position where they might feel obliged either to state how they voted or indeed to lie about this as a result of peer pressure is also likely to have a negative effect on their dignity.

Ethics principle 2 - Study 1

Informed consent

At no point were the students told they were taking part in a research study, so they did not give their consent to participate.

Ethics principle 3 - Study 3

Right to withdraw

At no point were the students told they were taking part in a research study, so they did not realise there was any research to withdraw from.

Would you give this study ethical approval?

To be approved a study would need to meet all criteria. If it fails to meet just one, it should not gain approval.

As is stated in this course, applying ethical principles is never that straightforward and there are often cases where people have differing opinions. Don't worry if you had different answers; instead concentrate on which aspects of the study we have linked to each of the three ethical principles.

The ethics panel that considered Study 1 did not give approval. Following their feedback the psychologist revised the design of the study and resubmitted it.

milgram experiment ethical considerations

The revised study is to involve recruiting participants through a poster that will ask for volunteers to take part in a study on political communication. Volunteers will be informed that they have the right to withdraw from the study at any point and will be first asked to complete a ‘consent to participate’ form and then a questionnaire asking them about their background, likes and dislikes, and also whom they voted for in the last election. They will then attend a session that evening where the study will take place. The session will involve watching a party political broadcast and then answering questions in a group on how well the speaker communicated their ideas. At the end of the broadcast the researcher will apologise to the participants saying their questionnaires have all been accidentally lost, and ask them to state in front of the whole group of participants whom they voted for in the last election. By comparing how the participants said they had voted in the questionnaire and in front of the group, it will be possible to see if any had changed their mind as a result of being part of a group.

Ethics principle 1 - Study 2

The revised study still involves the participants’ being required to say how they voted publicly in front of a group of people; that they are still being asked the question at all can be seen as not respecting their values, and they are also still being put in a position where they might feel obliged to state how they voted, or indeed to lie about this as a result of peer pressure, and this is likely to have a negative effect on their dignity.

Ethics principle 2 - Study 2

Although the study now asks for volunteers and asks them to complete a consent form, this cannot be considered ‘informed’ consent because the researcher has not told them what the study is really about. Instead, the participants are being deceived about the real purpose of the study. A degree of deception is sometimes necessary in some forms of psychological research, but the degree of deception involved here is unethical as the participants are not being told in advance that they will be asked publicly about a confidential matter (voting behaviour).

Ethics principle 3 - Study 2

The correct answer is a.

Participants in this study would be aware that they were taking part in research, and were also told explicitly that they could withdraw from the study at any point. One view could be that as they did not know the true nature of the study, they were not able to withdraw from it, but this issue is dealt with under ‘informed consent’. Note that a study should not be given approval if it fails to meet any of the ethical principles.

To be approved a study would need to meet all criteria. If it fails to meet just one, it should not receive approval.

The ethics panel were still not convinced and did not give approval to Study 2 either. Undeterred, the researcher revised the design of the study and resubmitted it yet again.

milgram experiment ethical considerations

The new study will involve recruiting participants through a poster placed on a college noticeboard, which will ask for volunteers to take part in a study on ‘social pressures and voting behaviour’. Volunteers will be provided with an accurate summary of the proposed study which will also explain that they have the right to withdraw at any point, and should feel free to do so. After having read the summary they will be asked to sign a consent form. The study will involve providing participants with transcripts of fictitious election speeches from three candidates for Student Union President of a fictitious college. The participants will read the speeches and indicate on a form whom they would vote for. Following this, they will be told that an overwhelming number of students are voting for one of the other candidates and again asked to indicate on a form whom they would vote for.

Ethics principle 1 - Study 3

Participants in this study are not being asked to reveal how they voted in a real election and are also not having to reveal any shift in their voting behaviour to a group. This study is therefore very unlikely to have any negative effect on the participants' psychological wellbeing, health, values or dignity.

Ethics principle 2 - Study 3

The study now tells potential participants what is involved before asking for their consent. It is possible that telling the participants what the study is exploring will affect how they respond and is therefore unlikely to provide any useful results. However, judging the ethics of research is a different matter from judging whether the design of the study will produce useful results.

Participants in this study would be aware that they were taking part in research, and were also told explicitly that they could withdraw from the study at any point.

As the study met all of the criteria, it should be approved.

In the end this research was approved by the ethics committee. One thing that is important to bear in mind is that obtaining approval involves a dialogue between the researcher(s) and the relevant ethics committee. This dialogue invariably involves an interpretation of the principles and a negotiation of what can be regarded as acceptable research conduct.

This activity focused on three key ethics principles, namely the right to withdraw, informed consent and the wellbeing of participants. However, ethics panels evaluating real research need to take into consideration a more complex set of issues.

Here is a PDF summary of the activity to print or save in your files.

Activity 2: Researching animals and humans

This activity explores the ethics of animal research and the guidelines that govern the use of animals (although, as you will see, not all animals) in psychological research.

Psychology is often thought to be just about human beings. However, there are important areas of psychological research that involve non-human animals.

Research with non-human animals poses two important questions:

  • How relevant are studies of non-human animals to human psychology? Aren’t human beings unique and different from other animals?
  • Is it acceptable to carry out experiments on non-human animals in the interests of science?

The tasks in this online activity give you the opportunity to consider these questions for yourself and clarify your own opinions and understanding of the issues they raise. The activity will help you to recognise why psychologists carry out research with non-human animals and identify the ethical issues involved in such research.

milgram experiment ethical considerations

Why carry out research with animals?

You have been introduced to a number of ethics principles that apply to research on human participants. But why might psychologists want to do research with animals other than humans?

Here are some reasons. Select ‘Reveal comment’ to read a more detailed explanation:

To find out about the evolution of psychological functions .

By studying how different species adapt their behaviour to their environments, and trying to identify innate factors in adaptation, it is hoped that the interactions of genes, environments and learning can be understood better, thus shedding light on how evolution may have shaped human psychological processes.

To better understand psychological principles that apply across different species .

If similar processes are found in a range of species, this helps researchers to describe and (they hope) explain the basic principles of behaviour and other processes, such as attachment and learning.

To do experiments that would be unethical with humans .

Arguably, as you will learn later in this activity, it can be seen as less problematic to use non-human animals in experiments that involve potentially harmful conditions such as deprivation, pain or confinement, or for example, to explore the effects of punishment.

To better understand what is special about humans .

A substantial amount of research with non-human animals is carried out to identify psychological functions such as language and empathy, which other animals may not possess. Attempts to teach language to chimpanzees, for example, have met with only limited success, clarifying the specialised language abilities of humans.

The ethics of animal research

In this course you were introduced to a number of ethics principles that apply to research on human participants. These include informed consent from participants, ensuring their right to withdraw, protecting their welfare and evaluating the costs and benefits of the study.

Which of these do not apply to animals?

Of course, animals cannot give informed consent, and, given that most are kept in captivity anyway, the ‘right to withdraw’ does not really apply.

However, there are separate ethical guidelines for work with animals, which are also issued by the British Psychological Society.

They include provisions such as:

  • The ‘smallest number of animals sufficient to accomplish the research goals’ should be used in any study.
  • The costs and benefits of any study must be carefully evaluated.
  • The welfare of the animal must be taken into account and researchers must ‘seek to minimise any pain, suffering or distress that might arise’ from any experiment.
  • Researchers should use alternatives to animal research whenever possible, including data collected by other researchers, lower species (leeches, cell cultures, etc.) or, increasingly, computer simulations.

What emerges from these studies are the ‘3 Rs’ of animal research. These are to:

  • refine procedures to minimise suffering
  • reduce the number of individual animals used
  • replace animals with other alternatives.

These guidelines are interpreted and applied by ethics committees of research institutions and other bodies (including the Home Office) that grant special licences for keeping animals and using them in research.

Task 1: Evaluating the ethics of research

  • Use alternatives to animal research whenever possible, including data collected by other researchers, lower species (leeches, cell cultures, etc.) or increasingly, computer simulations.

Based on the guidelines on animal research, do you think an ethics committee would approve each of the following studies?

1. A researcher at a UK university is applying for a licence to replicate Harlow’s studies of deprivation. Infant monkeys would be raised in isolation with different types of ‘surrogate mother’ (inanimate objects that were either cloth-covered or made of wire with a milk bottle attached).

Do you think an ethics committee would give this study approval?

An ethics committee, and the Home Office, would probably not give approval for this study to be carried out. Not only does it raise issues about animal welfare and deprivation, but Harlow has already carried out this research, and it is highly unlikely that simply repeating the study would be considered a benefit great enough to outweigh the cost to the animals involved.

2. A researcher in a UK animal research laboratory is interested in addiction. He is proposing a study in which a small radio receiver would be implanted into the brains of seventy-five rhesus monkeys. These receivers would allow the researcher to activate areas of the brain thought to be associated with ‘pleasure’. The study would help shed light on brain mechanisms involved in addiction.

Several issues would need to be considered here. One is the number of monkeys involved. Ethics committees must ensure that the smallest number needed is actually used in the study. It is unlikely that as many as seventy-five monkeys would be absolutely necessary, so the researcher would probably be asked to make a very strong case, or reduce the number. Also, the ethics committee would want to hear what would happen to the monkeys after the experiment. Would they be able to live normally after the experiment is over?

3. A laboratory has been contracted by the Ministry of Defence to evaluate whether pigeons could be used in a guided missile to direct it towards an enemy aircraft in order to destroy it. The research would involve three pigeons being trained in a Skinner box to peck at targets on a radar screen.

Believe it or not, a study such as this, which drew on the ideas of B.F. Skinner, was carried out in the USA during the Second World War, as part of the so-called ‘Project Orcon’ or ‘Pigeon Project’. In this study there are no obvious ethical concerns (except for the more general issues of animal welfare), given that the pigeons’ ability to guide a missile is simply being evaluated, and a standard Skinner-box procedure is being proposed. So the project would probably receive ethical clearance, as the committee are likely to consider it to involve a discrimination learning task using specific stimuli (target on a radar screen), and there are no obvious animal welfare issues involved. (The pigeons would not actually be used in the attacks, or at least that is not what this project is about.)

4. A researcher has applied for permission to carry out an observational study in the Kalahari Desert of communication among pied babblers (wild birds) and with other species. The study would involve two researchers wearing camouflage hiding in the bushes, observing the behaviour of the pied babblers and recording their mating calls.

Research such as this is being done regularly around the world. While most of the ethical guidance that you have learned about so far in this activity has referred to laboratory work, the BPS also regulates observational work in the animals’ natural habitats. Researchers would be asked by the committee to follow careful protocols to ensure that disruption to animals’ lives is kept to a minimum. Disruption might have a detrimental effect not only on the birds’ lives, but might also impact on the inferences that can be drawn from the study. This is because the observed behaviour might be a reaction to intrusion by researchers, rather than something that occurs naturally, in the wild.

Different animals, different guidelines

Although human beings are animals, for most people there is a strong conceptual division between human and non-human animals. Humans – or homo sapiens – are seen as being fundamentally different even from pan paniscus (the bonobo chimpanzee (Figure 9) – genetically our closest relative in the animal world).

milgram experiment ethical considerations

It is this widespread belief that humans occupy a special place in nature that underpins the whole notion that it might be appropriate to carry out some types of research with non-human animals, where human research ethics would not permit such research to be done with humans.

As you have seen, it’s not that there are no ethical considerations in research with other species, but rather that different, and less stringent, considerations apply.

There is, however, a further issue here. It is not just that humans are believed to be different from non-human animals. There are also differences between animals. In fact, what is meant by the word ‘animal’? While there are many people who believe that the animal world ends with the bonobo chimpanzee, we rarely think about where it begins.

The final task in this activity encourages you to explore this question, to examine your own views and to critically consider the issues involved.

Task 2: Reflecting on different species

milgram experiment ethical considerations

Have a look at the list below and consider the following:

  • Which one would you consider to be an ‘animal’?
  • Which ones you would permit animal research on (with BPS guidance in mind)?

For each answer try to think of reasons why you referred to some of these species as ‘animals’ but not others, or why you think that research on some might be more appropriate than others. Once you have indicated your responses, go to the next page, where you will find the current legal status of research on these species and read about some of the issues involved in determining which animals can and cannot be used in research.

Animal research and the law

Only one of the organisms included in the list is not an animal: the E. coli bacterium. At least, that is the straightforward scientific answer. However, you may be surprised to learn that in many countries the law regulates what is, and what is not, an animal – at least where research is concerned. Such definitions determine which species are covered by the guidelines for psychologists working with animals. According to the Animals (Scientific Procedures) Act 1986, which legislates for research in the UK, only species that are vertebrates (possess a spinal column) and one single invertebrate, the octopus ( octopus vulgaris ), are legally defined as ‘animals’ when it comes to research. The same category of living things is also protected by the Animal Welfare Act 2006. Under this definition, the only ‘animals’ in the list are the rat, cat, dog, horse, dolphin and gorilla: the coral, woodlouse, crab, mosquito and scorpion are not legally ‘animals’. As for the second question, there are really no right or wrong answers. You were asked ‘which ones you would permit animal research on’ and that is a matter of personal ethics. However, you now know which of the species listed are protected by law (at least in the UK).

So there are many animals, or should we say ‘organisms’, that are not covered by any legislation, and for research purposes do not count as ‘animals’. There is no law protecting lobsters, spiders or mosquitoes (and especially not the E. coli bacteria) from being used in any way whatsoever in research. The British Psychological Society (BPS) does recommend that the ethics standards of work with protected species should also be maintained with organisms not covered by the relevant legislation, but this is just a recommendation.

Even among the vertebrates, further distinctions are made. Can you guess which animals receive extra protection? It is horses, cats, dogs and primates. In the case of primates, the reason is undoubtedly their genetic proximity to humans and the frequency with which they are used in research, but what about cats, dogs and horses?

The reason is that these animals are regarded by many humans as occupying a special place in the animal world, given that they are kept as pets. However, humans don’t always agree on which animals are pets. Whereas most people in the UK would not consider eating a dog, in parts of Asia dogs are considered a delicacy. Horses, donkeys and a host of other animals that people in some countries treat as companions adorn many a menu in other countries. So, some animals are a ‘man’s (or a woman’s) best friend’ whilst others are ‘vermin’ or ‘food’. In the same way, some species are regarded as ‘beautiful’, ‘cute’ or ‘intelligent’, while others are treated as less so. What all of this suggests is that differentiation within the animal world is not always based on strict scientific criteria, but rather on cultural sensitivities.

milgram experiment ethical considerations

In this activity you were given an opportunity to consider a number of questions relevant to psychological research on animals. These included why psychologists study animals, how the rights of the animals used in research are protected, and how the appropriateness of animals for research is determined.

However, the controversy surrounding this research, especially the study on the effects of deprivation, had a different kind of impact on psychology. It made researchers more aware of the need to regulate research on animals and treat them more humanely. It was therefore in the aftermath of Harlow’s study that rules guiding psychological research on animals began to be tightened.

Activity 3: Researching animals

You have just learned about the ethics of animal research; now you have the opportunity of hearing from two psychologists working with animals in two films, Researching Animals ..

Reading about research that psychologists have conducted with animals is often fascinating, but seeing how the research is conducted is even better. Film A (8 minutes) introduces the work of Alex Thornton with meerkats; in Film B (17 minutes) you will learn about Tetsuro Matsuzawa’s work with chimpanzees.

First, watch the two films without interruption. After doing so, read ‘Issues to consider’ below, then watch them again, keeping these issues in mind. Make sure that you take some notes.

Copy this transcript to the clipboard

Transcript: Film A: Researching Animals (meerkats)

Transcript: film b: researching animals (chimpanzees), issues to consider.

  • Alex Thornton has looked at how meerkats teach their young to catch scorpions. To what extent is this process similar to or different from that which human parents use?
  • Tetsuro Matsuzawa found that, when it comes to completing the photographic memory task, chimpanzees are superior to humans. What explanation does he give for this finding? Can the performance of the chimpanzees be explained by conditioning?
  • Think about the location where the two researchers conduct their studies. What are the advantages and disadvantages of researching animals in their natural habitat compared to captivity?
  • Compare the reasons why Alex Thornton and Tetsuro Matsuzawa study animals. Which of them is interested in animal behaviour not just for its own sake, but also as a way of learning about human capacities?

This free course provided an introduction to studying sociology. It took you through a series of exercises designed to develop your approach to study and learning at a distance and helped to improve your confidence as an independent learner.

Acknowledgements

Grateful acknowledgement is made to the following sources.

Philip Banyard

Course image: Phil Dolby in Flickr made available under Creative Commons Attribution 2.0 Licence .

Figures 1, 2, 3 and 4: Milgram, S (1974) Obedience to Authority, Harper & Row. Copyright © 1974 by Stanley Milgram

Figure 5: Copyright © Andresr/iStock

Figure 6: Copyright © Michael Flippo/iStock

Angel Eye Media

With thanks to:

Kalahari Meerkat Project

iBrain Festival of Neuroscience at

Muziekcentrum De Bijloke

Primate Research Institute, Kyoto University.

Don't miss out:

If reading this text has inspired you to learn more, you may be interested in joining the millions of people who discover our free learning resources and qualifications by visiting The Open University - www.open.edu/ openlearn/ free-courses

Copyright © 2016 The Open University

  • Bipolar Disorder
  • Therapy Center
  • When To See a Therapist
  • Types of Therapy
  • Best Online Therapy
  • Best Couples Therapy
  • Managing Stress
  • Sleep and Dreaming
  • Understanding Emotions
  • Self-Improvement
  • Healthy Relationships
  • Student Resources
  • Personality Types
  • Guided Meditations
  • Verywell Mind Insights
  • 2024 Verywell Mind 25
  • Mental Health in the Classroom
  • Editorial Process
  • Meet Our Review Board
  • Crisis Support

The Process of Conducting Ethical Research in Psychology

Tom Merton / Getty Images

Earlier in psychology history, many experiments were performed with highly questionable and even outrageous violations of ethical considerations. Milgram's infamous obedience experiment , for example, involved deceiving human subjects into believing that they were delivering painful, possibly even life-threatening, electrical shocks to another person.

These controversial psychology experiments played a major role in the development of the ethical guidelines and regulations that psychologists must abide by today. When performing studies or experiments that involve human participants, psychologists must submit their proposal to an institutional review board (IRB) for approval. ​These committees help ensure that experiments conform to ethical and legal guidelines.

Ethical codes, such as those established by the American Psychological Association, are designed to protect the safety and best interests of those who participate in psychological research. Such guidelines also protect the reputations of psychologists, the field of psychology itself and the institutions that sponsor psychology research.

Ethical Guidelines for Research With Human Subjects

When determining ethical guidelines for research , most experts agree that the cost of conducting the experiment must be weighed against the potential benefit to society the research may provide. While there is still a great deal of debate about ethical guidelines, there are some key components that should be followed when conducting any type of research with human subjects.

Participation Must Be Voluntary

All ethical research must be conducted using willing participants.   Study volunteers should not feel coerced, threatened or bribed into participation. This becomes especially important for researchers working at universities or prisons, where students and inmates are often encouraged to participate in experiments.

Researchers Must Obtain Informed Consent

Informed consent is a procedure in which all study participants are told about procedures and informed of any potential risks.   Consent should be documented in written form. Informed consent ensures that participants know enough about the experiment to make an informed decision about whether or not they want to participate.

Obviously, this can present problems in cases where telling the participants the necessary details about the experiment might unduly influence their responses or behaviors in the study. The use of deception in psychology research is allowed in certain instances, but only if the study would be impossible to conduct without the use of deception, if the research will provide some sort of valuable insight and if the subjects will be debriefed and informed about the study's true purpose after the data has been collected.

Researchers Must Maintain Participant Confidentiality

Confidentiality is an essential part of any ethical psychology research.   Participants need to be guaranteed that identifying information and individual responses will not be shared with anyone who is not involved in the study.

While these guidelines provide some ethical standards for research, each study is different and may present unique challenges. Because of this, most colleges and universities have a Human Subjects Committee or Institutional Review Board that oversees and grants approval for any research conducted by faculty members or students. These committees provide an important safeguard to ensure academic research is ethical and does not pose a risk to study participants.

American Psychological Association. Ethical principles of psychologists and code of conduct .

By Kendra Cherry, MSEd Kendra Cherry, MS, is a psychosocial rehabilitation specialist, psychology educator, and author of the "Everything Psychology Book."

Encyclopedia Britannica

  • History & Society
  • Science & Tech
  • Biographies
  • Animals & Nature
  • Geography & Travel
  • Arts & Culture
  • Games & Quizzes
  • On This Day
  • One Good Fact
  • New Articles
  • Lifestyles & Social Issues
  • Philosophy & Religion
  • Politics, Law & Government
  • World History
  • Health & Medicine
  • Browse Biographies
  • Birds, Reptiles & Other Vertebrates
  • Bugs, Mollusks & Other Invertebrates
  • Environment
  • Fossils & Geologic Time
  • Entertainment & Pop Culture
  • Sports & Recreation
  • Visual Arts
  • Demystified
  • Image Galleries
  • Infographics
  • Top Questions
  • Britannica Kids
  • Saving Earth
  • Space Next 50
  • Student Center

Stanley Milgram

  • When did science begin?
  • Where was science invented?

Blackboard inscribed with scientific formulas and calculations in physics and mathematics

Milgram experiment

Our editors will review what you’ve submitted and determine whether to revise the article.

  • Open University - OpenLearn - Psychological research, obedience and ethics: 1 Milgram’s obedience study
  • Social Science LibreTexts - The Milgram Experiment- The Power of Authority
  • Verywell Mind - What was the Milgram Experiment?
  • BCcampus Open Publishing - Ethics in Law Enforcement - The Milgram Experiment
  • Nature - Modern Milgram experiment sheds light on power of authority
  • SimplyPsychology - Stanley Milgram Shock Experiment: Summary, Results, & Ethics
  • University of California - College of Natural Resources - Milgrams Experiment on Obedience to Authority

Milgram experiment , controversial series of experiments examining obedience to authority conducted by social psychologist Stanley Milgram . In the experiment, an authority figure, the conductor of the experiment, would instruct a volunteer participant, labeled the “teacher,” to administer painful, even dangerous, electric shocks to the “learner,” who was actually an actor. Although the shocks were faked, the experiments are widely considered unethical today due to the lack of proper disclosure, informed consent, and subsequent debriefing related to the deception and trauma experienced by the teachers. Some of Milgram’s conclusions have been called into question. Nevertheless, the experiments and their results have been widely cited for their insight into how average people respond to authority.

Milgram conducted his experiments as an assistant professor at Yale University in the early 1960s. In 1961 he began to recruit men from New Haven , Connecticut , for participation in a study he claimed would be focused on memory and learning . The recruits were paid $4.50 at the beginning of the study and were generally between the ages of 20 and 50 and from a variety of employment backgrounds. When they volunteered, they were told that the experiment would test the effect of punishment on learning ability. In truth, the volunteers were the subjects of an experiment on obedience to authority. In all, about 780 people, only about 40 of them women, participated in the experiments, and Milgram published his results in 1963.

Milgram experiment

Volunteers were told that they would be randomly assigned either a “teacher” or “learner” role, with each teacher administering electric shocks to a learner in another room if the learner failed to answer questions correctly. In actuality, the random draw was fixed so that all the volunteer participants were assigned to the teacher role and the actors were assigned to the learner role. The teachers were then instructed in the electroshock “punishment” they would be administering, with 30 shock levels ranging from 15 to 450 volts. The different shock levels were labeled with descriptions of their effects, such as “Slight Shock,” “Intense Shock,” and “Danger: Severe Shock,” with the final label a grim “XXX.” Each teacher was given a 45-volt shock themselves so that they would better understand the punishment they believed the learner would be receiving. Teachers were then given a series of questions for the learner to answer, with each incorrect answer generally earning the learner a progressively stronger shock. The actor portraying the learner, who was seated out of sight of the teacher, had pre-recorded responses to these shocks that ranged from grunts of pain to screaming and pleading, claims of suffering a heart condition, and eventually dead silence. The experimenter, acting as an authority figure, would encourage the teachers to continue administering shocks, telling them with scripted responses that the experiment must continue despite the reactions of the learner. The infamous result of these experiments was that a disturbingly high number of the teachers were willing to proceed to the maximum voltage level, despite the pleas of the learner and the supposed danger of proceeding.

Milgram’s interest in the subject of authority, and his dark view of the results of his experiments, were deeply informed by his Jewish identity and the context of the Holocaust , which had occurred only a few years before. He had expected that Americans, known for their individualism , would differ from Germans in their willingness to obey authority when it might lead to harming others. Milgram and his students had predicted only 1–3% of participants would administer the maximum shock level. However, in his first official study, 26 of 40 male participants (65%) were convinced to do so and nearly 80% of teachers that continued to administer shocks after 150 volts—the point at which the learner was heard to scream—continued to the maximum of 450 volts. Teachers displayed a range of negative emotional responses to the experiment even as they continued to obey, sometimes pleading with the experimenters to stop the experiment while still participating in it. One teacher believed that he had killed the learner and was moved to tears when he eventually found out that he had not.

Milgram experiment

Milgram included several variants on the original design of the experiment. In one, the teachers were allowed to select their own voltage levels. In this case, only about 2.5% of participants used the maximum shock level, indicating that they were not inclined to do so without the prompting of an authority figure. In another, there were three teachers, two of whom were not test subjects, but instead had been instructed to protest against the shocks. The existence of peers protesting the experiment made the volunteer teachers less likely to obey. Teachers were also less likely to obey in a variant where they could see the learner and were forced to interact with him.

The Milgram experiment has been highly controversial, both for the ethics of its design and for the reliability of its results and conclusions. It is commonly accepted that the ethics of the experiment would be rejected by mainstream science today, due not only to the handling of the deception involved but also to the extreme stress placed on the teachers, who often reacted emotionally to the experiment and were not debriefed . Some teachers were actually left believing they had genuinely and repeatedly shocked a learner before having the truth revealed to them later. Later researchers examining Milgram’s data also found that the experimenters conducting the tests had sometimes gone off-script in their attempts to coerce the teachers into continuing, and noted that some teachers guessed that they were the subjects of the experiment. However, attempts to validate Milgram’s findings in more ethical ways have often produced similar results.

APS

Member Article

Replicating milgram.

  • Institutional Review Board (IRB)
  • IRBs: Navigating the Maze

Last month, we featured IRB best practices (“IRBs: Navigating the Maze” November 2007 Observer ), and got the ball rolling with strategies and tips that psychological scientists have found to work. Here, we continue the dissemination effort with the second of three articles by researchers who share their experiences with getting their research through IRB hoops. Jerry Burger from Santa Clara University managed to do the seemingly impossible — he conducted a partial replication of the infamous Milgram experiment. Read on for valuable advice, and look for similar coverage in upcoming Observers .

“It can’t be done.”

These are the first words I said to Muriel Pearson, producer for ABC News’ Primetime , when she approached me with the idea of replicating Stanley Milgram’s famous obedience studies. Milgram’s work was conducted in the early 1960s before the current system of professional guidelines and IRBs was in place. It is often held up as the prototypic example of why we need policies to protect the welfare of research participants. Milgram’s participants were placed in an emotionally excruciating situation in which an experimenter instructed them to continue administering electric shocks to another individual despite hearing that person’s agonizing screams of protest. The studies ignited a debate about the ethical treatment of participants. And the research became, as I often told my students, the study that can never be replicated.

Nonetheless, I was intrigued. Although more than four decades have passed since Milgram conducted his research, his obedience studies continue to occupy an important place in social psychology textbooks and classes. The haunting black-and-white images of ordinary citizens delivering what appear to be dangerous, if not deadly, electric shocks and the implications of the findings for atrocities like the Holocaust and Abu Ghraib are not easily dismissed. Yet because Milgram’s procedures are clearly out-of-bounds by today’s ethical standards, many questions about the research have gone unanswered. Chief among these is one that inevitably surfaces when I present Milgram’s findings to students: Would people still act that way today?

The challenge was to develop a variation of Milgram’s procedures that would allow useful comparisons with the original investigations while protecting the well-being of the participants. But meeting this challenge would raise another: I would also need to assuage the apprehension my IRB would naturally experience when presented with a proposal to replicate the study that can never be replicated.

I went to great lengths to recreate Milgram’s procedures (Experiment Five), including such details as the words used in the memory test and the experimenter’s lab coat. But I also made several substantial changes. First, we stopped the procedures at the 150-volt mark. This is the first time participants heard the learner’s protests through the wall and his demands to be released. When we look at Milgram’s data, we find that this point in the procedure is something of a “point of no return.” Of the participants who continued past 150 volts, 79 percent went all the way to the highest level of the shock generator (450 volts). Knowing how people respond up to this point allowed us to make a reasonable estimate of what they would do if allowed to continue to the end.

Stopping the study at this juncture also avoided exposing participants to the intense stress Milgram’s participants often experienced in the subsequent parts of the procedure.

Second, we used a two-step screening process for potential participants to exclude any individuals who might have a negative reaction to the experience. Potential participants were asked in an initial phone interview if they had ever been diagnosed with a psychiatric disorder; if they were currently receiving psychotherapy; if they were currently taking any medications for emotional difficulties; if they had any medical conditions that might be affected by stress; if they ever had any problems with alcohol or drug use; and if they had ever experienced serious trauma, such as child abuse, domestic violence, or combat. Individuals who responded “yes” to any of these questions (about 30 percent) were excluded from the study. During the second step in the screening process, participants completed measures of anxiety and depression and were interviewed in person by a licensed clinical psychologist. The clinicians were shown the anxiety and depression data and were allowed to interview participants for as long as needed (about 30 minutes on average). The clinicians were instructed to err on the side of caution and to exclude anyone who they judged might have a negative reaction to the experiment procedures. More than 38 percent of the interviewed participants were excluded at this point.

Third, participants were told at least three times (twice in writing) that they could withdraw from the study at any time and still receive their $50 for participation. Fourth, like Milgram, we administered a sample shock to our participants (with their consent). However, we administered a very mild 15-volt shock rather than the 45-volt shock Milgram gave his participants. Fifth, we allowed virtually no time to elapse between ending the session and informing participants that the learner had received no shocks. Within a few seconds after ending the study, the learner entered the room to reassure the participant he was fine. Sixth, the experimenter who ran the study also was a clinical psychologist who was instructed to end the session immediately if he saw any signs of excessive stress. Although each of these safeguards came with a methodological price (e.g., the potential effect of screening out certain individuals, the effect of emphasizing that participants could leave at any time), I wanted to take every reasonable measure to ensure that our participants were treated in a humane and ethical manner.

Of course, I also needed IRB approval. I knew from my own participation on the IRB that the proposal would be met with concern and perhaps a little fear by the board’s members. I work at a relatively small university, and our IRB consists of individuals from a variety of academic backgrounds. I knew that few members would be comfortable or confident when assessing a potentially controversial proposal from another discipline. Given the possibility of a highly visible mistake, the easy response would have been to say “no.” To address these concerns, I created a list of individuals who were experts on Milgram’s studies and the ethical questions surrounding this research. I offered to make this list available to the IRB. More important, Steven Breckler, a social psychologist who currently serves as the executive director for science at the American Psychological Association, graciously provided an assessment of the proposal’s ethical issues that I shared with the IRB.

In the end, all the extra steps and precautions paid off. The IRB carefully reviewed and then approved the procedures. More than a year after collecting the data, I have no indication that any participant was harmed by his or her participation in the study. On the contrary, I was constantly surprised by participants’ enthusiasm for the research both during the debriefing and in subsequent communications. We also produced some interesting findings. Among other things, we found that today people obey the experimenter in this situation at about the same rate they did 45 years ago. ABC devoted an entire 60-minute Primetime broadcast to the research and its implications. Finally, it is my hope that other investigators will use the 150-volt procedure and thereby jump-start research on some of the important questions that motivated Stanley Milgram nearly half a century ago. ♦

milgram experiment ethical considerations

So far the trials seemed to have been conducted on people Americans. Might you not get significantly different results if trials are conducted on people living in other countries, eg Singapore?

milgram experiment ethical considerations

Lau Kim Boo, You made a very good point! Different cultures may react differently. I think in order to get valid results, for any study involving psychology or the make up of a human being, broad, diversified research, evaluations need to be conducted. I, for one, would really like to see the results of this ‘universal’ experiment! Thank you

milgram experiment ethical considerations

Hi! I was reading through your post and think its great that you were able to take Milgram’s study, revise it and have it approved by the IRB. However, my question is that your study still contained deception, possible coercion, and psychological anguish. Why did the IRB approve your study? and if you feel as though it did not contain these things, can you please tell me why?

milgram experiment ethical considerations

In Jerry Burger’s study he actually used a diverse amount of people from all cultures. In Milgram’s and Burger’s experiments they realized that this was a universal experience.

milgram experiment ethical considerations

“So far the trials seemed to have been conducted on people Americans. ”

That is not correct. It has been replicated in Germany and in Australia.

APS regularly opens certain online articles for discussion on our website. Effective February 2021, you must be a logged-in APS member to post comments. By posting a comment, you agree to our Community Guidelines and the display of your profile information, including your name and affiliation. Any opinions, findings, conclusions, or recommendations present in article comments are those of the writers and do not necessarily reflect the views of APS or the article’s author. For more information, please see our Community Guidelines .

Please login with your APS account to comment.

About the Author

Jerry Burger is a professor of psychology at Santa Clara University. His research interests include social influence, particularly compliance, and the perception of and motivation for personal control.

milgram experiment ethical considerations

How the Classics Changed Research Ethics

Some of history’s most controversial psychology studies helped drive extensive protections for human research participants. Some say those reforms went too far.

Using an Automated Wizard to Process Minimal-Risk Research

A user-friendly decision tool could help researchers identify minimal-risk research, easing the administrative burden on both investigators and institutional reviewers.

Department of Health & Human Services

Revision to the Common Rule: Implications for Behavioral and Social Sciences Research

The various provisions of the US government’s revised rule on human subjects research will provide increased flexibility and less burden for behavioral and social sciences researchers. William T. Riley and Farheen Akbar from the Office of Behavior and Social Sciences Research explain.

Privacy Overview

CookieDurationDescription
__cf_bm30 minutesThis cookie, set by Cloudflare, is used to support Cloudflare Bot Management.
CookieDurationDescription
AWSELBCORS5 minutesThis cookie is used by Elastic Load Balancing from Amazon Web Services to effectively balance load on the servers.
CookieDurationDescription
at-randneverAddThis sets this cookie to track page visits, sources of traffic and share counts.
CONSENT2 yearsYouTube sets this cookie via embedded youtube-videos and registers anonymous statistical data.
uvc1 year 27 daysSet by addthis.com to determine the usage of addthis.com service.
_ga2 yearsThe _ga cookie, installed by Google Analytics, calculates visitor, session and campaign data and also keeps track of site usage for the site's analytics report. The cookie stores information anonymously and assigns a randomly generated number to recognize unique visitors.
_gat_gtag_UA_3507334_11 minuteSet by Google to distinguish users.
_gid1 dayInstalled by Google Analytics, _gid cookie stores information on how visitors use a website, while also creating an analytics report of the website's performance. Some of the data that are collected include the number of visitors, their source, and the pages they visit anonymously.
CookieDurationDescription
loc1 year 27 daysAddThis sets this geolocation cookie to help understand the location of users who share the information.
VISITOR_INFO1_LIVE5 months 27 daysA cookie set by YouTube to measure bandwidth that determines whether the user gets the new or old player interface.
YSCsessionYSC cookie is set by Youtube and is used to track the views of embedded videos on Youtube pages.
yt-remote-connected-devicesneverYouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
yt-remote-device-idneverYouTube sets this cookie to store the video preferences of the user using embedded YouTube video.
yt.innertube::nextIdneverThis cookie, set by YouTube, registers a unique ID to store data on what videos from YouTube the user has seen.
yt.innertube::requestsneverThis cookie, set by YouTube, registers a unique ID to store data on what videos from YouTube the user has seen.

milgram experiment ethical considerations

Reference Library

Collections

  • See what's new
  • All Resources
  • Student Resources
  • Assessment Resources
  • Teaching Resources
  • CPD Courses
  • Livestreams

Study notes, videos, interactive activities and more!

Psychology news, insights and enrichment

Currated collections of free resources

Browse resources by topic

  • All Psychology Resources

Resource Selections

Currated lists of resources

Study Notes

Explanations for Obedience - Milgram (1963)

Last updated 22 Mar 2021

  • Share on Facebook
  • Share on Twitter
  • Share by Email

Milgram (1963) conducted one of the most famous and influential psychological investigations of obedience. He wanted to find out if ordinary American citizens would obey an unjust order from an authority figure and inflict pain on another person because they were instructed to.

Milgram’s sample consisted of 40 male participants from a range of occupations and backgrounds. The participants were all volunteers who had responded to an advert in a local paper, which offered $4.50 to take part in an experiment on ‘punishment and learning’.

The 40 participants were all invited to a laboratory at Yale University and upon arrival they met with the experimenter and another participant, Mr Wallace, who were both confederates.

The experimenter explained that one person would be randomly assigned the role of teacher and the other, a learner. However, the real participant was always assigned the role of teacher. The experimenter explained that the teacher, the real participant, would read the learner a series of word pairs and then test their recall. The learner, who was positioned in an adjacent room, would indicate his choice using a system of lights. The teacher was instructed to administer an electric shock ever time the learner made a mistake and to increase the voltage after each mistake.

The teacher watched the learner being strapped to the electric chair and was given a sample electric shock to convince them that the procedure was real. The learner wasn’t actually strapped to the chair and gave predetermined answers to the test. As the electric shocks increased the learner’s screams, which were recorded, became louder and more dramatic. At 180 volts the learner complained of a weak heart. At 300 volts he banged on the wall and demanded to leave and at 315 volts he became silent, to give the illusions that was unconscious, or even dead.

The experiment continued until the teacher refused to continue, or 450 volts was reached. If the teacher tried to stop the experiment, the experimenter would respond with a series of prods, for example: ‘The experiment requires that you continue.’ Following the experiment the participants were debriefed.

Milgram found that all of the real participants went to at least 300 volts and 65% continued until the full 450 volts. He concluded that under the right circumstances ordinary people will obey unjust orders.

Milgram’s study has been heavily criticised for breaking numerous ethical guidelines, including: deception , right to withdraw and protection from harm.

Milgram deceived his participants as he said the experiment was on ‘punishment and learning’, when in fact he was measuring obedience, and he pretended the learner was receiving electric shocks. In addition, it was very difficult for participants with withdraw from the experiment, as the experimenter prompted the participants to continue. Finally, many of the participants reported feeling exceptionally stressed and anxious while taking part in the experiment and therefore they were not protect from psychological harm. This is an issue, as Milgram didn’t respect his participants, some of whom felt very guilt following the experiment, knowing that they could have harmed another person. However, it must be noted that it was essential for Milgram to deceive his participants and remove their right to withdraw to test obedience and produce valid results. Furthermore, he did debrief his participants following the experiment and 83.7% of participants said that they were happy to have taken part in the experiment and contribute to scientific research.

Milgram’s study has been criticised for lacking ecological validity. Milgram tested obedience in a laboratory, which is very different to real-life situations of obedience, where people are often asked to follow more subtle instructions, rather than administering electric shocks. As a result we are unable to generalise his findings to real life situations of obedience and cannot conclude that people would obey less severe instructions in the same way.

Finally, Milgram’s research lacked population validity. Milgram used a bias sample of 40 male volunteers, which means we are unable to generalise the results to other populations, in particular females, and cannot conclude if female participants would respond in a similar way.

  • Ethical Issues
  • Right to withdraw
  • Protection from harm

You might also like

milgram experiment ethical considerations

Great lesson starter for obedience!

16th February 2016

Conformity - Asch (1951)

Conformity to social roles as investigated by zimbardo, dispositional explanation for obedience: authoritarian personality, dispositional explanation for obedience: the authoritarian personality.

Quizzes & Activities

Resistance to Social Change

Video: the authority hoax - you don't know what you would do.

1st November 2016

Lights, Camera, Action! A Shocking Rendition of Milgram

28th January 2017

Our subjects

  • › Criminology
  • › Economics
  • › Geography
  • › Health & Social Care
  • › Psychology
  • › Sociology
  • › Teaching & learning resources
  • › Student revision workshops
  • › Online student courses
  • › CPD for teachers
  • › Livestreams
  • › Teaching jobs

Boston House, 214 High Street, Boston Spa, West Yorkshire, LS23 6AD Tel: 01937 848885

  • › Contact us
  • › Terms of use
  • › Privacy & cookies

© 2002-2024 Tutor2u Limited. Company Reg no: 04489574. VAT reg no 816865400.

  • Foundations
  • Write Paper

Search form

  • Experiments
  • Anthropology
  • Self-Esteem
  • Social Anxiety

milgram experiment ethical considerations

  • Foundations >

Milgram Experiment Ethics

In recent years, psychologists and social scientists have begun to question the Milgram experiment ethics, and whether the experiment should have been allowed at all.

This article is a part of the guide:

  • Social Psychology Experiments
  • Milgram Experiment
  • Bobo Doll Experiment
  • Stanford Prison Experiment
  • Asch Experiment

Browse Full Outline

  • 1 Social Psychology Experiments
  • 2.1 Asch Figure
  • 3 Bobo Doll Experiment
  • 4 Good Samaritan Experiment
  • 5 Stanford Prison Experiment
  • 6.1 Milgram Experiment Ethics
  • 7 Bystander Apathy
  • 8 Sherif’s Robbers Cave
  • 9 Social Judgment Experiment
  • 10 Halo Effect
  • 11 Thought-Rebound
  • 12 Ross’ False Consensus Effect
  • 13 Interpersonal Bargaining
  • 14 Understanding and Belief
  • 15 Hawthorne Effect
  • 16 Self-Deception
  • 17 Confirmation Bias
  • 18 Overjustification Effect
  • 19 Choice Blindness
  • 20.1 Cognitive Dissonance
  • 21.1 Social Group Prejudice
  • 21.2 Intergroup Discrimination
  • 21.3 Selective Group Perception

This notorious experiment was designed as a response to the notorious trials of Nazi war criminals, who claimed that they were 'just following orders'. Milgram wanted to establish whether people really would obey authority figures, even when the instructions given were morally wrong.

The main thing to remember, when judging the experiment is that modern day criticisms have the benefit of hindsight. A few decades ago, Europe had been mentally scarred by the atrocities committed during the Second World War, and was looking for answers.

Even a few years later, in the 1960s, these wounds remained; as a Jew himself, Milgram was trying to establish whether the claim of war-criminals, that they were just obeying orders, was a reasonable defense or not.

At the time, the Milgram experiment ethics seemed reasonable, but by the stricter controls in modern psychology, this experiment would not be allowed today. Milgram's generation needed conclusive answers about the 'final solution', and some closure on this chapter of human history. Was human nature inherently evil or could reasonable people be coerced by authority into unnatural actions?

The Milgram experiment once again became relevant in the 1970s Mai Lai massacre , with society questioning the motives behind the, as well as other atrocities committed by the Americans in Vietnam. Whilst the actions of the soldiers concerned cannot be condoned, it showed the horrible effects on the psyche, and morals, of young men when they are exposed to death and suffering on a daily basis.

The main concerns raised about the Milgram Experiment ethics are based on a number of factors.

Modern ethical standards assert that participants in any experiment must not be deceived, and that they must be made aware of any consequences. In the interest of fairness, follow up research, performed after the experiment, indicated that there were no long term psychological effects on the participants.

However, the fact that these people thought that they had caused suffering to another human being, could have caused severe emotional distress.

Whilst the Milgram experiment appeared to have no long term effects on the participants, it is essential that psychological studies do have strict guidelines; the Stanford Prison Experiment is an example of one such study that crossed the line, and actually caused measurable psychological distress to the participants.

In some cases, this emotional scarring lasted for months and years after the study, so questioning the Milgram Experiment ethics is a necessary part of science.

The long term effect of carrying on performing similar studies would be destructive for research without deception . Much of psychology researches explore areas where the involved subjects are more skeptical than the average of the population.

They might never agree to participate in research which is harmless or aimed to help the subjects if it was well known that researchers commonly use deception.

In conclusion, whilst there is no doubt that the experiment, in its original format, would not be allowed, it is important to remember that Stanley Milgram was not a bad person. He was genuinely trying to uncover the reasons why humans could become embroiled in great evil.

In modern times, with questionable practices being carried out at Abu Ghraib and Guantanamo Bay , the Milgram and Stanford Prison experiments are once again becoming relevant.

  • Psychology 101
  • Flags and Countries
  • Capitals and Countries

Martyn Shuttleworth (Jul 4, 2008). Milgram Experiment Ethics. Retrieved Sep 26, 2024 from Explorable.com: https://explorable.com/milgram-experiment-ethics

You Are Allowed To Copy The Text

The text in this article is licensed under the Creative Commons-License Attribution 4.0 International (CC BY 4.0) .

This means you're free to copy, share and adapt any parts (or all) of the text in the article, as long as you give appropriate credit and provide a link/reference to this page.

That is it. You don't need our permission to copy the article; just include a link/reference back to this page. You can use it freely (with some kind of link), and we're also okay with people reprinting in publications like books, blogs, newsletters, course-material, papers, wikipedia and presentations (with clear attribution).

milgram experiment ethical considerations

Related articles

Stanley Milgram Experiment

Deception and Research

Informed Consent Policy

Want to stay up to date? Follow us!

Save this course for later.

Don't have time for it all now? No problem, save it as a course and come back to it later.

Footer bottom

  • Privacy Policy

milgram experiment ethical considerations

  • Subscribe to our RSS Feed
  • Like us on Facebook
  • Follow us on Twitter

U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings

The PMC website is updating on October 15, 2024. Learn More or Try it out now .

  • Advanced Search
  • Journal List
  • Scientific Reports

Logo of scirep

A novel experimental approach to study disobedience to authority

Emilie a. caspar.

1 Moral and Social Brain Lab, Department of Experimental Psychology, Ghent University, Henri Dunantlaan, 2, 9000 Ghent, Belgium

2 Center for Research in Cognition and Neuroscience, Université Libre de Bruxelles, Brussels, Belgium

Associated Data

Data are made available on OSF (DOI: https://doi.org/10.17605/OSF.IO/2BKJC ).

Fifty years after the experiments of Stanley Milgram, the main objective of the present paper is to offer a paradigm that complies with up-to-date ethical standards and that can be adapted to various scientific disciplines, ranging from sociology and (social) psychology to neuroscience. Inspired by subsequent versions of Milgram-like paradigms and by combining the strengths of each, this paper presents a novel experimental approach to the study of (dis)obedience to authority. Volunteers are recruited in pairs and take turns to be ‘agents’ or ‘victims’, making the procedure fully reciprocal. For each trial, the agents receive an order from the experimenter to send a real, mildly painful electric shock to the ‘victim’, thus placing participants in an ecological set-up and avoiding the use of cover stories. Depending on the experimental condition, ‘agents’ receive, or do not receive, a monetary gain and are given, or are not given, an aim to obey the experimenter’s orders. Disobedience here refers to the number of times ‘agents’ refused to deliver the real shock to the ‘victim’. As the paradigm is designed to fit with brain imaging methods, I hope to bring new insights and perspectives in this area of research.

Introduction

The experiment of Stanley Milgram is one of the most (in)famous in psychology 1 , within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. Milgram’s classical studies famously suggested a widespread willingness to obey authority, to the point of inflicting irreversible harm to another person just met a few minutes before. Beyond the studies of Milgram, the history of nations is also plagued by horrendous acts of obedience that have caused wars and the loss of countless lives 2 . History has fortunately shown that some individuals do resist the social constraint of receiving orders when their own morality is of greater importance than the social costs associated with defying orders (e.g., 3 , 4 ). To understand the factors that prevent an individual from complying with immoral orders, research on disobedience should focus on two main axes: (1) what social and situational factors support disobedience and (2) what individual differences support disobedience.

The first axe has already been largely investigated in past studies. From Milgram’s studies, important situational factors supporting disobedience have already been established 5 . For instance, disobedience increases if the experimenter is not physically present in the room or if two experimenters provide opposing views regarding the morality of the experiment. Subsequent versions and interpretations of Milgram’s studies 6 – 8 as well as historical research 4 , 9 also suggested the importance of several social (e.g. presence of a supporting group) and situational factors (e.g. family history, proximity with the ‘victim’, intensity of the pain; money) supporting resistance to immoral orders. However, the second axe regarding individual differences has been less systematically approached. A few studies 10 , 11 previously explored personality traits that may influence disobedience (e.g. empathic concern, risk-taking) but most of these studies, however, have used relatively weak and potentially biased methods, such as self-reported questionnaires and methods based on cover stories. These studies are not sufficient to explain why, in a given situation, some people will refuse immoral orders and rescue threatened human beings while others will comply with such orders. With the current literature on disobedience, we have no idea about which neuro-cognitive processes drive inter-individual differences regarding the degree of disobedience. This aim could be achieved by offering a novel experimental approach that would make it possible to use novel techniques that give us a more direct access to the functioning of the brain and cognition, such as functional near-infrared spectroscopy (fNIRS), electroencephalography or Magnetic Resonance Imagery (MRI). Regrettably, the original paradigm and those bearing close similarity are not adapted to reliably answer those questions as they were not designed to fit with neuroimaging measurements. By combining the strengths of previous work on disobedience into a single experimental paradigm and adapting it to fit with cognitive and brain imaging measurements, this novel experimental approach could help to better understand, together with individual, social, and cultural factors, which mechanisms make it possible for an individual to refuse to comply with immoral orders.

There were several challenges to consider in order to develop such a paradigm, both ethical and methodological. Studying obedience and resistance to immoral orders involves putting volunteers in a situation where they have to make a decision on whether or not to commit ‘immoral acts’ under orders. A balance has to be found between what is acceptable from an ethical perspective and what is necessary for the research question. Milgram’s studies on obedience raised undeniable ethical issues 12 – 14 , mostly associated with high stress and the use a cover story, which involves deception. Some variants of Milgram’s studies were realized with immersive virtual reality to prevent the ethical issues associated with Milgram’s paradigm 15 , but the transparency of the fake scenario presented to participants does not capture decision-making in an ecological set-up. Other Milgram-based variants, such as the 150-V method, appear to replicate Milgram’s results 16 with respect to the actual ethical standards, but methodological concerns are still present 17 as cover stories are still used, which lead to interpretation issues. Beyond ethical considerations, the use of deception also indeed involves a doubt about whether or not volunteers truly believed the cover story. As a consequence, a reasonable doubt remains on how to interpret the results and this is one of the main critics associated with Milgram’s studies and following versions. Recent work on the reports of Milgram’s volunteers suggested that there are no strong and reliable evidence that participants believed in the cover story 8 , 14 , 18 . Others suggested that since the stress of participants was visible on video recordings during the experiment (e.g. hand shaking, nervousness), this suggests that participants actually believed that they were torturing another human being 19 . However, this interpretation has been challenged by another study showing that participants can have physiological reactions to stress even in an obviously-fake experimental set-up 15 . These contrasting interpretations of Milgram’s studies actually reinforce the idea that results can hardly be interpreted when cover stories are used 20 . To answer those criticisms, a real scenario had thus to be created, where participants made decisions that have real consequences on another human being.

An additional challenge is that methods relying on the original paradigm of Milgram, such as the virtual reality version 15 or the 150-V method 16 are not adapted to neuroimaging measurements. More specifically, with such Milgram-like experimental approaches, only a single trial would be recorded for the entire experimental session, that is, when the volunteer stops the experiment (if this happens). For cognitive and neuroimaging data collection, a single trial per participant is not a reliable result, which requires the averaging of several trials to obtain a good signal-to-noise ratio.

Another challenge at the methodological and conceptual levels it that several experimenters 1 , 5 , 21 , 22 including myself 21 – 27 , noted that volunteers are extremely obedient when coming to an experiment. Personally, I have tested about 800 volunteers to investigate the mechanisms by which coercive instructions influence individual cognition and moral behaviors. For instance, by using behavioral, electrophysiological and neuroimaging methods, we have observed that when people obey orders to send real shocks to someone else, their sense of agency 23 , their feeling of responsibility 28 , empathy for the pain of the victim and interpersonal guilt 26 are attenuated compared to a situation where they are free to decide which action to execute. Out of 800 volunteers tested, only 27 disobeyed my orders (i.e. 3.3%): 21 for prosocial reasons (i.e. they refused to administer an electric shock to another individual), 3 by contradiction (i.e. by systematically pressing the other button, not matter the content of the order), and 3 for antisocial reasons (i.e. by administering shocks despite my order not to do so). Although convenient to study how obedience affects cognition, this rate is indubitably an issue when studying disobedience. If participants almost never disobey, we can’t study the mechanisms through which resistance to immoral orders may develop in a given situation. Several reasons for not disobeying the experimenter’s orders have been suggested. Some consider that being obedient is part of the human nature as massive and destructive obedience has been observed through countless historical events 2 . Another current view on the experiments of Milgram is that volunteers were actually happy to participate and to contribute to the acquisition of scientific data 17 , thus explaining the high obedience rate observed. This effect has been referred to as ‘engaged followership’ 29 . If that interpretation is correct, the volunteer’s willingness to come and help the experimenter acquiring scientific data creates an extra difficulty to obtain disobedience in an experimental setup. However, this interpretation is challenged by several studies reported by Milgram, which displayed a higher disobedience rate than his original study. For instance disobedience increases when the shocks’ receiver sits in the same room as the participant or when the authoritative experimenter is not physically present in the room 5 . If participants were indeed only guided by their willingness to help to acquire scientific data, this should be the case in any experimental set-up. As some studies involve a higher disobedience rate compared to the initial version of Milgram’s study 1 , they could thus, at a first glance, be used for studying disobedience. However, even if some versions of the initial study of Milgram offer a highly disobedience rate, thus making it possible to study the mechanisms through which resistance to immoral orders may develop in a given situation, these experimental set-ups are still not adapted for cognitive and neuroimaging measurements and still rely on the use of a cover story.

Taking all the presented challenges into account (i.e. not using cover stories to avoid interpretation issues; obtaining a fair rate of disobedience; using an experimental approach that also fits with cognitive and neuroimaging measurements; respecting ethical standards), the present paper presents a set of experiments that combine the strengths of past experimental work on (dis)obedience. Volunteers were openly involved and active (= real social situation) rather than having to act in fictitious scenarios (= imagined social situation, e.g. Slater et al., 2006). They were confronted with moral decisions to follow or not the orders from an experimenter to inflict a real painful shock to a ‘victim’ in exchange (or not) for a small monetary gain, thus avoiding the use of cover stories. Since the aim here is to develop a paradigm that could be used both in behavioral and neuroimaging studies, some basic characteristics had to be considered. For instance, to fit with a Magnetic Resonance Imagery (MRI) scanning environment, neither the ‘victim’, nor the experimenter were in the same room as the agent. A real-time video was thus used to display a video of the victim’s hand receiving shocks on the agent’s screen and headphones were used so the participant could hear the experimenter’s orders.

Another method to study disobedience would be to select participants who are more likely to disobey than others. Each volunteer was thus also asked to complete a series of personality questionnaires to evaluate if a specific profile is associated with a greater prosocial disobedience rate. Systematic post-experimental interviews were conducted at the end of each experiment in order to understand the decisions of volunteers to follow or not the orders of the experimenter and to ask them how they felt during the experiment.

Participants

A hundred eighty naive volunteers (94 females) were recruited in same gender dyads (= 90 dyads). During the recruitment procedure, I ensured that the participants in each dyad were neither close friends (by mixing people studying different academic courses), nor relatives. To estimate the sample size a priori, I calculated the total sample size based on an effect size f of (0.3). To achieve a power of 0.85 for this effect size, the estimated sample size was 168 for 6 groups 30 . I increased the sample size slightly to 180 in order to prevent loss of data in case of withdrawals. Volunteers were randomly assigned to one of the 6 variants of the task (N = 30/variant). One volunteer was not taken into account because they only played the role of the ‘victim’ to replace a participant who did not show up. No volunteers withdrew from the experiment. For the remaining 179 volunteers, the mean age was 22.63 years old (SD = 2.77, range:18–35). A Univariate ANOVA with Age as the dependent variable and Variant as the fixed factor confirmed that age of the volunteers did not differ between the different variant of the tasks ( p  > 0.1, BF 10  = 0.167). Volunteers received between €10 and €19.60 for their participation. All volunteers provided written informed consent prior to the experiment. The study was approved by the Ethics Committee of the Erasme Hospital (reference number: P2019/484). All methods were performed in accordance with the relevant guidelines and regulations.

Method and Material

Six experimental set-ups were created in a between-subject design. In all six set-ups, volunteers were invited by pairs. One person was assigned to start as agent and the other one to start as ‘victim’. Their roles were switched mid-way, ensuring reciprocity. Compared to the experimental design of Milgram, both volunteers were real participants, not confederates. The reciprocity also avoided volunteers to be stuck in the role of the person providing pain to the other, thus attenuating the potential psychological distress of being in a perpetrator role only. Volunteers were given the possibility to choose the role they wanted to start with. In the case none of them had a preference, role assignment was decided by a coin flip, but volunteers were reminded that they could still decide themselves. This procedure allows to ensure that participants do not think that this procedure is a trick.

Volunteers were first given the instructions of the task. Then, they signed the consent forms in front of each other, so both were aware of the other’s consent. The experimenter was never present in the same room, but rather gave the instructions through headphones. This was for two reasons. First, Milgram’s studies show that disobedience increases if the experimenter is not physically present in the room. Second, in the case of MRI scanning, the experimenter would not be able to give direct verbal instructions to the volunteers in the MRI room due to the high noise of the scanner. Here, agents were isolated in a room and were provided headphones to hear the experimenter’s instructions (see Fig.  1 ). They were told that this was done to avoid attentional interferences through the experimenter’s physical presence in the room. In this series of studies, instructions were pre-recorded but a real setup with a microphone connected to the headphones could also work. Pre-recordings allow perfect timing of the events, important for neuroimaging or electroencephalography recordings. The instructions were “ give a shock ” or “ don’t give a shock ”. To increase the authenticity of the procedure, each sentence was recorded 6 times with small variations in the voice and displayed randomly. In addition, the audio recordings included a background sound similar to interphone communications.

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig1_HTML.jpg

Experimental setup. Schematic representation of the experimental setup. Volunteers were in different rooms. The experimenter was located in a third, separated room. The agent heard on a trial basis the orders of the experiment through headphones and had to decide to press the ‘SHOCK’ or ‘NO SHOCK’ button. A real-time camera feedback displayed the hand of the victim of the agent’s screen so to allow to keep track on the consequences of their actions.

Shocks were delivered using a constant current stimulator (Digitimer DS7A) connected to two electrodes placed on the back of victims’ left hand, visible to the agent through the camera display. Individual pain thresholds were determined for the two volunteers before starting the experiment. This threshold was determined by increasing stimulation in steps of 1 mA (Caspar et al., 2016). I approximated an appropriate threshold by asking a series of questions about their pain perception during the calibration (1. «  Is it uncomfortable?  »—2. «  Is it painful?  »—3. «  Could you cope with a maximum of 100 of these shocks?  »—4. «  Could I increase the threshold?  »). When roles were reversed, I briefly re-calibrated the pain threshold of the new victim by increasing the stimulation again from 0 in steps of 3 mA up to the previously determined threshold, to confirm that the initial estimate was still appropriate, and to allow re-familiarisation. The mean stimulation level selected by this procedure was 36.3 mA (SD = 17.5, V = 300, pulse duration: 200 µs). I chose this instead of other types of pain (e.g. financial) because it produces a clear muscle twitch on the victim’s hand each time a shock is sent. This allows volunteers to have a clear and visible feedback of the consequences of their actions and to be fully aware that shocks were real.

There was a total of 96 trials per experimental condition. In the coerced condition, the experimenter asked to give a shock in 64 trials and asked not to give a shock 32 trials. This ratio was chosen on the assumption that the volunteer’s willingness to refuse immoral orders would increase with the number of times they were instructed to inflict pain to the “victim”.

On each trial, a picture of two rectangles, a red one labelled ‘SHOCK’ and a green one labelled ‘NO SHOCK’, was displayed in the bottom left and right of the screen. The key-outcome mapping varied randomly on a trial-wise basis, but the outcome was always fully congruent with the mapping seen by the participant. Agents could then press one of the two buttons. Pressing the SHOCK key delivered a shock to the victim while pressing the NO SHOCK key did not deliver any shocks. This procedure of randomized button mapping allows to have a better control over motor preparation, an aspect that can be important for neuroimaging data.

In half of the variants of the task (i.e., 3/6), the “Aim” variants, participants were given a reason for obeying the orders of the experimenter, while this was not the case in the other half, the “No aim” variants. In the “No Aim” variants, I did not provide any reasons for obeying to the participants and I simply explained the task. If participants asked about the aim, I simply told them that they would know at the end of the experiment, without providing further justifications. In the “Aim” variants, volunteers were told that researchers observed a specific brain activity in the motor cortex in another study when participants were given instructions. We explained that the present study was a control study to measure different aspects linked to motor activity when they press buttons, in order to see if the button pressing was related to brain activity measured over the motor cortex. To increase the veracity of the aim, electrodes were also placed on their fingers and connected to a real electromyography (EMG) apparatus to supposedly record their muscle activity. Volunteers were instructed to press the two buttons only with their right and left index fingers, as naturally as possible, and to avoid producing too ample movements to create clean EMG data. In the case volunteers asked if they really had to follow orders, I told them that for ethical reasons I could not force them to do anything, but that it would be better for the sake of the experiment. Telling them explicitly that they could disobey the orders would not be beneficial in the quest of studying ‘real’ disobedience.

In 4 out of 6 variants of the task, the “Free-choice” variants, a second experimental condition was used, the free-choice condition. In this condition, volunteers were told that they could freely decide in each trial to shock the ‘victim’ or not. In this condition, they did not receive instructions. In 4 out of 6 variants of the task, the “Monetary reward” variants, agents received a monetary reward of + €0.05 for each shock delivered. In the other 2 variants, volunteers were not rewarded for each shock delivered (i.e. “No monetary reward” variants). To resume, the 6 variants of the same task were the following: (1) No Aim + Monetary reward + Free-choice condition; (2) No Aim + No monetary reward + Free-choice condition; (3) Aim + Monetary reward + Free-choice condition; (4) Aim + No monetary reward + Free-choice condition; (5) No Aim + Monetary reward + No free-choice condition; (6) Aim + Monetary reward + No free-choice condition (see Table ​ Table1 1 ).

Schematic representation of each variant of the experimental task.

Variants of the taskAim for obedienceMonetary rewardFree-choice condition
Variant 1
Variant 2
Variant 3
Variant 4
Variant 5
Variant 6

Before the experimental session, volunteers filled in six questionnaires. Those questionnaires included (1) the Money Attitude Scale (e.g. “ I put money aside on a regular basis for the future ”) 31 , (2) the Moral Foundation Questionnaire (e.g. “ Whether or not someone showed a lack of respect for authority ”) 32 , (3) the Aggression-Submission-Conventionalism scale (e.g., “ We should believe what our leaders tell us ”) 33 , (4) the short dark triad scale (e.g., “ Most people can be manipulated ”) 34 , the Interpersonal Reactivity Index (e.g. “ When I see someone get hurt, I tend to remain calm ”) 35 . At the end of the experimental session, they were asked to fill in two more questionnaires: (1) A debriefing assessing what they felt during the experiment and the reasons for choosing to obey or disobey the orders of the experimenter (Supplementary Information S1) and (2) a questionnaire on social identification with the experimenter (e.g., “ I feel strong ties with this experimenter ”) 36 . At the end of the experiment a debriefing was conducted for each volunteer, separately. Volunteers were then paid, again separately.

General data analyses

Each result was analyzed with both frequentist and Bayesian statistics 37 . Bayesian statistics assess the likelihood of the data under both the null and the alternative hypothesis. BF 10 corresponds to the p (data| H 1 )/ p (data| H 0 ). Generally, a BF between 1/3 and 3 indicates that the data is similarly likely under the H 1 and H 0 , and that the data does not adjudicate which is more likely. A BF 10 below 1/3 or above 3 is interpreted as supporting H 0 and H 1 , respectively. For instance, BF 10  = 20 would mean that the data are 20 times more likely under H 1 than H 0 providing very strong support for H 1 , while BF 10  = 0.05 would mean that the data are 20 times more likely under H 0 than H 1 providing very strong support for H 0 38 . BF and p values were calculated using JASP 39 and the default priors implemented in JASP. All analyses were two-tailed.

Number of shocks given in the free-choice condition

In the free-choice condition, volunteers were told that they were entirely free to decide to deliver a shock or not to the ‘victim’ on each of the 96 free-choice trials. On average, agents administered shocks to the victim on 31.86% of the trials (SD = 34.98, minimum: 0%, maximum: 100%) in the free-choice condition, corresponding to 30.59/96 shocks. A paired-sample t-test indicated that agents delivered less frequently a shock in the free-choice condition than in the coerced condition (68.03%, SD = 41.11, t (119)  = -9.919, p  < 0.001, Cohen’s d = − 0.906, BF 10  = 1.987e + 14). This result supports the fact that individuals can inflict more harm to others when they obey orders than when they act freely.

Prosocial disobedience across variants

In the present study, I was interested in prosocial disobedience, that is, when agents refuse the orders of the experimenter to send a painful shock to the ‘victim’. Table ​ Table2 2 displays the number of volunteers who reported that they voluntarily disobeyed in each variant of the task.

Number of volunteers who reported that they voluntarily disobeyed the orders of the experimenter.

Variant 1Variant 2Variant 3Variant 4Variant 5Variant 6
Voluntary disobedience (‘Yes’)23/3024/308/3016/3024/3013/30

In this experiment, the main variable of interest was not to consider how many participants disobeyed in each variant only, but also how frequently they disobeyed. A percentage of prosocial disobedience was calculated for each volunteer, corresponding to the number of trials in which participants chose to disobey (i.e., sending no shocks while ordered by the experimenter to do so) divided by the total number of trials corresponding to the order to send a shock, multiplied by 100. I compared the prosocial disobedience rate across variants of the task, gender of participants and order of the role. I conducted a univariate ANOVA with prosocial disobedience as the dependent variable and Aim (aim given, no aim given), Monetary reward (+ €0.05 or not), Free-choice (presence or absence of a free-choice condition), Gender and Order of the Role (agent first, victim first) as fixed factors (see Fig.  2 ). Both frequentist and Bayesian statistics strongly supported a main effect of Aim (F (1,155)  = 14.248, p  < 0.001, η 2 partial  = 0.084, BF incl  = 158.806). Prosocial disobedience was lower when an aim for obedience was given to volunteers (20.4%, CI 95  = 12.8–28.1) than when no aim was given (43.3%, CI 95  = 35.6–51). Both frequentist and Bayesian statistics also supported a main effect of Monetary reward (F (1,155)  = 12.335, p  = 0.001, η 2 partial  = 0.074, BF incl  = 28.930). Prosocial disobedience was lower when a monetary reward was given for each shock (25.1%, CI 95  = 18.5–31.7) than when no monetary reward was given (45.4%, CI 95  = 35.9–54.8). The frequentist approach showed a main effect of Gender (F (1,155)  = 5.128, p  = 0.025, η 2 partial  = 0.032), with a lower prosocial disobedience rate for female volunteers (25.7%, CI 95  = 18.2–33.2) then for male volunteers (38%, CI 95  = 30–46). However, the Bayesian version of the same analysis revealed a lack of sensitivity (BF incl  = 0.871). All other main effects or interactions supported H 0 or a lack of sensitivity (all p s > 0.1 & BFs incl  ≥ 0.4.291E-7 & ≤ 1.178).

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig2_HTML.jpg

Graphical representation of the percentages of prosocial disobedience in each variant of the task.

The following results report two-tailed Pearson correlations between prosocial disobedience and several other variables, including (1) the reasons given for disobeying, (2) the feeling of responsibility, badness and how sorry they experienced during the experiment, (3) the identification with the experimenter, (4) the perceived level of pain of the victim, (5) identification with the ‘victim’, and (6) individual differences measured through self-report questionnaires. I applied a False Discovery Rate (FDR) approach with the Benjamini and Hochberg method 40 to each p-value for each of those correlations but for the sake of clarity these variables are reported in different sub-sections.

Reasons for prosocial disobedience

All participants who reported that they voluntarily disobeyed the orders of the experimenter (N = 108) were presented a list of 10 reasons that they had to rate from “Not at all” to “Extremely” (see Supplementary Information S1). The reason ‘ I wanted to make more money ’ was only considered for the data of volunteers who had a variant with a monetary reward for each shock (N = 68). Both frequentist and Bayesian statistics showed that the percentage of prosocial disobedience positively correlated with moral reasons (r = 0.550, p FDR  < 0.001, BF 10  = 1.700e + 7), positively correlated with disobedience by contradiction (r = 0.329, p FDR  < 0.001, BF 10  = 47.53) and negatively correlated with the willingness to make more money (r = − 0.485, p FDR  < 0.001, BF 10  = 822.16). Other correlations were in favor of H 0 or were inconclusive (all p s FDR  > 0.076, all BFs 10  ≥ 0.120 & ≤ 1.446).

Feeling responsible, bad and sorry

Both frequentist and Bayesian statistics showed strong positive correlations between prosocial disobedience and how responsible (r = 0.299, p FDR  < 0.001, BF 10  = 343.98) and how bad (r = 0.301, p FDR  < 0.001, BF 10  = 384.65) they felt during the task (see Figs.  3 A and B). The more responsible and worse they felt during the task, the more they refused the order to send a shock to the ‘victim’. How sorry they felt was inconclusive ( p FDR  > 0.08, BF 10  = 0.929).

An external file that holds a picture, illustration, etc.
Object name is 41598_2021_2334_Fig3_HTML.jpg

Graphical representation of Pearson correlations between prosocial disobedience and ( A ) feeling of responsibility, ( B ) how bad agents felt during the task when they administered shocks to the ‘victim’, and ( C ) how painful they estimated the shock delivered to the ‘victim’ was. All tests were two-tailed.

Identification with the experimenter

Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and personal identification ( p FDR  > 0.5, BF 10  = 0.121) and bonding with the experimenter ( p FDR  > 0.5, BF 10  = 0.117). The relationship between the charisma of the experimenter and prosocial disobedience was also slightly in favor of H 0 ( p FDR  > 0.1, BF 10  = 0.530).

Estimated pain of the ‘victim’

The frequentist approach showed a positive correlation between the perceived pain of the ‘victim’ and prosocial disobedience (r = 0.189, p FDR  = 0.048). The higher they considered the ‘victim’ to be in pain, the more frequently they refused to deliver the shock. The Bayesian version of the same analysis slightly supported this relationship (BF 10  = 2.236), see Fig.  3 C.

Identification with the ‘victim’

In the post-session questionnaire, volunteers had to identify to what extent they considered that the other participant could be part of their group and to what extent they identified with the other participant. Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and the perception that the other participant could be part of one’s own group ( p FDR  > 0.8, BF 10  = 0.096). The relationship between prosocial disobedience and the identification with the other participant also slightly supported H 0 ( p FDR  > 0.1, BF 10  = 0.511).

Correlations between the behavior of pairs of participants

As we used a role reversal procedure, the behavior of those who were agents first could influence the behavior of those who turned agents afterwards. A Pearson correlation between prosocial disobedience of agents first and prosocial disobedience of victims who turned agents afterwards. The correlation was positive (r = 0.514, p  < 0.001, BF 10  = 60,068.704), suggesting participants who were agents second tend to act similarly as those who were agents first.

Individual differences associated with prosocial disobedience

Another approach to ensure a reliable prosocial disobedience rate when recruiting volunteers would be to target individuals with a profile that is most frequently associated with disobedient behaviors. Both frequentist and Bayesian statistics for exploratory correlations were two-tailed. Cronbach’s α for each subscale is presented in Supplementary Information S2. Both frequentist and Bayesian statistics showed a negative correlation between scores on the Authority subscale (r = -0.259, p FDR  < 0.001, BF 10  = 41.372) and the Purity subscale (r = -0.303, p FDR  < 0.001, BF 10  = 424.97) from the MFQ questionnaire. The lower volunteers scored on authority and purity, the higher was their prosocial disobedience rate. Other correlations were in favor of H 0 or were inconclusive (all p s FDR  ≥ 0.048, all BFs 10  ≥ 0.100 & ≤ 2.314).

Reasons for obedience

If participants reported that they did not voluntarily disobey the orders of the experimenter, they were asked in an open question to explain their decision to comply with those orders. After reading all the answers, three categories were extracted from the reasons provided: (1) ‘For science’ reasons; participants reported that they obeyed to allow reliable data acquisition (e.g., Participant 91: “ Pour ne pas fausser l’étude ”—English translation: “ To avoid biasing the stud y”); (2) ‘For respect of authority’ reasons; participants reported that they had to follow the orders of the authoritative figure (e.g., Participant 13: “ Pour moi c’est normal de suivre un ordre ”—English translation: “ In my opinion, it’s normal to follow an order ”), and (3) ‘For lack of side-effects’ reasons; participants reported that since the shocks delivered were calibrated on one’s own pain threshold, obeying orders to shock was not problematic (e.g., Participant 115: “ Douleur supportable pour l'autre, je n'ai accepté de faire subir que ce que j'aurais été prêt à subir moi-même ”—English translation: “ The pain was tolerable for the other participant, I have accepted to inflict the intensity of the pain that I would have been ready to undergo myself ”). An independent, naive judge classified the response of participants in one or several of those three established categories. Analyses of the frequencies revealed that the reason “For Science” was mentioned 31/70 times, the reason “For lack of side-effects” was mentioned 17/70 times and the reason “For respect of authority” was mentioned 31/70 times.

The aim of the present paper was to present a novel experimental approach to study (dis)obedience to immoral orders, by combining the strength of past experimental work and by adapting it to cognitive and neuroimaging measurements. Although other versions were proposed since Milgram’s studies, like a study in an immersive virtual environment 15 or the 150-V method 16 , some methodological concerns remained as those methods still involved cover stories or fake experimental set-ups. Here, the experimental approach was significantly different as it was based on an entirely transparent method that involved the administration of real electric shocks to another individual. This approach has the advantage to solve some of the main ethical and methodological concerns associated with the use of cover stories. It also has the advantage that it be can used both to study how social and situational factors influence disobedience as well as individual factors. For social and situational factors, the proposed paradigm can be adapted to evaluate for instance the influence of a supporting group, the use of high or low monetary rewards or how priming disobedience with a documentary influence disobedience. For individual factors, the paradigm allows to investigate how personality traits influence disobedience or to study the neuro-cognitive processes underlying disobedience.

Some novel theories combining a multi-method approach based on social psychology, neuroeconomics and neuroscience could thus emerge to understand better the mechanisms supporting disobedience. For instance, one could evaluate how empathy for the pain of the victim predicts disobedience and how the presence of a supporting group influences our capacity to feel empathy 41 and/or compassion for the ‘victim’ 42 . It could also be argued that the presence of a supporting group diffuses responsibility between individuals and increases obedience, by influencing how our brain processes agency and responsibility over our actions 28 , 43 – 45 . As the results obtained in the present study also indicated that feeling bad for the shocks delivered was statistically associated with prosocial disobedience, one could evaluate how the neural correlates of guilt 46 predicts prosocial disobedience and what historical, cultural and individual factors influence the feeling of guilt.

Six variants of the same task were tested in the present study, some inducing a higher prosocial disobedience rate than others. Statistical results showed that providing a reason—or aim—to justify obedience strongly decreased disobedience. Providing a monetary reward, even one as small as €0.05, also strongly decreased disobedience. Variant 2, in which volunteers were not given an aim or monetary reward, showed the highest disobedience rates. However, to study disobedience in ecological way, the paradigm should capture disobedience of participants even if they know that they are losing something (i.e., monetary rewards or the ‘trust’ of the experimenter asking them help for the study). Defying the orders of an authority generally involves social and/or monetary costs in real-life situations. I would thus not recommend using an experimental paradigm in which volunteers have no costs associated with defying the orders of the experimenter, as it would reduce the ecology of the disobedience act. Variants 3 and 6 involve two types of costs for resisting the orders of the experimenter: a monetary loss and deceiving the experimenter. In Variant 3, descriptive statistics showed that prosocial disobedience was lower compared to Variant 6. The main difference between these two variants was the presence of a free-choice condition. In my former studies 23 , 27 , volunteers frequently justified obedience in the coerced condition because they were given freedom in the free-choice condition (e.g. Participant 89 – English Translation: “ (…) In addition, I knew I could chose freely in the other condition not to send shocks—what I did ). In the present debriefings, some volunteers also reported that the presence of a free-choice condition was giving them enough freedom to accept to follow the orders in the coerced condition. In the supplementary analyses, results showed that when the monetary reward and the aim for obeying are identical, being given a free-choice condition reduces disobedience in the coerced condition. Therefore, Variant 6 appears to provide a good balance between reaching a reliable disobedience rate and finding volunteers who would refuse to produce physical harm on another human beings despite the monetary or social costs associated with defying orders.

Another approach would be to pre-select people who are predicted to be more disobedient. Personality questionnaires indicated that scoring low on the authority and on the purity subscale of the MFQ was strongly associated with a higher prosocial disobedience rate. The link between one’s own relationship to authority and prosocial disobedience observed here replicates another study conducted on the first generation of Rwandese after the 1994 genocide 47 . One’s own relationship to authority thus appears to be a reliable predictor variable in order to pre-select a sample that is more likely to disobey immoral orders.

In the present paper, administering a real mildly painful shock in exchange or not for a small monetary gain was described as an ‘immoral’ act. The notion of what is moral or not can highly differ between individuals 48 , for both academics and volunteers participating in an experiment. Humans are indeed sensitive to different competing issues of morality, a key reason for rescuing persecuted people 49 . In accordance with this observation, the present results indicated that moral reasons were a critical factor associated with the prosocial disobedience rate: the more shocking partners was considered as immoral, the more volunteers disobeyed. However, considering an action as against one’s own moral values does not necessarily translate to a refusal—especially when this order is in line with the Law. An extreme example is soldiers who have perpetrated acts that transgressed their moral beliefs but were issued by their superior in combat 50 . A core question for future research remains: Why are some people capable of putting their own moral standards above the social costs associated with defying orders?

Results indicate that the more volunteers felt responsible during the task, and the worse they felt for sending shocks to the ‘victim’, the higher was their prosocial disobedience. In another study, we observed that obeying orders reduced the feeling of responsibility, how bad and how sorry volunteers felt compared to being free to decide 26 . One hypothesis is that individuals who have preserved a feeling of responsibility and feeling bad—even under command—could more easily defy immoral orders. However, future studies are necessary to better understand the neuro-cognitive processes that prevent an individual from complying with immoral orders. As this paradigm is adapted to neuroimaging measurements, a whole range of studies could now be conducted.

It has been previously suggested that a strong identification with the experimenter giving orders is associated with higher obedience 36 . However, in the present paper, correlations between prosocial disobedience and identification with the experimenter were in favor of H0 with both the frequentist and the Bayesian approaches. In a former study, we also observed that identification to the experimenter was not a critical aspect for explain (dis)obedience. We observed that the generation of Rwandese born after the genocide and tested in Rwanda reported a higher identification to the experimenter than the same generation of Rwandese but tested in Belgium 47 . However, the latter group had a higher prosocial disobedience rate than the former group. Future studies must thus be conducted to understand how the identification with the person giving orders could influence obedience and its weight compared to other social, cultural and individual variables.

Although some volunteers reported that they felt a bit stressed and anxious during the task when they were in the role of the agent, the overwhelming majority did not report any negative psychological feelings. None of the participants withdrew from the experiment and none reported long-term negative psychological effects.

Nowadays, it has become difficult to find volunteers who do not know Milgram’s studies given the high media coverage, including movies, radio soaps, books, podcast and documentaries. One could expect that knowing Milgram would prevent people to obey. However, for the large majority of volunteers, it appears that this is not the case. In previous studies that I conducted with a relatively similar paradigm, the disobedience rate was drastically low (i.e. 3.3%) even if participants were university students knowing Milgram’s studies. In the present study, almost all the volunteers who participated in the present study knew Milgram and explicitly mentioned him during the oral debriefings or before starting the experiment. Yet for those who disobeyed, almost none reported that the reason for disobedience was that they thought it was the aim of the experiment. Further, there was no statistical relationship between prosocial disobedience and believing that it was the aim of the study. It does not mean that knowing Milgram would not influence at all disobedience. It rather suggests that knowing Milgram is not the main factor influencing one’s decision to obey or not an experimenter. It is also possible that since in this experiment shocks were real and not fake such as in Milgram’s studies, participants considered that this was indeed not a study aiming to replicate Milgram.

As far as I have observed, the main problem associated with knowing Milgram’s studies is that volunteers believe that I also have hidden aims and procedures when they enter the experimental room. Several volunteers reported that they only realized that my explanations for the task were true when they were explicitly offered the choice to decide which role to play first and/or when they started receiving the shocks. This is a general concern in psychological studies: The high use of cover stories can also impact other research, as volunteers start to develop a mistrust in what researchers tell them.

Results indicated that who were agents second tend to act similarly as those who were agents first, by sending a relatively similar amount of shocks. Of note, this is an effect that we also observed in past studies on the effect of obeying orders on cognition 23 , 26 , 43 . Nonetheless, in none of those studies we observed that the order of the role had a statistical influence on the neuro-cognitive processes targeted. However, the influence on role reversal on disobedience and related neuro-cognitive processes has still to be investigated in future studies.

The present paradigm is ecological in the sense that volunteers are facing decisions that have a real, physical impact on another human being. However, at the moment I only have little evidences that this paradigm has ecological validity to reflect obedience in real life situations, especially regarding “destructive disobedience” 17 . Caution is indicated when making inference from laboratory studies to complex social behaviours, such as those observed during genocides 16 . My main evidence at the moment is that the very low rate of prosocial disobedience observed in the first generation of post-genocide Rwandans tested in Rwanda using this paradigm 47 is consistent with the fact that deference to authority had already been emphasized by academics as an important factor in the 1994 genocide 4 , 51 . Individual scores on deference to authority in Caspar et al. 47 was the best predictive factor for prosocial disobedience in that former paradigm, thus suggesting some ecological validity. A promising approach would be to recruit “Righteous Among the Nations”, individuals who really saved lives during genocides. Testing this population with the present paradigm would put the ecological validity of this paradigm to the test.

People’s ability to question and resist immoral orders is a fundamental aspect of individual autonomy and of successful societies. As Howard Zinn famously wrote: “ Historically, the most terrible things—war, genocide, and slavery—have resulted not from disobedience, but from obedience ”. Understanding how individuals differ in the extent to which they comply with orders has undeniably several societal implications. They range from understanding how evolving in highly hierarchical environments — such as the military or prisons—influences moral behaviours, to developing interventions that would help to prevent blind obedience and help to resist calls to violence in vulnerable societies. However, since Milgram’s studies, the topic of disobedience has been mostly studied by social psychologists using adapted versions of the initial paradigm developed by Milgram. I hope that with this novel approach, (dis)obedience research will be given a new boost and will be considered by other scientific disciplines seeking to understand better human behaviours.

Supplementary Information

Acknowledgements.

Emilie A. Caspar was funded by the F.R.S-FNRS.

Author contributions

E.A.C. developed the study concept and the method. Testing, data collection and data analysis were performed by E.A.C. E.A.C. wrote the manuscript.

Data availability

Competing interests.

The author declares no competing interests.

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

The online version contains supplementary material available at 10.1038/s41598-021-02334-8.

What is Psychology?

Answers to Your Psychology Questions

What are the Milgram Experiment Ethical Issues?

The Milgram Experiment was a series of experimental studies that took place in the 1960s to investigate how willing subjects were to obey an authority figure even when their actions directly conflicted with their personal conscience. The experiments proved to be extremely controversial and were considered to be highly unethical at the time, and although they have been replicated many times over the years using more stringent ethical controls, the results have remained fairly consistent. So what are the Milgram Experiment Ethical Issues and what did the results of the experiment show?

The experiments were conducted by Stanley Milgram, a psychologist at Yale University, and were designed to see if ordinary people would blindly follow immoral orders, just like many had done during the Nazi era. As part of the experiment, participants were instructed to teach a “learner” pairs of words and administer an electric shock if they gave an incorrect response. With each incorrect response, the electric shock was amplified, despite the fact the “learner” had already warned the subject he had a heart problem.

In reality there were no electric shocks and the “learner” was an actor, but the participant had no way of knowing this and as far as the teacher was concerned, they really were obeying instructions and administering electric shocks to another individual in a separate room.

Interestingly, despite hearing a recorded soundtrack of screams of pain as well as believing that the subject had a heart condition, 65 percent of the participants from the first experiment continued until the end and administered the potentially fatal 450 volt electric shock, although all of them questioned what they were being told to do at some point in the process.

The conclusions drawn from the Milgram experiment were summarized in an article written in 1974 titled ‘The Perils of Obedience’.

“Ordinary people, simply doing their jobs, and without any particular hostility on their part, can become agents in a terrible destructive process. Moreover, even when the destructive effects of their work become patently clear, and they are asked to carry out actions incompatible with fundamental standards of morality, relatively few people have the resources needed to resist authority.”

What are the Milgram Experiment ethical issues and has the design of the experiment been altered over the years to accommodate current standards for the ethical treatment of participants?

One of the biggest criticisms of the Milgram Experiments was the extreme psychological stress inflicted on the participants. Many were reported to be very uncomfortable about what they were being asked to do, although a large number of the participants in the early experiments were later said to be very grateful to have taken part.

There were also many questions raised about the implications of the experiments. The early experiments were conducted in the shadow of the Adolf Eichmann Nazi war crimes trial and there were many uncomfortable parallels drawn between Milgram’s results and the blind obedience exhibited by thousands of Nazi accomplices during the Holocaust.

The Milgram Experiments have been repeated in different guises numerous times over the years. In all cases, the results have remained largely consistent, even when modern ethical guidelines were followed and participants were given details of the experiment.

Related Articles:

  • What are the Zimbardo Prison Experiment Ethical Issues? The Stanford Prison Experiment was designed to explore the psychological impact of the prison environment on prisoners and prison guards. Although Zimbardo intended the experiment to last two weeks, it was abruptly halted after only six days due to the increasingly disturbing behavioral traits being exhibited by the "prisoners" and...
  • What are the Cognitive Consequences of Forced Compliance? What are the cognitive consequences of forced compliance? Or to put it another way, does a person’s private opinion change if he is forced to say something that contradicts that opinion? This type of behavior was first discussed in the paper Cognitive Consequences of Forced Compliance written by eminent psychologists,...
  • What are the Interpersonal Dynamics in a Simulated Prison? What are the interpersonal dynamics in a simulated prison? The psychological effects of a simulated prison are an interesting area of study and one that was explored in a highly controversial psychological experiment known as the Stanford Prison Experiment. In 1971, Professor Philip Zimbardo came up with the idea of...
  • Stanford Prison Experiment Summary The Stanford Prison Experiment Summary is a famous psychology experiment that was designed to study the psychological impact of becoming a prison guard or prisoner. The experiment was conducted by Professor of Psychology, Philip Zimbardo, at Stanford University in 1971. Although it was originally intended to last for two weeks,...

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

The Milgram Experiment Ethical Issues in 2017

Cortney Taianne Winfield at Purdue University Global

  • Purdue University Global

Discover the world's research

  • 25+ million members
  • 160+ million publication pages
  • 2.3+ billion citations
  • Recruit researchers
  • Join for free
  • Login Email Tip: Most researchers use their institutional email address as their ResearchGate login Password Forgot password? Keep me logged in Log in or Continue with Google Welcome back! Please log in. Email · Hint Tip: Most researchers use their institutional email address as their ResearchGate login Password Forgot password? Keep me logged in Log in or Continue with Google No account? Sign up

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Open access
  • Published: 25 November 2021

A novel experimental approach to study disobedience to authority

  • Emilie A. Caspar 1 , 2  

Scientific Reports volume  11 , Article number:  22927 ( 2021 ) Cite this article

6787 Accesses

6 Citations

16 Altmetric

Metrics details

  • Human behaviour

Fifty years after the experiments of Stanley Milgram, the main objective of the present paper is to offer a paradigm that complies with up-to-date ethical standards and that can be adapted to various scientific disciplines, ranging from sociology and (social) psychology to neuroscience. Inspired by subsequent versions of Milgram-like paradigms and by combining the strengths of each, this paper presents a novel experimental approach to the study of (dis)obedience to authority. Volunteers are recruited in pairs and take turns to be ‘agents’ or ‘victims’, making the procedure fully reciprocal. For each trial, the agents receive an order from the experimenter to send a real, mildly painful electric shock to the ‘victim’, thus placing participants in an ecological set-up and avoiding the use of cover stories. Depending on the experimental condition, ‘agents’ receive, or do not receive, a monetary gain and are given, or are not given, an aim to obey the experimenter’s orders. Disobedience here refers to the number of times ‘agents’ refused to deliver the real shock to the ‘victim’. As the paradigm is designed to fit with brain imaging methods, I hope to bring new insights and perspectives in this area of research.

Similar content being viewed by others

milgram experiment ethical considerations

On the cognitive mechanisms supporting prosocial disobedience in a post-genocidal context

milgram experiment ethical considerations

Neural mechanisms of deception in a social context: an fMRI replication study

milgram experiment ethical considerations

Psychopathic traits mediate guilt-related anterior midcingulate activity under authority pressure

Introduction.

The experiment of Stanley Milgram is one of the most (in)famous in psychology 1 , within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. Milgram’s classical studies famously suggested a widespread willingness to obey authority, to the point of inflicting irreversible harm to another person just met a few minutes before. Beyond the studies of Milgram, the history of nations is also plagued by horrendous acts of obedience that have caused wars and the loss of countless lives 2 . History has fortunately shown that some individuals do resist the social constraint of receiving orders when their own morality is of greater importance than the social costs associated with defying orders (e.g., 3 , 4 ). To understand the factors that prevent an individual from complying with immoral orders, research on disobedience should focus on two main axes: (1) what social and situational factors support disobedience and (2) what individual differences support disobedience.

The first axe has already been largely investigated in past studies. From Milgram’s studies, important situational factors supporting disobedience have already been established 5 . For instance, disobedience increases if the experimenter is not physically present in the room or if two experimenters provide opposing views regarding the morality of the experiment. Subsequent versions and interpretations of Milgram’s studies 6 , 7 , 8 as well as historical research 4 , 9 also suggested the importance of several social (e.g. presence of a supporting group) and situational factors (e.g. family history, proximity with the ‘victim’, intensity of the pain; money) supporting resistance to immoral orders. However, the second axe regarding individual differences has been less systematically approached. A few studies 10 , 11 previously explored personality traits that may influence disobedience (e.g. empathic concern, risk-taking) but most of these studies, however, have used relatively weak and potentially biased methods, such as self-reported questionnaires and methods based on cover stories. These studies are not sufficient to explain why, in a given situation, some people will refuse immoral orders and rescue threatened human beings while others will comply with such orders. With the current literature on disobedience, we have no idea about which neuro-cognitive processes drive inter-individual differences regarding the degree of disobedience. This aim could be achieved by offering a novel experimental approach that would make it possible to use novel techniques that give us a more direct access to the functioning of the brain and cognition, such as functional near-infrared spectroscopy (fNIRS), electroencephalography or Magnetic Resonance Imagery (MRI). Regrettably, the original paradigm and those bearing close similarity are not adapted to reliably answer those questions as they were not designed to fit with neuroimaging measurements. By combining the strengths of previous work on disobedience into a single experimental paradigm and adapting it to fit with cognitive and brain imaging measurements, this novel experimental approach could help to better understand, together with individual, social, and cultural factors, which mechanisms make it possible for an individual to refuse to comply with immoral orders.

There were several challenges to consider in order to develop such a paradigm, both ethical and methodological. Studying obedience and resistance to immoral orders involves putting volunteers in a situation where they have to make a decision on whether or not to commit ‘immoral acts’ under orders. A balance has to be found between what is acceptable from an ethical perspective and what is necessary for the research question. Milgram’s studies on obedience raised undeniable ethical issues 12 , 13 , 14 , mostly associated with high stress and the use a cover story, which involves deception. Some variants of Milgram’s studies were realized with immersive virtual reality to prevent the ethical issues associated with Milgram’s paradigm 15 , but the transparency of the fake scenario presented to participants does not capture decision-making in an ecological set-up. Other Milgram-based variants, such as the 150-V method, appear to replicate Milgram’s results 16 with respect to the actual ethical standards, but methodological concerns are still present 17 as cover stories are still used, which lead to interpretation issues. Beyond ethical considerations, the use of deception also indeed involves a doubt about whether or not volunteers truly believed the cover story. As a consequence, a reasonable doubt remains on how to interpret the results and this is one of the main critics associated with Milgram’s studies and following versions. Recent work on the reports of Milgram’s volunteers suggested that there are no strong and reliable evidence that participants believed in the cover story 8 , 14 , 18 . Others suggested that since the stress of participants was visible on video recordings during the experiment (e.g. hand shaking, nervousness), this suggests that participants actually believed that they were torturing another human being 19 . However, this interpretation has been challenged by another study showing that participants can have physiological reactions to stress even in an obviously-fake experimental set-up 15 . These contrasting interpretations of Milgram’s studies actually reinforce the idea that results can hardly be interpreted when cover stories are used 20 . To answer those criticisms, a real scenario had thus to be created, where participants made decisions that have real consequences on another human being.

An additional challenge is that methods relying on the original paradigm of Milgram, such as the virtual reality version 15 or the 150-V method 16 are not adapted to neuroimaging measurements. More specifically, with such Milgram-like experimental approaches, only a single trial would be recorded for the entire experimental session, that is, when the volunteer stops the experiment (if this happens). For cognitive and neuroimaging data collection, a single trial per participant is not a reliable result, which requires the averaging of several trials to obtain a good signal-to-noise ratio.

Another challenge at the methodological and conceptual levels it that several experimenters 1 , 5 , 21 , 22 including myself 23 , 24 , 25 , 26 , 27 , noted that volunteers are extremely obedient when coming to an experiment. Personally, I have tested about 800 volunteers to investigate the mechanisms by which coercive instructions influence individual cognition and moral behaviors. For instance, by using behavioral, electrophysiological and neuroimaging methods, we have observed that when people obey orders to send real shocks to someone else, their sense of agency 23 , their feeling of responsibility 28 , empathy for the pain of the victim and interpersonal guilt 26 are attenuated compared to a situation where they are free to decide which action to execute. Out of 800 volunteers tested, only 27 disobeyed my orders (i.e. 3.3%): 21 for prosocial reasons (i.e. they refused to administer an electric shock to another individual), 3 by contradiction (i.e. by systematically pressing the other button, not matter the content of the order), and 3 for antisocial reasons (i.e. by administering shocks despite my order not to do so). Although convenient to study how obedience affects cognition, this rate is indubitably an issue when studying disobedience. If participants almost never disobey, we can’t study the mechanisms through which resistance to immoral orders may develop in a given situation. Several reasons for not disobeying the experimenter’s orders have been suggested. Some consider that being obedient is part of the human nature as massive and destructive obedience has been observed through countless historical events 2 . Another current view on the experiments of Milgram is that volunteers were actually happy to participate and to contribute to the acquisition of scientific data 17 , thus explaining the high obedience rate observed. This effect has been referred to as ‘engaged followership’ 29 . If that interpretation is correct, the volunteer’s willingness to come and help the experimenter acquiring scientific data creates an extra difficulty to obtain disobedience in an experimental setup. However, this interpretation is challenged by several studies reported by Milgram, which displayed a higher disobedience rate than his original study. For instance disobedience increases when the shocks’ receiver sits in the same room as the participant or when the authoritative experimenter is not physically present in the room 5 . If participants were indeed only guided by their willingness to help to acquire scientific data, this should be the case in any experimental set-up. As some studies involve a higher disobedience rate compared to the initial version of Milgram’s study 1 , they could thus, at a first glance, be used for studying disobedience. However, even if some versions of the initial study of Milgram offer a highly disobedience rate, thus making it possible to study the mechanisms through which resistance to immoral orders may develop in a given situation, these experimental set-ups are still not adapted for cognitive and neuroimaging measurements and still rely on the use of a cover story.

Taking all the presented challenges into account (i.e. not using cover stories to avoid interpretation issues; obtaining a fair rate of disobedience; using an experimental approach that also fits with cognitive and neuroimaging measurements; respecting ethical standards), the present paper presents a set of experiments that combine the strengths of past experimental work on (dis)obedience. Volunteers were openly involved and active (= real social situation) rather than having to act in fictitious scenarios (= imagined social situation, e.g. Slater et al., 2006). They were confronted with moral decisions to follow or not the orders from an experimenter to inflict a real painful shock to a ‘victim’ in exchange (or not) for a small monetary gain, thus avoiding the use of cover stories. Since the aim here is to develop a paradigm that could be used both in behavioral and neuroimaging studies, some basic characteristics had to be considered. For instance, to fit with a Magnetic Resonance Imagery (MRI) scanning environment, neither the ‘victim’, nor the experimenter were in the same room as the agent. A real-time video was thus used to display a video of the victim’s hand receiving shocks on the agent’s screen and headphones were used so the participant could hear the experimenter’s orders.

Another method to study disobedience would be to select participants who are more likely to disobey than others. Each volunteer was thus also asked to complete a series of personality questionnaires to evaluate if a specific profile is associated with a greater prosocial disobedience rate. Systematic post-experimental interviews were conducted at the end of each experiment in order to understand the decisions of volunteers to follow or not the orders of the experimenter and to ask them how they felt during the experiment.

Participants

A hundred eighty naive volunteers (94 females) were recruited in same gender dyads (= 90 dyads). During the recruitment procedure, I ensured that the participants in each dyad were neither close friends (by mixing people studying different academic courses), nor relatives. To estimate the sample size a priori, I calculated the total sample size based on an effect size f of (0.3). To achieve a power of 0.85 for this effect size, the estimated sample size was 168 for 6 groups 30 . I increased the sample size slightly to 180 in order to prevent loss of data in case of withdrawals. Volunteers were randomly assigned to one of the 6 variants of the task (N = 30/variant). One volunteer was not taken into account because they only played the role of the ‘victim’ to replace a participant who did not show up. No volunteers withdrew from the experiment. For the remaining 179 volunteers, the mean age was 22.63 years old (SD = 2.77, range:18–35). A Univariate ANOVA with Age as the dependent variable and Variant as the fixed factor confirmed that age of the volunteers did not differ between the different variant of the tasks ( p  > 0.1, BF 10  = 0.167). Volunteers received between €10 and €19.60 for their participation. All volunteers provided written informed consent prior to the experiment. The study was approved by the Ethics Committee of the Erasme Hospital (reference number: P2019/484). All methods were performed in accordance with the relevant guidelines and regulations.

Method and Material

Six experimental set-ups were created in a between-subject design. In all six set-ups, volunteers were invited by pairs. One person was assigned to start as agent and the other one to start as ‘victim’. Their roles were switched mid-way, ensuring reciprocity. Compared to the experimental design of Milgram, both volunteers were real participants, not confederates. The reciprocity also avoided volunteers to be stuck in the role of the person providing pain to the other, thus attenuating the potential psychological distress of being in a perpetrator role only. Volunteers were given the possibility to choose the role they wanted to start with. In the case none of them had a preference, role assignment was decided by a coin flip, but volunteers were reminded that they could still decide themselves. This procedure allows to ensure that participants do not think that this procedure is a trick.

Volunteers were first given the instructions of the task. Then, they signed the consent forms in front of each other, so both were aware of the other’s consent. The experimenter was never present in the same room, but rather gave the instructions through headphones. This was for two reasons. First, Milgram’s studies show that disobedience increases if the experimenter is not physically present in the room. Second, in the case of MRI scanning, the experimenter would not be able to give direct verbal instructions to the volunteers in the MRI room due to the high noise of the scanner. Here, agents were isolated in a room and were provided headphones to hear the experimenter’s instructions (see Fig.  1 ). They were told that this was done to avoid attentional interferences through the experimenter’s physical presence in the room. In this series of studies, instructions were pre-recorded but a real setup with a microphone connected to the headphones could also work. Pre-recordings allow perfect timing of the events, important for neuroimaging or electroencephalography recordings. The instructions were “ give a shock ” or “ don’t give a shock ”. To increase the authenticity of the procedure, each sentence was recorded 6 times with small variations in the voice and displayed randomly. In addition, the audio recordings included a background sound similar to interphone communications.

figure 1

Experimental setup. Schematic representation of the experimental setup. Volunteers were in different rooms. The experimenter was located in a third, separated room. The agent heard on a trial basis the orders of the experiment through headphones and had to decide to press the ‘SHOCK’ or ‘NO SHOCK’ button. A real-time camera feedback displayed the hand of the victim of the agent’s screen so to allow to keep track on the consequences of their actions.

Shocks were delivered using a constant current stimulator (Digitimer DS7A) connected to two electrodes placed on the back of victims’ left hand, visible to the agent through the camera display. Individual pain thresholds were determined for the two volunteers before starting the experiment. This threshold was determined by increasing stimulation in steps of 1 mA (Caspar et al., 2016). I approximated an appropriate threshold by asking a series of questions about their pain perception during the calibration (1. «  Is it uncomfortable?  »—2. «  Is it painful?  »—3. «  Could you cope with a maximum of 100 of these shocks?  »—4. «  Could I increase the threshold?  »). When roles were reversed, I briefly re-calibrated the pain threshold of the new victim by increasing the stimulation again from 0 in steps of 3 mA up to the previously determined threshold, to confirm that the initial estimate was still appropriate, and to allow re-familiarisation. The mean stimulation level selected by this procedure was 36.3 mA (SD = 17.5, V = 300, pulse duration: 200 µs). I chose this instead of other types of pain (e.g. financial) because it produces a clear muscle twitch on the victim’s hand each time a shock is sent. This allows volunteers to have a clear and visible feedback of the consequences of their actions and to be fully aware that shocks were real.

There was a total of 96 trials per experimental condition. In the coerced condition, the experimenter asked to give a shock in 64 trials and asked not to give a shock 32 trials. This ratio was chosen on the assumption that the volunteer’s willingness to refuse immoral orders would increase with the number of times they were instructed to inflict pain to the “victim”.

On each trial, a picture of two rectangles, a red one labelled ‘SHOCK’ and a green one labelled ‘NO SHOCK’, was displayed in the bottom left and right of the screen. The key-outcome mapping varied randomly on a trial-wise basis, but the outcome was always fully congruent with the mapping seen by the participant. Agents could then press one of the two buttons. Pressing the SHOCK key delivered a shock to the victim while pressing the NO SHOCK key did not deliver any shocks. This procedure of randomized button mapping allows to have a better control over motor preparation, an aspect that can be important for neuroimaging data.

In half of the variants of the task (i.e., 3/6), the “Aim” variants, participants were given a reason for obeying the orders of the experimenter, while this was not the case in the other half, the “No aim” variants. In the “No Aim” variants, I did not provide any reasons for obeying to the participants and I simply explained the task. If participants asked about the aim, I simply told them that they would know at the end of the experiment, without providing further justifications. In the “Aim” variants, volunteers were told that researchers observed a specific brain activity in the motor cortex in another study when participants were given instructions. We explained that the present study was a control study to measure different aspects linked to motor activity when they press buttons, in order to see if the button pressing was related to brain activity measured over the motor cortex. To increase the veracity of the aim, electrodes were also placed on their fingers and connected to a real electromyography (EMG) apparatus to supposedly record their muscle activity. Volunteers were instructed to press the two buttons only with their right and left index fingers, as naturally as possible, and to avoid producing too ample movements to create clean EMG data. In the case volunteers asked if they really had to follow orders, I told them that for ethical reasons I could not force them to do anything, but that it would be better for the sake of the experiment. Telling them explicitly that they could disobey the orders would not be beneficial in the quest of studying ‘real’ disobedience.

In 4 out of 6 variants of the task, the “Free-choice” variants, a second experimental condition was used, the free-choice condition. In this condition, volunteers were told that they could freely decide in each trial to shock the ‘victim’ or not. In this condition, they did not receive instructions. In 4 out of 6 variants of the task, the “Monetary reward” variants, agents received a monetary reward of + €0.05 for each shock delivered. In the other 2 variants, volunteers were not rewarded for each shock delivered (i.e. “No monetary reward” variants). To resume, the 6 variants of the same task were the following: (1) No Aim + Monetary reward + Free-choice condition; (2) No Aim + No monetary reward + Free-choice condition; (3) Aim + Monetary reward + Free-choice condition; (4) Aim + No monetary reward + Free-choice condition; (5) No Aim + Monetary reward + No free-choice condition; (6) Aim + Monetary reward + No free-choice condition (see Table 1 ).

Before the experimental session, volunteers filled in six questionnaires. Those questionnaires included (1) the Money Attitude Scale (e.g. “ I put money aside on a regular basis for the future ”) 31 , (2) the Moral Foundation Questionnaire (e.g. “ Whether or not someone showed a lack of respect for authority ”) 32 , (3) the Aggression-Submission-Conventionalism scale (e.g., “ We should believe what our leaders tell us ”) 33 , (4) the short dark triad scale (e.g., “ Most people can be manipulated ”) 34 , the Interpersonal Reactivity Index (e.g. “ When I see someone get hurt, I tend to remain calm ”) 35 . At the end of the experimental session, they were asked to fill in two more questionnaires: (1) A debriefing assessing what they felt during the experiment and the reasons for choosing to obey or disobey the orders of the experimenter (Supplementary Information S1) and (2) a questionnaire on social identification with the experimenter (e.g., “ I feel strong ties with this experimenter ”) 36 . At the end of the experiment a debriefing was conducted for each volunteer, separately. Volunteers were then paid, again separately.

General data analyses

Each result was analyzed with both frequentist and Bayesian statistics 37 . Bayesian statistics assess the likelihood of the data under both the null and the alternative hypothesis. BF 10 corresponds to the p (data| H 1 )/ p (data| H 0 ). Generally, a BF between 1/3 and 3 indicates that the data is similarly likely under the H 1 and H 0 , and that the data does not adjudicate which is more likely. A BF 10 below 1/3 or above 3 is interpreted as supporting H 0 and H 1 , respectively. For instance, BF 10  = 20 would mean that the data are 20 times more likely under H 1 than H 0 providing very strong support for H 1 , while BF 10  = 0.05 would mean that the data are 20 times more likely under H 0 than H 1 providing very strong support for H 0 38 . BF and p values were calculated using JASP 39 and the default priors implemented in JASP. All analyses were two-tailed.

Number of shocks given in the free-choice condition

In the free-choice condition, volunteers were told that they were entirely free to decide to deliver a shock or not to the ‘victim’ on each of the 96 free-choice trials. On average, agents administered shocks to the victim on 31.86% of the trials (SD = 34.98, minimum: 0%, maximum: 100%) in the free-choice condition, corresponding to 30.59/96 shocks. A paired-sample t-test indicated that agents delivered less frequently a shock in the free-choice condition than in the coerced condition (68.03%, SD = 41.11, t (119)  = -9.919, p  < 0.001, Cohen’s d = − 0.906, BF 10  = 1.987e + 14). This result supports the fact that individuals can inflict more harm to others when they obey orders than when they act freely.

Prosocial disobedience across variants

In the present study, I was interested in prosocial disobedience, that is, when agents refuse the orders of the experimenter to send a painful shock to the ‘victim’. Table 2 displays the number of volunteers who reported that they voluntarily disobeyed in each variant of the task.

In this experiment, the main variable of interest was not to consider how many participants disobeyed in each variant only, but also how frequently they disobeyed. A percentage of prosocial disobedience was calculated for each volunteer, corresponding to the number of trials in which participants chose to disobey (i.e., sending no shocks while ordered by the experimenter to do so) divided by the total number of trials corresponding to the order to send a shock, multiplied by 100. I compared the prosocial disobedience rate across variants of the task, gender of participants and order of the role. I conducted a univariate ANOVA with prosocial disobedience as the dependent variable and Aim (aim given, no aim given), Monetary reward (+ €0.05 or not), Free-choice (presence or absence of a free-choice condition), Gender and Order of the Role (agent first, victim first) as fixed factors (see Fig.  2 ). Both frequentist and Bayesian statistics strongly supported a main effect of Aim (F (1,155)  = 14.248, p  < 0.001, η 2 partial  = 0.084, BF incl  = 158.806). Prosocial disobedience was lower when an aim for obedience was given to volunteers (20.4%, CI 95  = 12.8–28.1) than when no aim was given (43.3%, CI 95  = 35.6–51). Both frequentist and Bayesian statistics also supported a main effect of Monetary reward (F (1,155)  = 12.335, p  = 0.001, η 2 partial  = 0.074, BF incl  = 28.930). Prosocial disobedience was lower when a monetary reward was given for each shock (25.1%, CI 95  = 18.5–31.7) than when no monetary reward was given (45.4%, CI 95  = 35.9–54.8). The frequentist approach showed a main effect of Gender (F (1,155)  = 5.128, p  = 0.025, η 2 partial  = 0.032), with a lower prosocial disobedience rate for female volunteers (25.7%, CI 95  = 18.2–33.2) then for male volunteers (38%, CI 95  = 30–46). However, the Bayesian version of the same analysis revealed a lack of sensitivity (BF incl  = 0.871). All other main effects or interactions supported H 0 or a lack of sensitivity (all p s > 0.1 & BFs incl  ≥ 0.4.291E-7 & ≤ 1.178).

figure 2

Graphical representation of the percentages of prosocial disobedience in each variant of the task.

The following results report two-tailed Pearson correlations between prosocial disobedience and several other variables, including (1) the reasons given for disobeying, (2) the feeling of responsibility, badness and how sorry they experienced during the experiment, (3) the identification with the experimenter, (4) the perceived level of pain of the victim, (5) identification with the ‘victim’, and (6) individual differences measured through self-report questionnaires. I applied a False Discovery Rate (FDR) approach with the Benjamini and Hochberg method 40 to each p-value for each of those correlations but for the sake of clarity these variables are reported in different sub-sections.

Reasons for prosocial disobedience

All participants who reported that they voluntarily disobeyed the orders of the experimenter (N = 108) were presented a list of 10 reasons that they had to rate from “Not at all” to “Extremely” (see Supplementary Information S1). The reason ‘ I wanted to make more money ’ was only considered for the data of volunteers who had a variant with a monetary reward for each shock (N = 68). Both frequentist and Bayesian statistics showed that the percentage of prosocial disobedience positively correlated with moral reasons (r = 0.550, p FDR  < 0.001, BF 10  = 1.700e + 7), positively correlated with disobedience by contradiction (r = 0.329, p FDR  < 0.001, BF 10  = 47.53) and negatively correlated with the willingness to make more money (r = − 0.485, p FDR  < 0.001, BF 10  = 822.16). Other correlations were in favor of H 0 or were inconclusive (all p s FDR  > 0.076, all BFs 10  ≥ 0.120 & ≤ 1.446).

Feeling responsible, bad and sorry

Both frequentist and Bayesian statistics showed strong positive correlations between prosocial disobedience and how responsible (r = 0.299, p FDR  < 0.001, BF 10  = 343.98) and how bad (r = 0.301, p FDR  < 0.001, BF 10  = 384.65) they felt during the task (see Figs.  3 A and B). The more responsible and worse they felt during the task, the more they refused the order to send a shock to the ‘victim’. How sorry they felt was inconclusive ( p FDR  > 0.08, BF 10  = 0.929).

figure 3

Graphical representation of Pearson correlations between prosocial disobedience and ( A ) feeling of responsibility, ( B ) how bad agents felt during the task when they administered shocks to the ‘victim’, and ( C ) how painful they estimated the shock delivered to the ‘victim’ was. All tests were two-tailed.

Identification with the experimenter

Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and personal identification ( p FDR  > 0.5, BF 10  = 0.121) and bonding with the experimenter ( p FDR  > 0.5, BF 10  = 0.117). The relationship between the charisma of the experimenter and prosocial disobedience was also slightly in favor of H 0 ( p FDR  > 0.1, BF 10  = 0.530).

Estimated pain of the ‘victim’

The frequentist approach showed a positive correlation between the perceived pain of the ‘victim’ and prosocial disobedience (r = 0.189, p FDR  = 0.048). The higher they considered the ‘victim’ to be in pain, the more frequently they refused to deliver the shock. The Bayesian version of the same analysis slightly supported this relationship (BF 10  = 2.236), see Fig.  3 C.

Identification with the ‘victim’

In the post-session questionnaire, volunteers had to identify to what extent they considered that the other participant could be part of their group and to what extent they identified with the other participant. Both frequentist and Bayesian statistics strongly supported H 0 regarding the relationship between prosocial disobedience and the perception that the other participant could be part of one’s own group ( p FDR  > 0.8, BF 10  = 0.096). The relationship between prosocial disobedience and the identification with the other participant also slightly supported H 0 ( p FDR  > 0.1, BF 10  = 0.511).

Correlations between the behavior of pairs of participants

As we used a role reversal procedure, the behavior of those who were agents first could influence the behavior of those who turned agents afterwards. A Pearson correlation between prosocial disobedience of agents first and prosocial disobedience of victims who turned agents afterwards. The correlation was positive (r = 0.514, p  < 0.001, BF 10  = 60,068.704), suggesting participants who were agents second tend to act similarly as those who were agents first.

Individual differences associated with prosocial disobedience

Another approach to ensure a reliable prosocial disobedience rate when recruiting volunteers would be to target individuals with a profile that is most frequently associated with disobedient behaviors. Both frequentist and Bayesian statistics for exploratory correlations were two-tailed. Cronbach’s α for each subscale is presented in Supplementary Information S2. Both frequentist and Bayesian statistics showed a negative correlation between scores on the Authority subscale (r = -0.259, p FDR  < 0.001, BF 10  = 41.372) and the Purity subscale (r = -0.303, p FDR  < 0.001, BF 10  = 424.97) from the MFQ questionnaire. The lower volunteers scored on authority and purity, the higher was their prosocial disobedience rate. Other correlations were in favor of H 0 or were inconclusive (all p s FDR  ≥ 0.048, all BFs 10  ≥ 0.100 & ≤ 2.314).

Reasons for obedience

If participants reported that they did not voluntarily disobey the orders of the experimenter, they were asked in an open question to explain their decision to comply with those orders. After reading all the answers, three categories were extracted from the reasons provided: (1) ‘For science’ reasons; participants reported that they obeyed to allow reliable data acquisition (e.g., Participant 91: “ Pour ne pas fausser l’étude ”—English translation: “ To avoid biasing the stud y”); (2) ‘For respect of authority’ reasons; participants reported that they had to follow the orders of the authoritative figure (e.g., Participant 13: “ Pour moi c’est normal de suivre un ordre ”—English translation: “ In my opinion, it’s normal to follow an order ”), and (3) ‘For lack of side-effects’ reasons; participants reported that since the shocks delivered were calibrated on one’s own pain threshold, obeying orders to shock was not problematic (e.g., Participant 115: “ Douleur supportable pour l'autre, je n'ai accepté de faire subir que ce que j'aurais été prêt à subir moi-même ”—English translation: “ The pain was tolerable for the other participant, I have accepted to inflict the intensity of the pain that I would have been ready to undergo myself ”). An independent, naive judge classified the response of participants in one or several of those three established categories. Analyses of the frequencies revealed that the reason “For Science” was mentioned 31/70 times, the reason “For lack of side-effects” was mentioned 17/70 times and the reason “For respect of authority” was mentioned 31/70 times.

The aim of the present paper was to present a novel experimental approach to study (dis)obedience to immoral orders, by combining the strength of past experimental work and by adapting it to cognitive and neuroimaging measurements. Although other versions were proposed since Milgram’s studies, like a study in an immersive virtual environment 15 or the 150-V method 16 , some methodological concerns remained as those methods still involved cover stories or fake experimental set-ups. Here, the experimental approach was significantly different as it was based on an entirely transparent method that involved the administration of real electric shocks to another individual. This approach has the advantage to solve some of the main ethical and methodological concerns associated with the use of cover stories. It also has the advantage that it be can used both to study how social and situational factors influence disobedience as well as individual factors. For social and situational factors, the proposed paradigm can be adapted to evaluate for instance the influence of a supporting group, the use of high or low monetary rewards or how priming disobedience with a documentary influence disobedience. For individual factors, the paradigm allows to investigate how personality traits influence disobedience or to study the neuro-cognitive processes underlying disobedience.

Some novel theories combining a multi-method approach based on social psychology, neuroeconomics and neuroscience could thus emerge to understand better the mechanisms supporting disobedience. For instance, one could evaluate how empathy for the pain of the victim predicts disobedience and how the presence of a supporting group influences our capacity to feel empathy 41 and/or compassion for the ‘victim’ 42 . It could also be argued that the presence of a supporting group diffuses responsibility between individuals and increases obedience, by influencing how our brain processes agency and responsibility over our actions 28 , 43 , 44 , 45 . As the results obtained in the present study also indicated that feeling bad for the shocks delivered was statistically associated with prosocial disobedience, one could evaluate how the neural correlates of guilt 46 predicts prosocial disobedience and what historical, cultural and individual factors influence the feeling of guilt.

Six variants of the same task were tested in the present study, some inducing a higher prosocial disobedience rate than others. Statistical results showed that providing a reason—or aim—to justify obedience strongly decreased disobedience. Providing a monetary reward, even one as small as €0.05, also strongly decreased disobedience. Variant 2, in which volunteers were not given an aim or monetary reward, showed the highest disobedience rates. However, to study disobedience in ecological way, the paradigm should capture disobedience of participants even if they know that they are losing something (i.e., monetary rewards or the ‘trust’ of the experimenter asking them help for the study). Defying the orders of an authority generally involves social and/or monetary costs in real-life situations. I would thus not recommend using an experimental paradigm in which volunteers have no costs associated with defying the orders of the experimenter, as it would reduce the ecology of the disobedience act. Variants 3 and 6 involve two types of costs for resisting the orders of the experimenter: a monetary loss and deceiving the experimenter. In Variant 3, descriptive statistics showed that prosocial disobedience was lower compared to Variant 6. The main difference between these two variants was the presence of a free-choice condition. In my former studies 23 , 27 , volunteers frequently justified obedience in the coerced condition because they were given freedom in the free-choice condition (e.g. Participant 89 – English Translation: “ (…) In addition, I knew I could chose freely in the other condition not to send shocks—what I did ). In the present debriefings, some volunteers also reported that the presence of a free-choice condition was giving them enough freedom to accept to follow the orders in the coerced condition. In the supplementary analyses, results showed that when the monetary reward and the aim for obeying are identical, being given a free-choice condition reduces disobedience in the coerced condition. Therefore, Variant 6 appears to provide a good balance between reaching a reliable disobedience rate and finding volunteers who would refuse to produce physical harm on another human beings despite the monetary or social costs associated with defying orders.

Another approach would be to pre-select people who are predicted to be more disobedient. Personality questionnaires indicated that scoring low on the authority and on the purity subscale of the MFQ was strongly associated with a higher prosocial disobedience rate. The link between one’s own relationship to authority and prosocial disobedience observed here replicates another study conducted on the first generation of Rwandese after the 1994 genocide 47 . One’s own relationship to authority thus appears to be a reliable predictor variable in order to pre-select a sample that is more likely to disobey immoral orders.

In the present paper, administering a real mildly painful shock in exchange or not for a small monetary gain was described as an ‘immoral’ act. The notion of what is moral or not can highly differ between individuals 48 , for both academics and volunteers participating in an experiment. Humans are indeed sensitive to different competing issues of morality, a key reason for rescuing persecuted people 49 . In accordance with this observation, the present results indicated that moral reasons were a critical factor associated with the prosocial disobedience rate: the more shocking partners was considered as immoral, the more volunteers disobeyed. However, considering an action as against one’s own moral values does not necessarily translate to a refusal—especially when this order is in line with the Law. An extreme example is soldiers who have perpetrated acts that transgressed their moral beliefs but were issued by their superior in combat 50 . A core question for future research remains: Why are some people capable of putting their own moral standards above the social costs associated with defying orders?

Results indicate that the more volunteers felt responsible during the task, and the worse they felt for sending shocks to the ‘victim’, the higher was their prosocial disobedience. In another study, we observed that obeying orders reduced the feeling of responsibility, how bad and how sorry volunteers felt compared to being free to decide 26 . One hypothesis is that individuals who have preserved a feeling of responsibility and feeling bad—even under command—could more easily defy immoral orders. However, future studies are necessary to better understand the neuro-cognitive processes that prevent an individual from complying with immoral orders. As this paradigm is adapted to neuroimaging measurements, a whole range of studies could now be conducted.

It has been previously suggested that a strong identification with the experimenter giving orders is associated with higher obedience 36 . However, in the present paper, correlations between prosocial disobedience and identification with the experimenter were in favor of H0 with both the frequentist and the Bayesian approaches. In a former study, we also observed that identification to the experimenter was not a critical aspect for explain (dis)obedience. We observed that the generation of Rwandese born after the genocide and tested in Rwanda reported a higher identification to the experimenter than the same generation of Rwandese but tested in Belgium 47 . However, the latter group had a higher prosocial disobedience rate than the former group. Future studies must thus be conducted to understand how the identification with the person giving orders could influence obedience and its weight compared to other social, cultural and individual variables.

Although some volunteers reported that they felt a bit stressed and anxious during the task when they were in the role of the agent, the overwhelming majority did not report any negative psychological feelings. None of the participants withdrew from the experiment and none reported long-term negative psychological effects.

Nowadays, it has become difficult to find volunteers who do not know Milgram’s studies given the high media coverage, including movies, radio soaps, books, podcast and documentaries. One could expect that knowing Milgram would prevent people to obey. However, for the large majority of volunteers, it appears that this is not the case. In previous studies that I conducted with a relatively similar paradigm, the disobedience rate was drastically low (i.e. 3.3%) even if participants were university students knowing Milgram’s studies. In the present study, almost all the volunteers who participated in the present study knew Milgram and explicitly mentioned him during the oral debriefings or before starting the experiment. Yet for those who disobeyed, almost none reported that the reason for disobedience was that they thought it was the aim of the experiment. Further, there was no statistical relationship between prosocial disobedience and believing that it was the aim of the study. It does not mean that knowing Milgram would not influence at all disobedience. It rather suggests that knowing Milgram is not the main factor influencing one’s decision to obey or not an experimenter. It is also possible that since in this experiment shocks were real and not fake such as in Milgram’s studies, participants considered that this was indeed not a study aiming to replicate Milgram.

As far as I have observed, the main problem associated with knowing Milgram’s studies is that volunteers believe that I also have hidden aims and procedures when they enter the experimental room. Several volunteers reported that they only realized that my explanations for the task were true when they were explicitly offered the choice to decide which role to play first and/or when they started receiving the shocks. This is a general concern in psychological studies: The high use of cover stories can also impact other research, as volunteers start to develop a mistrust in what researchers tell them.

Results indicated that who were agents second tend to act similarly as those who were agents first, by sending a relatively similar amount of shocks. Of note, this is an effect that we also observed in past studies on the effect of obeying orders on cognition 23 , 26 , 43 . Nonetheless, in none of those studies we observed that the order of the role had a statistical influence on the neuro-cognitive processes targeted. However, the influence on role reversal on disobedience and related neuro-cognitive processes has still to be investigated in future studies.

The present paradigm is ecological in the sense that volunteers are facing decisions that have a real, physical impact on another human being. However, at the moment I only have little evidences that this paradigm has ecological validity to reflect obedience in real life situations, especially regarding “destructive disobedience” 17 . Caution is indicated when making inference from laboratory studies to complex social behaviours, such as those observed during genocides 16 . My main evidence at the moment is that the very low rate of prosocial disobedience observed in the first generation of post-genocide Rwandans tested in Rwanda using this paradigm 47 is consistent with the fact that deference to authority had already been emphasized by academics as an important factor in the 1994 genocide 4 , 51 . Individual scores on deference to authority in Caspar et al. 47 was the best predictive factor for prosocial disobedience in that former paradigm, thus suggesting some ecological validity. A promising approach would be to recruit “Righteous Among the Nations”, individuals who really saved lives during genocides. Testing this population with the present paradigm would put the ecological validity of this paradigm to the test.

People’s ability to question and resist immoral orders is a fundamental aspect of individual autonomy and of successful societies. As Howard Zinn famously wrote: “ Historically, the most terrible things—war, genocide, and slavery—have resulted not from disobedience, but from obedience ”. Understanding how individuals differ in the extent to which they comply with orders has undeniably several societal implications. They range from understanding how evolving in highly hierarchical environments — such as the military or prisons—influences moral behaviours, to developing interventions that would help to prevent blind obedience and help to resist calls to violence in vulnerable societies. However, since Milgram’s studies, the topic of disobedience has been mostly studied by social psychologists using adapted versions of the initial paradigm developed by Milgram. I hope that with this novel approach, (dis)obedience research will be given a new boost and will be considered by other scientific disciplines seeking to understand better human behaviours.

Data availability

Data are made available on OSF (DOI: https://doi.org/10.17605/OSF.IO/2BKJC ).

Milgram, S. Behavioral study of obedience. Psychol. Sci. Public Interest 67 (4), 371–378. https://doi.org/10.1037/h0040525 (1963).

Article   CAS   Google Scholar  

H. Zinn, The Zinn Reader: Writings on Disobedience and Democracy . Seven Stories Press, 1997.

Roisin, J. Dans la nuit la plus noire se cache l'humanité: Récits des justes du Rwanda. Les Impressions nouvelles (2017).

Fox, N. & Nyseth Brehm, H. I decided to save them: Factors that shaped participation in rescue efforts during genocide in Rwanda. Soc. Forces 96 (4), 1625–1648. https://doi.org/10.1093/sf/soy018 (2018).

Article   Google Scholar  

S. Milgram, Obedience to Authority: an Experiment View . Harper & Row, 1974.

Blass, T. Understanding behavior in the Milgram obedience experiment: The role of personality, situations, and their interactions. J. Pers. Soc. Psychol. 60 (3), 398–413. https://doi.org/10.1037/0022-3514.60.3.398 (1991).

Article   ADS   Google Scholar  

Dolinski, D. & Grzyb, T. The (doubtful) role of financial reward in obedience to authority. J. Soc. Psychol. 159 (4), 490–496. https://doi.org/10.1080/00224545.2018.1505708 (2019).

Article   PubMed   Google Scholar  

Haslam, N., Loughnan, S. & Perry, G. Meta-milgram: An empirical synthesis of the obedience experiments. PLoS ONE 9 (4), e93927. https://doi.org/10.1371/journal.pone.0093927 (2014).

Article   ADS   CAS   PubMed   PubMed Central   Google Scholar  

Fagin-Jones, S. & Midlarsky, E. Courageous altruism: Personal and situational correlates of rescue during the Holocaust. J. Posit. Psychol. 2 (2), 136–147. https://doi.org/10.1080/17439760701228979 (2007).

Bègue, L. et al. Personality predicts obedience in a milgram paradigm. J. Pers. 83 , 299–306. https://doi.org/10.1111/jopy.12104 (2015).

S. P. Oliner and P. M. Oliner, The altruistic personality: Rescuers of Jews in Nazi Europe . New York, NY, US: Free Press, 1988, pp. xxv, 419.

Baumrind, D. Some thoughts on ethics of research: After reading Milgram’s “Behavioral Study of Obedience”. Am. Psychol. 19 (6), 421–423. https://doi.org/10.1037/h0040128 (1964).

A. G. Miller, The obedience experiments: A case study of controversy in social science . New York, NY, England: Praeger Publishers, 1986, pp. ix, 295.

Perry, G. Behind the Shock Machine: The Untold Story of the Notorious Milgram Psychology Experiments (New Press, 2013).

Google Scholar  

Slater, M. et al. A virtual reprise of the stanley milgram obedience experiments. PLoS ONE 1 (1), e39. https://doi.org/10.1371/journal.pone.0000039 (2006).

Article   ADS   PubMed   PubMed Central   Google Scholar  

Burger, J. M. Replicating Milgram: Would people still obey today?. Am. Psychol. 64 (1), 1–11. https://doi.org/10.1037/a0010932 (2009).

Miller, A. G. Reflections on “replicating milgram” (Burger, 2009). Am. Psychol. 64 (1), 20–27. https://doi.org/10.1037/a0014407 (2009).

Griggs, R. A. & Whitehead, G. I. Coverage of recent criticisms of Milgram’s obedience experiments in introductory social psychology textbooks. Theory Psychol. 25 (5), 564–580. https://doi.org/10.1177/0959354315601231 (2015).

T. Blass, Obedience to Authority: Current Perspectives on the Milgram Paradigm . Psychology Press, 1999.

Kelman, H. C. Human use of human subjects: The problem of deception in social psychological experiments. Psychol. Bull. 67 (1), 1–11. https://doi.org/10.1037/h0024072 (1967).

Article   CAS   PubMed   Google Scholar  

Beauvois, J.-L., Courbet, D. & Oberlé, D. The prescriptive power of the television host. A transposition of Milgram’s obedience paradigm to the context of TV game show. Eur. Rev. Appl. Psychol. 62 (3), 111–119. https://doi.org/10.1016/j.erap.2012.02.001 (2012).

Frank, J. D. Experimental studies of personal pressure and resistance: I. experimental production of resistance. J. Gen. Psychol. 30 (1), 23–41. https://doi.org/10.1080/00221309.1943.10544454 (1944).

Caspar, E. A., Christensen, J. F., Cleeremans, A. & Haggard, P. Coercion changes the sense of agency in the human brain. Curr. Biol. 26 (5), 585–592. https://doi.org/10.1016/j.cub.2015.12.067 (2016).

Article   CAS   PubMed   PubMed Central   Google Scholar  

Caspar, E. A., Vuillaume, L., Magalhães de Saldanha da Gama, P. A. & Cleeremans, A. The Influence of (Dis)belief in free will on immoral behavior. Front. Psychol. https://doi.org/10.3389/fpsyg.2017.00020 (2017).

Article   PubMed   PubMed Central   Google Scholar  

Caspar, E. A., Cleeremans, A. & Haggard, P. Only giving orders? An experimental study of the sense of agency when giving or receiving commands. PLoS ONE 13 (9), e0204027. https://doi.org/10.1371/journal.pone.0204027 (2018).

Caspar, E. A., Ioumpa, K., Keysers, C. & Gazzola, V. Obeying orders reduces vicarious brain activation towards victims’ pain. Neuroimage 222 , 117251. https://doi.org/10.1016/j.neuroimage.2020.117251 (2020).

Caspar, E. A., LoBue, S., Magalhães de Saldanha da Gama, P. A., Haggard, P. & Cleeremans, A. The effect of military training on the sense of agency and outcome processing. Nat. Commun. https://doi.org/10.1038/s41467-020-18152-x (2020).

Beyer, F., Sidarus, N., Bonicalzi, S., & Haggard, P. Beyond self-serving bias: diffusion of responsibility reduces sense of agency and outcome monitoring. Soc Cogn Affect Neurosci. 12 (1), 138–145. https://doi.org/10.1093/scan/nsw160 (2017).

Haslam, S. A., Reicher, S. D., Millard, K. & McDonald, R. “Happy to have been of service”: The Yale archive as a window into the engaged followership of participants in Milgram’s “obedience” experiments. Br. J. Soc. Psychol. 54 (1), 55–83. https://doi.org/10.1111/bjso.12074 (2015).

Faul, F., Erdfelder, E., Lang, A.-G. & Buchner, A. G*Power 3: A flexible statistical power analysis program for the social, behavior, and biomedical sciences. Behav. Res. Methods 39 , 175–191. https://doi.org/10.3758/BF03193146 (2007).

Yamauchi, K. T. & Templer, D. J. The development of a money attitude scale. J. Pers. Assess. 46 (5), 522–528. https://doi.org/10.1207/s15327752jpa4605_14 (1982).

Graham, J. et al. Mapping the moral domain. J. Pers. Soc. Psychol. 101 (2), 366–385. https://doi.org/10.1037/a0021847 (2011).

Dunwoody, P. T. & Funke, F. The aggression-submission-conventionalism scale: Testing a new three factor measure of authoritarianism. J. Soc. Polit. Psychol. 4 (2), 571–600. https://doi.org/10.5964/jspp.v4i2.168 (2016).

Jones, D. N. & Paulhus, D. L. Introducing the short dark triad (SD3): A brief measure of dark personality traits. Assessment 21 (1), 28–41. https://doi.org/10.1177/1073191113514105 (2014).

M. Davis, A multidimensional approach to individual differences in empathy. JSAS Catalog Sel. Doc. Psychol. , vol. 10, (1980).

Steffens, N. K., Haslam, S. A. & Reicher, S. D. Up close and personal: Evidence that shared social identity is a basis for the “special” relationship that binds followers to leaders. Leadersh. Q. 25 (2), 296–313. https://doi.org/10.1016/j.leaqua.2013.08.008 (2014).

Dienes, Z. Bayesian versus orthodox statistics: Which Side Are You On?. Perspect. Psychol. Sci. 6 (3), 274–290. https://doi.org/10.1177/1745691611406920 (2011).

Marsman, M. & Wagenmakers, E.-J. Bayesian benefits with JASP. Europ. J. Develop. Psychol. 14 (5), 545–555. https://doi.org/10.1080/17405629.2016.1259614 (2017).

JASP Team, ‘JASP (Version 0.14.10)’. 2019.

Benjamini, Y. & Hochberg, Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. J. Roy. Stat. Soc.: Ser. B (Methodol.) 57 (1), 289–300. https://doi.org/10.1111/j.2517-6161.1995.tb02031.x (1995).

Article   MathSciNet   MATH   Google Scholar  

Singer, T. et al. Empathy for Pain involves the affective but not sensory components of pain. Science 303 (5661), 1157–1162. https://doi.org/10.1126/science.1093535 (2004).

Article   ADS   CAS   PubMed   Google Scholar  

Valk, S. L. et al. Structural plasticity of the social brain: Differential change after socio-affective and cognitive mental training. Sci. Adv. 3 (10), e1700489. https://doi.org/10.1126/sciadv.1700489 (2017).

Caspar, E. A., Beyer, F., Cleeremans, A. & Haggard, P. The obedient mind and the volitional brain: A neural basis for preserved sense of agency and sense of responsibility under coercion. PLoS ONE 16 (10), e0258884. https://doi.org/10.1371/journal.pone.0258884 (2021).

Beyer, F., Sidarus, N., Bonicalzi, S. & Haggard, P. Beyond self-serving bias: diffusion of responsibility reduces sense of agency and outcome monitoring. Social Cognit. Affect. Neurosci. 12 (1), 138–145. https://doi.org/10.1093/scan/nsw160 (2017).

Haggard, P. Sense of agency in the human brain. Nat. Rev. Neurosci. https://doi.org/10.1038/nrn.2017.14 (2017).

Yu, H. et al. A generalizable multivariate brain pattern for interpersonal guilt. Cereb. Cortex 30 (6), 3558–3572. https://doi.org/10.1093/cercor/bhz326 (2020).

E. Caspar, D. Gishoma, and P. A. M. D. S. da Gama, ‘Obedience to authority in the aftermath of a genocide. A social neuroscience study in Rwanda’. PsyArXiv, Jun. 23, 2021. https://doi.org/10.31234/osf.io/a8r7y .

B. Gert, and J. Gert, ‘The Definition of Morality’, in The Stanford Encyclopedia of Philosophy , Fall 2020., E. N. Zalta, Ed. Metaphysics Research Lab, Stanford University, 2020. Accessed: Apr. 07, 2021. [Online]. Available: https://plato.stanford.edu/archives/fall2020/entries/morality-definition/

Gross, M. L. Jewish rescue in holland and france during the second world war: Moral cognition and collective action*. Soc. Forces 73 (2), 463–496. https://doi.org/10.1093/sf/73.2.463 (1994).

Shay, J. Moral injury. Psychoanal. Psychol. 31 (2), 182–191. https://doi.org/10.1037/a0036090 (2014).

Article   MathSciNet   Google Scholar  

Prunier, G. The Rwanda Crisis: History of a Genocide (C. Hurst & Co., 1998).

Download references

Acknowledgements

Emilie A. Caspar was funded by the F.R.S-FNRS.

Author information

Authors and affiliations.

Moral and Social Brain Lab, Department of Experimental Psychology, Ghent University, Henri Dunantlaan, 2, 9000, Ghent, Belgium

Emilie A. Caspar

Center for Research in Cognition and Neuroscience, Université Libre de Bruxelles, Brussels, Belgium

You can also search for this author in PubMed   Google Scholar

Contributions

E.A.C. developed the study concept and the method. Testing, data collection and data analysis were performed by E.A.C. E.A.C. wrote the manuscript.

Corresponding author

Correspondence to Emilie A. Caspar .

Ethics declarations

Competing interests.

The author declares no competing interests.

Additional information

Publisher's note.

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Supplementary information., rights and permissions.

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ .

Reprints and permissions

About this article

Cite this article.

Caspar, E.A. A novel experimental approach to study disobedience to authority. Sci Rep 11 , 22927 (2021). https://doi.org/10.1038/s41598-021-02334-8

Download citation

Received : 21 April 2021

Accepted : 15 November 2021

Published : 25 November 2021

DOI : https://doi.org/10.1038/s41598-021-02334-8

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

By submitting a comment you agree to abide by our Terms and Community Guidelines . If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

milgram experiment ethical considerations

IMAGES

  1. How to Make Milgram's Experiment Ethical

    milgram experiment ethical considerations

  2. PPT

    milgram experiment ethical considerations

  3. The Milgram Experiment

    milgram experiment ethical considerations

  4. Milgram Shock Experiment: Summary, Results, & Ethics

    milgram experiment ethical considerations

  5. Ethical Considerations

    milgram experiment ethical considerations

  6. Ethical Considerations

    milgram experiment ethical considerations

VIDEO

  1. Milgram Experiment: Shocking Obedience to Authority Revealed

  2. The Impact of Milgram’s Psychology Experiment || Psychology || Ettienne-Murphy

  3. Ethical Considerations in Research

  4. Will You OBEY Orders

  5. How Evil are You? Milgram Experiments (Hindi)

  6. The Milgram Experiment: Shocking Obedience

COMMENTS

  1. Milgram Shock Experiment

    Stanley Milgram, a psychologist at Yale University, carried out one of the most famous studies of obedience in psychology. He conducted an experiment focusing on the conflict between obedience to authority and personal conscience. Milgram (1963) examined justifications for acts of genocide offered by those accused at the World War II, Nuremberg ...

  2. The Milgram Experiment: Theory, Results, & Ethical Issues

    Ethical Issues in the Milgram Experiment Just two years before Milgram began his experiments, the American Psychological Association (APA) updated its code of ethics, which included the expectation that researchers would "respect the integrity and protect the welfare of the person or group with whom he is working" (APA, 1959, p. 280).

  3. Milgram Experiment: Overview, History, & Controversy

    Replications of the Milgram Experiment . While Milgram's research raised serious ethical questions about the use of human subjects in psychology experiments, his results have also been consistently replicated in further experiments. One review further research on obedience and found that Milgram's findings hold true in other experiments.

  4. The Milgram Experiment: Summary, Conclusion, Ethics

    The goal of the Milgram experiment was to test the extent of humans' willingness to obey orders from an authority figure. Participants were told by an experimenter to administer increasingly powerful electric shocks to another individual. Unbeknownst to the participants, shocks were fake and the individual being shocked was an actor.

  5. Milgram's Obedience Experiment: Insights and Implications

    The psychological distress experienced by some subjects raised serious concerns about the balance between scientific inquiry and ethical considerations. Despite these controversies, Milgram's work remains a cornerstone of obedience research, sparking countless debates and inspiring numerous follow-up studies.

  6. Credibility and Incredulity in Milgram's Obedience Experiments: A

    Gina Perry is an Australian writer and author of Behind the Shock Machine: The Untold Story of the Notorious Milgram Psychology Experiments (2012) and The Lost Boys: Inside Muzafer Sherif's Robber Cave Experiment (2018). Both works draw on extensive archival research and interviews with experimental participants. She completed her PhD at the University of Melbourne, where she is an associate ...

  7. What can Milgram and Zimbardo teach ethics committees and qualitative

    Philip Zimbardo's (1973) Stanford Prison Study and Stanley Milgram's (1974) Obedience study are convenient shorthand fall guys for justifying the necessity of ethics review. As with Adam and Eve's original sin producing the fall of man in the Christian faith, Zimbardo and Milgram are cast in this role, not only for use in psychology, but emblematic of the need to evaluate behavioral ...

  8. ETHICS ROUNDS: A classic study, revisited

    A second theme to emerge in the articles involves the relationship between methodology and ethics. Burger begins his discussion section by noting "People learning about Milgram's obedience studies often ask whether similar results would be found today. Ethical concerns prevent researchers from providing a definitive answer to that question."

  9. PDF The Milgram Experiment & Its Contemporary Replication Was It Ethical

    I won't be in the experiment any more! I refuse to go on!". 180 V: "I can't stand the pain!". 270 V: agonizing scream. 300 V: shout in desperation that he would no longer answer. 315 V: violent scream; vehemently reaffirmed that he was no longer a participant. 330 V: shrieked in agony. >330 V: no sounds.

  10. Ethics, Deception, and 'Those Milgram Experiments'

    Ethics, Deception, and 'Those Milgram Experiments' C. D. HERRERA ... experiment will be is a tall order for psychologists, or anyone else. At the same time, critics themselves have difficulty in showing what is wrong with deception, and how subjects in these experiments suffer. Hence, it becomes unclear what the psychologists, including Milgram ...

  11. Modern Milgram experiment sheds light on power of authority

    Milgram's original experiments were motivated by the trial of Nazi Adolf Eichmann, who famously argued that he was 'just following orders' when he sent Jews to their deaths. The new findings ...

  12. How the Classics Changed Research Ethics

    Nearly 60 years have passed since Stanley Milgram's infamous "shock box" study sparked an international focus on ethics in psychological research. Countless historians and psychology instructors assert that Milgram's experiments—along with studies like the Robbers Cave and Stanford prison experiments—could never occur today; ethics ...

  13. Contesting the "Nature" Of Conformity: What Milgram and Zimbardo's

    Abstract. Understanding of the psychology of tyranny is dominated by classic studies from the 1960s and 1970s: Milgram's research on obedience to authority and Zimbardo's Stanford Prison Experiment. Supporting popular notions of the banality of evil, this research has been taken to show that people conform passively and unthinkingly to both the ...

  14. Psychological research, obedience and ethics

    One of the best known studies in the history of psychology is the research on obedience carried out by Stanley Milgram in the 1960s. In his research Milgram demonstrated the lengths to which people are willing to go just because someone in authority tells them to do something. The studies Milgram conducted also raised the issue of ethics in ...

  15. 3 Key Ethical Considerations in Psychological Research

    Earlier in psychology history, many experiments were performed with highly questionable and even outrageous violations of ethical considerations. Milgram's infamous obedience experiment, for example, involved deceiving human subjects into believing that they were delivering painful, possibly even life-threatening, electrical shocks to another ...

  16. Milgram experiment

    Milgram experiment The setup of the "shock generator" equipment for Stanley Milgram's experiment on obedience to authority in the early 1960s. The volunteer teachers were unaware that the shocks they were administering were not real. Milgram included several variants on the original design of the experiment.

  17. Replicating Milgram

    Jerry Burger from Santa Clara University managed to do the seemingly impossible — he conducted a partial replication of the infamous Milgram experiment. Read on for valuable advice, and look for similar coverage in upcoming Observers. "It can't be done.". These are the first words I said to Muriel Pearson, producer for ABC News ...

  18. Ethical Problems

    The ethical issues involved with the Milgram experiment are as follows: deception, protection of participants involved, and the right to withdrawal. The experiment was deemed unethical, because the participants were led to believe that they were administering shocks to real people. The participants were unaware that the learner was an associate ...

  19. Explanations for Obedience

    Deception. Ethical Issues. Right to withdraw. Protection from harm. Milgram (1963) conducted one of the most famous and influential psychological investigations of obedience. He wanted to find out if ordinary American citizens would obey an unjust order from an authority figure and inflict pain on another person because they were instructed to.

  20. Milgram Experiment Ethics

    You can use it freely (with some kind of link), and we're also okay with people reprinting in publications like books, blogs, newsletters, course-material, papers, wikipedia and presentations (with clear attribution). Critique of the Milgram Experiment Ethics made experiments using deception illegal. The Stabley Milgram Experiment would never ...

  21. A novel experimental approach to study disobedience to authority

    The experiment of Stanley Milgram is one of the most (in)famous in psychology 1, within and beyond academia. Several variables account for this notoriety, such as the method used, the ethical issues associated, the enthralling results or the societal impact of the research topic. ... Beyond ethical considerations, the use of deception also ...

  22. What are the Milgram Experiment Ethical Issues?

    The Milgram Experiment was a series of experimental studies that took place in the 1960s to investigate how willing subjects were to obey an authority figure even when their actions directly conflicted with their personal conscience. The experiments proved to be extremely controversial and were considered to be highly unethical at the time, and ...

  23. The Milgram Experiment Ethical Issues in 2017

    Kaplan University. PS 499. May 29, 2017. The Milgram Experiment was done in by Stanley Milgram who was a psychologist at Yale. University and conducted an experiment focusing on the conflict ...

  24. A novel experimental approach to study disobedience to authority

    The experiment of Stanley Milgram is one of the most ... Beyond ethical considerations, the use of deception also indeed involves a doubt about whether or not volunteers truly believed the cover ...