Evidence of Brainwashing in Cults (Benjamin Zablocki, 2001)

NOTE: The following article is taken from the 5th chapter of Misunderstanding Cults: Searching for Objectivity in a Controversial Field, entitled, Towards a Demystified and Disinterested Scientific Theory of Brainwashing.


I have attempted to test the model as much as possible with the limited data that currently exist. I have relied on three sources of evidence. The first and most important of these consists of ethnographic studies of a wide variety of contemporary American charismatic cults conducted by myself and others. The first-hand opportunities I have had to watch (at least the public face of) charismatic resocialization in numerous cult situations has convinced me of the need to theorize about this phenomenon. The second source of data consists of interviews with former leaders of charismatic groups. Although I have only a handful of such interviews, they are particularly valuable for elucidating the process from the perspective of ‘management,’ rather than from the perspective of the subjects. The third source of data consists of reports of ex-members of cults, drawing heavily on scientifically sampled interviews that my students and I have conducted. Most of these respondents were interviewed at least twice over a roughly twenty-five-year period.

Because evidence in this field of study tends to be so bitterly contested, it is perhaps necessary to point out that my own studies in this area were all subject to rigorous and competitive peer review. Five of my studies were reviewed and funded by three organizations — the National Institute of Mental Health (2), and the National Institute of Health (1) — over a period extending from 1964 to 2001. On all of these I was the principal investigator, and the research designs are in the public record. During this same period, other research of mine in this same field of study was funded by peer-reviewed faculty research grants from all of the universities with which I have been affiliated: the University of California at Berkeley, the California Institute of Technology, Columbia University, and Rutgers University. It is a strange anomaly that this body of work seems to be generally respected throughout the social and behavioural sciences, with the exception of a small field, the sociology of new religious movements, where some try their best to hold it up to ridicule and disesteem.

Ethnographic Accounts

Bainbridge (1997) has argued that most ethnographic studies of cults have failed to find evidence of brainwashing. But it is more accurate to say that ethnographers have been divided on this subject. Lalich, Ofshe, Kent, and myself have found such evidence abundantly (Kent and Krebs 1998; Lalich 1993; Ofshe, Eisenberg et. al. 1974; Zablocki 1980).  Even Barker, Beckford, and Richardson, who are among the most hostile to the brainwashing conjecture, have found evidence of attempted brainwashing, although they have claimed that these attempts are largely or entirely unsuccessful (Barker 1984; Beckford 1985; Richardson, Harder et. al. 1972). Still other ethnographers (Balch 1985; Rochford, Purvis et. al. 1989) seem ambivalent on the subject and not sure what to make of the evidence. Others such as Palmer (1994) and Hall (1987, 2000) have been fairly clear about the absence of brainwashing in their observations.

Such disparity is to be expected. There is no reason to believe that all cults practise brainwashing any more than that all cults are violent or that all cults make their members wear saffron robes. Most ethnographers who did discover evidence of brainwashing in the cults they investigated were surprised by the finding. The fact that evidence of this sort has been repeatedly discovered by researchers who were not particularly looking for it suggests that the process really exists in some cults. I have observed fully developed brainwashing processes in some cults, partially developed ones in others, and none whatsoever in others. As ethnographic work in cults continues to accumulate, we should expect to find a similar degree of heterogeneity in published reports. certainly , there is abundant evidence of uncritically obedient behaviour in charismatic cults (Ayella 1990; Davis 2000; Katchen 1997; Lalich 1999; Lifton 1999; Wallis 1977), and this behaviour needs to be explained. The presence or absence of brainwashing may ultimately turn out to contribute to such an explanation.

When I first studied the Bruderhof thirty-five years ago, using ethnographic methods, I noticed a strong isomorphism between the phases of Bruderhof resocialization and the phases of brainwashing in Chinese re-education centres described by Lifton. Since I could think of no other reason why the Bruderhof would support such a costly and labour-intensive resocialization program if it were not to create deployable agents with long-term loyalty to the community, I hypothesized that something akin to brainwashing must be going on. My observations over the next thirty-five years have only strengthened my confidence in the correctness of this hypothesis. Bruderhof members were never kept from leaving by force or force threat. But the community put a lot of time and energy into assuring that defections would made rare and difficult by imbuing in its members an uncritical acceptance of the teachings of the community and a terror of life outside the community.1

Some (but not all) of the other cultic groups I have lived with as a participant-observer have shown signs of a brainwashing process at work. Individuals being plucked suddenly out of the workday routine of the group, appearing to become haggard with lack of sleep for prolonged periods, secretiveness and agitation, alternating periods of shunning and warm communal embrace, all suggest the presence of such a process. Some of these people, years later, having left the cult, have confirmed to me that such a process is what they went through when I observed them under this stress. According to my ethnographic observations, some sort of fully or partially developed brainwashing process figures in the resocialization of at least half of the cults I have studied during at least some phases of their history.

Geronda Dositheos with personalized 'Geronda' sweater

Leader Accounts

A second source of evidence may be found in reports given by people who were actually responsible for practising brainwashing with their fellow cult members. Several cult leaders  who left their groups have since apologized to other ex-members for having subjected them to brainwashing methods. One such former cult leader put it this way:

“What you have to understand is that, for us, breaking the spirit … emptying out our ego, is very very important. And any means to that end … well, we would have said it was justified. And over the years we developed [by trial and error] ways of accomplishing this [task]. It was only after I was finished with [the cult] and living in the world again that I did some reading and realized how similar [our techniques] were to what the Communists did – to brainwashing. I think you would have to say that what we did was a kind of brainwashing even if we didn’t mean it to be so.’

In another case I interviewed the widow of a cult leader who had died and whose cult had disbanded soon thereafter. She said the following:

‘Those kinds of things definitely happened [on quite a few occasions]. It’s not like we sat down and said, hey we’re going to brainwash everybody. That would have been crazy. It’s more like we knew how important our mission was and how [vulnerable it was] to treachery. I think we got a little paranoid about being overcome by treachery within, especially after Gabe and Helen left and started saying those things about us. So everybody had to be tested. I had to be tested. Even he [the leader] had to be tested. We all knew it and we all [accepted it]. So we would pull a person out of the routine and put him in solitary for awhile. Nobody could talk to him except [my husband] and maybe a few others. I couldn’t even talk to him when I brought him meals. That was usually my job … At first it was just isolation and observation and having deep long talks far into the night about the mission. We didn’t know anything about brainwashing or any of that stuff. But gradually the things you describe got in there too somehow. Especially the written confessions. I had to write a bunch of them towards the end when [X] was sick. Whatever you wrote was not enough. They always wanted more, and you always felt you were holding out on them. Finally your confessions would get crazy, they’d come from your wildest fantasies of what they might want. At the end I confessed that I was killing [my husband] by tampering with his food because I wanted to – I don’t know – be the leader in his place I guess. All of us knew it was bullshit but somehow it satisfied them when I wrote that … And, even though we knew it was bullshit, going through that changed us. I mean I know it changed me. It burned a bridge … [T]here was no going back. You really did feel you changed into being a different person in a weird sort of way.’

Perhaps the closest thing I have found to a smoking gun in this regard has to do with a sociology professor who became a charismatic cult leader. Two of this cult leader’s top lieutenants independently spoke to me on this subject. Both of these respondents described in great detail how they assisted in concerted campaigns to brainwash fellow cult members. Both felt guilty about this and found the memory painful to recount. One of them indicated that the brainwashing attempt was conscious and deliberate:

‘During her years in academia, Baxter became very interested in mass social psychology and group behaviour modification. She studied Robert Jay Lifton’s work on thought reform; she studied and admired ‘total’ communities such as Synanon, and directed methods of change, such as Alcoholic Anonymous. She spoke of these techniques as positive ways to change people.’ (Lalich 1993: 55)

In this cult, which has since disbanded, there seems to be general consensus among both leaders and followers that systematic brainwashing techniques were used on a regular basis and were successful in their aim of producing deployable agents.

Geronda Ephraim, Monks, Devotees

Ex-member Accounts

Our third source of evidence is the most controversial. There has been a misguided attempt to deny the validity of negative ex-member accounts as a source of data about cults. They’ve been condemned as ‘atrocity tales’ (Richardson 1998: 172), and Johnson (1998: 118) has dismissed them categorically by alleging that ‘the autobiographical elements of apostate narratives are further shaped by a concern that the targeted religious groups be painted in the worst possible light.’


The apostate role has been defined by Bromley (1997) largely in terms of the content of attitudes towards the former cult. If these attitudes are negative and expressed collectively in solidarity with other negatively disposed ex-members, they constitute evidence that the person must not be an ordinary ex-member but an ‘apostate.’ This is a direct violation of Robert Meron’s (1968) admonition that role sets be defined in terms of shared structural characteristics, not individual attitudes. What if this same logic were used to denigrate abused spouses who choose to be collectively vocal in their complaints? Nevertheless, this perspective on so-called ‘apostate accounts’ has been widely influential among cult scholars.

David Bromley is a sociologist theorist of great personal integrity but limited field experience. I think that if Bromley and his followers could just once sit down with a few hundred of these emotionally haunted ex-members whom they blithely label ‘apostates’,’ and listen to their stories, and see for themselves how badly most of them would like nothing more than to be able to put the cult experience behind them and get on with their lives, they would be deeply ashamed of the way they have subverted role theory to deny a voice to a whole class of people.

Fr. Silouanos Coutavas

Dawson (1995) has correctly pointed out that there are methodological problems involved in using accounts of any kind as data. We need to be careful not to rely only on ex-member accounts. Triangulation of data sources is essential. But even the reports of professional ethnographers are nothing more than accounts, and thus subject to the same sort of limitations. Ex-member accounts have been shown to have reliability and validity roughly equivalent to the accounts given by current cult members (Zablocki 1996).

Solomon (1981) has provided some empirical support for the argument that those with stormy exits from cults and those with anti-cult movement affiliations are more likely to allege that they have been brainwashed than those with relatively uneventful exits and no such affiliation. ‘Cult apologists’ have made much of the finding that ex-members affiliated with anti-cult organizations are more likely to allege brainwashing than those who are not. Their hatred of the anti-cult movement has blinded them to two important considerations: (1) The causal direction is by no means obvious — it is at least as likely that those who were brainwashed are more likely to seek out anti-cult organizations as support groups as that false memories of brainwashing are implanted by anti-cult groups into those ex-members who fall into their clutches; and (2) Although the percentages may be lower, some ex-members who don’t affiliate with anti-cult groups still allege brainwashing.


Many ex-members of cults find brainwashing the most plausible explanation of their own cult experiences. While some might be deluding themselves to avoid having to take responsibility for their own mistakes, it strains credulity to imagine that all are doing so. Here, just by way of example, are excerpts from interviews done with five ex-members of five different cults. None of these respondents was ever affiliated, even marginally, with an anti-cult organization:

‘They ask you to betray yourself so gradually that you never notice you’re giving up everything that makes you who you are and letting them fill you up with something they think is better and that they’ve taught you to believe is something better.’

‘What hurts most is that I thought these people were my new friends, my new family.  It wasn’t until after that I realized how I was manipulated little step by little step. Just like in Lifton; it’s really amazing when you think of it … couldn’t just be a coincidence …  I don’t know if you can understand it, but what hurts most is not that they did it but realizing that they planned it out so carefully from the beginning. That was so cold.’

‘I’ve never been able to explain it to people who weren’t there. I don’t really understand it myself. But black was white, night was day, whatever they told us to believe, it was like a test. The more outrageous the idea the greater the victory, when I could wrap my mind around it and really believe it down to my toes. And, most important, be prepared to act on it just like if it was proven fact. That’s the really scary part when I look back on it.’

‘In the frame of mind I was in [at the time], I welcomed the brainwashing. I thought of it like a purge. I needed to purge my old ways, my old self. I hated it and I felt really violent toward it … I wanted to wash it all away and make myself an empty vehicle for [the guru’s] divine plan … [Our] ideal was to be unthinking obedient foot soldiers in God’s holy army.’

Many wax particularly eloquent on this subject when interviewed in the aftermath of media events involving cultic mass suicides or murders. The fifth respondent said the following:

‘It makes me shudder and … thank God that I got out when I did. ‘Cause that could have been me doing that, could have been any of us. [I have] no doubt any one of us would have done that in the condition we all were in — killed ourselves, our kids, any that [the leaders] named enemies.’

I have quoted just five ex-members because of limitations of space. Many more could be found. Thousands of ex-members of various groups (only a small minority of whom have ever been interviewed by me) have complained of being brainwashed. Contrary to the allegations of some ‘cult apologists,’ very few of these are people who had been deprogrammed (and presumably brainwashed into believing that they had been brainwashed). The accounts of these people tend often to agree on the particulars of what happened to them, even though these people may never have talked with one another.

Another striking aspect of these brainwashing accounts by ex-members is that they are held to consistently for many years. I have interviewed many ex-cult members twenty to thirty years after leaving the cult, and have yet to have a single case of a person who alleged brainwashing immediately after leaving the cult, later recant and say it wasn’t true after all. More than anything else, this consistency over extended periods of time convinces me that ex-member accounts often may be relied on. Even if some of the details have been forgotten or exaggerated with the passage of time, the basic outline of what happened to them is probably pretty accurate. All in all, therefore, I think it is fair to conclude, both from accumulated ethnographic and ex-member data, that brainwashing happens to at least some people in some cults.

Greek Coffee (TX)

Incidence and Consequences

Finally, we come to the aspect of brainwashing theory for which our data are sketchiest, the one most in need of further research. How often does brainwashing actually occur (incidence) 2 and how significant are its consequences?

Defining what we mean by incidence is far from a simple matter. In the reporting of brainwashing there are numerous false positives and false negatives, and no consensus as to whether these errors lead to net underestimation or net overestimation. Several factors can produce false positives. Unless the term is precisely defined to respondents, some answers will reflect folk definitions of the term. It might mean little more to them than that they believe they were not treated nicely by their former cults. Other respondents may share our definition of the term, but answer falsely out of a desire to lay claim to the victim role or out of anger towards the cult. False negatives also can occur for several reasons. Most significantly, current members (as well as ex-members who still sympathize with the cult) may deny brainwashing to protect the cult. Others may understand the term differently than do the interviewers, and still others may be embarrassed to admit that they had been brainwashed. These errors can be minimized but hardly eliminated by in-depth interviewing in which respondents are asked not merely to label but to describe the process they went through.

There is insufficient space in this chapter to discuss these important methodological issues. I will therefore merely state the criteria upon which I base my own measurement. I treat incidence as a ratio of X to Y. in Y are included all those who were fully committed members of a cult for a year or more, but who are currently no longer affiliated with any cult. 3 In X are included those members of the Y set who both claim to have been brainwashed and who are able to give evidence of the particulars of their own brainwashing experience (at least through phase 2) consistent with those discussed in the previous section of this chapter.

In the handful of systematic studies that have been done, estimates of brainwashing incidence seem to cluster around 10% (plus or minus 5%) of former cult members (Katchen 1997; Wright 1987; Zablocki, Hostetler et. al. in press). However, there is tremendous variation in estimates for this number given by people working in this field. Ignoring those scholars who deny that brainwashing is ever attempted or ever successful, I have heard anecdotal estimates as low as <0.1% and as high as 80%, given by ethnographers.

Stuart Wright’s (1987) data on voluntarily exiting ex-members indicate that 9% say they had been brainwashed. This study is noteworthy because it examined ex-members of a variety of different cults rather than just one. It relied, however, on each respondent’s own definition of what it meant to be brainwashed.

My national longitudinal study (Zablocki 1980) relied primarily on a two-stage sampling procedure in which geographical regions were first selected and groups then sampled within these regions. I have followed 404 cases, most of them surveyed at least twice over intervals extending up to twenty-five years. Of those who were interviewed, 11% meet the criteria for having been brainwashed discussed above. Interestingly, all those in my sample who claim to have been brainwashed stick to their claims even after many years have passed. My own study is the only one that I know of that has repeatedly interviewed members and former members over several decades.

Another issue is whether overall incidence among the ex-member population is the most meaningful statistic to strive for given the heterogeneity among cults and types of cult member. Cults vary in the proportion of their members they attempt to brainwash from 0% to 100%. Since brainwashing significantly increases exit costs (according to hypothesis 8), it follows that examples of brainwashed individuals will be somewhat over-represented among current cult members and somewhat under-represented among ex-members.

The incidence, among ex-members, is higher (24% in my sample) when the relevant population is confined to a cult’s ‘inner circle,’ the core membership surrounding the leader. In an important  and neglected article, Wexler (1995) makes the point that it is simplistic to think of a cult as comprising only a leader and a homogeneous mass of followers. Most cults have a third category of membership, a corps of lieutenants, surrounding the leader, which Wexler refers to as a ‘decision elite.’ It follows from the hypotheses discussed earlier that we should expect attempts to brainwash to be concentrated among members in this category.

One study suggests that incidence is also higher among adults who grew up in cults (Katchen 1997). My own ethnographic observation supports the last point, and further suggests that cults under extreme stress become more likely to engage in brainwashing or to extend already existing brainwashing programs to a much wider circle of members.

With regard to consequences, we must distinguish between obedience consequences and traumatic consequences. Uncritical obedience is extinguished rapidly, certainly within a year of exiting if not sooner. The popular idea that former cult members can be programmed to carry obedience compulsions for specific acts to be performed long after membership in the cult has ceased is, in my opinion, wholly a myth based largely on a movie, The Manchurian Candidate. I know of nobody who has ever seen even a single successful instance of such programming. However, many brainwashed ex-members report that they would not feel safe visiting the cult, fearing that old habits of obedience might quickly be reinstilled.

There is evidence, in my data set, of persistent post-traumatic effects. The majority of those who claim to have been brainwashed say that they never fully get over the psychological insult, although its impact on their lives diminishes over time. The ability to form  significant bonds with others takes a long time to heal, and about a third wind up (as much as a quarter of a century later) living alone with few significant social ties. This is more than double the proportion of controls (cult participants who appeared not to have been brainwashed) that are socially isolated twenty-five years later. Visible effects also linger in the ability to form new belief commitments. In about half there is no new commitment to a belief community after two years. By twenty-five years, this has improved, although close to 25% still have formed no such commitment. Occupationally, they tend to do somewhat better, but often not until having separated from the cult for five to ten years.

Baptism at Holy Archangels (TX)


We can conclude from all of the above that those who claim that cultic brainwashing does not exist and those who claim it is pandemic to cults are both wrong. Brainwashing is an administratively costly and not always effective procedure that some cults use on some of their members. A few cults rely heavily on brainwashing and put all their members through it. Other cults do not use the procedure at all. During periods of stressful confrontation, either with external enemies or among internal factions, or in attempts to cope with failed apocalyptic prophecies, it is not uncommon for brainwashing suddenly to come to play a central role in the cult’s attempts to achieve order and social control. At such times, risk of uncritically obedient violent aggression or mass suicide may be heightened.

Hopefully, it will be clear from this chapter that brainwashing has absolutely nothing to do with the overthrow of ‘free will’ or any other such mystical or non-scientific concept. People who have been brainwashed are ‘not free’ only in the sense that all of us, hemmed in on all sides as we are by social and cultural constraints, are not free. The kinds of social constraints involved in brainwashing are much more intense than those involved in socializing many of us to eat with knives and forks rather than our hands. But the constraints involved differ only in magnitude and focus, not in kind. Any brainwashed cult member always retains the ability to leave the cult or defy the cult as long as he or she is willing to pay the mental and emotional price (which may be considerable) that the cult is able to exact for so doing.

As I finish this chapter, a number of European nations are debating the advisability of anti-brainwashing laws, some of which eventually may be used to inhibit freedom of religious expression. In light of this trend a number of colleagues have criticized me, not on the grounds that my facts are incorrect, but that my timing is unfortunate. One socked me with the following, particularly troubling, complaint: “Ben, if you had discovered evidence, in 1942, of a higher prevalence among Jews than non-Jews of the Tay-Sachs genetic defect, would you have published your findings in a German biology journal?” Ultimately, although I respect the sentiments behind my colleagues’ concerns, I must respectfully disagree with their fastidious caution. It never works to refuse to look at frightening facts. They only become larger, more   frightening, and more mystically permeated when banished to one’s peripheral vision. A direct, honest acknowledgement of the limited but significant role that brainwashing plays in producing uncritical obedience in some cults will serve, in the long run, to lessen paranoid reactions to ‘the threat of the cults,’ rather than increase them.


  1. Bruderhof members, particularly those in responsible positions, are never fully trusted until they have gone through the ordeal of having been put into the great exclusion (being sent away) and then spiritually fought their way back to the community. Such exclusion serves as the ultimate test of deployability. Is the conversion deep enough to hold even when away from daily reinforcement by participation in community life? The degree to which the Bruderhof stresses the importance of this ideal serves as additional evidence that the creation of deployable agents is a major aim of the socialization process.
  2. A related question is what portion of those a cult attempts to brainwash actually get brainwashed. No data have been collected on this issue to the best of my knowledge.
  3. I do not distinguish between voluntary and involuntary mode of exit in my measure because my sample includes only an insignificant number (less than one-half of one percent) who were deprogrammed out of their cults.

Brainwashing as a Scientific Concept (Benjamin Zablocki, 2001)

NOTE: The following article is taken from the 5th chapter of Misunderstanding Cults: Searching for Objectivity in a Controversial Field, entitled, Towards a Demystified and Disinterested Scientific Theory of Brainwashing.


What I am presenting here is not a ‘new’ theory of brainwashing but a conceptual model of the foundational theory developed in the mid-twentieth century by Lifton, Schein, and Sargant as it applies to charismatic collectivities. Because its scientific stature has been so frequently questioned, I will err on the side of formality by presenting a structured exposition of brainwashing theory in terms of eight definitions and twelve hypotheses. Each definition includes an operationalized form by which the trait may be observed. If either of the first two hypotheses disconfirmed, we must conclude that brainwashing is not being attempted in the cult under investigation. If any of the twelve hypotheses is disconfirmed, we must conclude that brainwashing is not successful in meeting its goals within that cult.

I do not pretend that the model outlined here is easy to test empirically, particularly for those researchers who either who either cannot or will not spend time immersing themselves in the daily lives of cults, or for those who are not willing, alternatively, to use as data the detailed retrospective accounts of ex-members. However, it should be clear that the model being proposed here stays grounded in what is empirically testable and does not involve mystical notions such as loss of free will or information disease (Conway and Siegelman 1978) that have characterized many of the extreme ‘anti-cult models.’

Nor do I pretend that this model represents the final and definitive treatment of this subject. Charismatic influence is still a poorly understood subject on which much additional research is needed. With few exceptions, sociology has treated it as if it were what engineers call a ‘black box,’ with charismatic inputs coming in one end and obedience outputs going out the other. What we have here is a theory that assists in the process of opening  this black box to see what is inside. It is an inductive theory, formed largely from the empirical generalizations of ethnographers and interviewers. The model itself presents an ideal-type image of brainwashing that does not attempt to convey the great variation among specific obedience-inducing processes that occur across the broad range of existing cults. Much additional refinement in both depth and breadth will certainly be needed.


EE 2018

D1. Charisma is defined, using the classical Weberian formula, as a condition of ‘devotion to the specific and exceptional sanctity, heroism, or exemplary character of an individual person, of the normative patterns or order revealed or ordained by him’ (Weber 1947: 328). Being defined this way, as a condition of devotion, leads us to recognize that charisma is not to be understood simply in terms of the characteristics of the leader, as it has come to be in popular usage, but requires an understanding of the relationship between leader and followers. In other words, charisma is a relational variable. It is defined operationally as a network of relationships in which authority is justified (for both superordinates and subordinates) in terms of the special characteristics discussed above.

D2. Ideological Totalism is a sociocultural system that places high valuation on total control over all aspects of the outer and inner lives of participants for the purpose of achieving the goals of an ideology defined as all important. Individual rights either do not exist under ideological totalism or they are clearly subordinated to the needs of the collectivity whenever the two come into conflict. Ideological totalism has been operationalized in terms of eight observable characteristics: milieu control, mystical manipulation, the demand for purity, the cult of confession, ‘sacred science,’ loading the language, doctrine over person, and the dispensing of existence (Lifton 1989: chap. 22).1

D3. Surveillance is defined as keeping watch over a person’s behaviour, and, perhaps, attitudes. As Hechter (1987) has shown, the need for surveillance is the greatest obstacle to goal achievement among ideological collectivities organized around the production of public goods. Surveillance is not only costly, it is also impractical for many activities in which agents of the collectivity may have to travel to act autonomously and at a distance. It follows from this that all collectivities pursuing public goals will be motivated to find ways to decrease the need for surveillance. Resources used for surveillance are wasted in the sense that they are unavailable for the achievement of collective goals.

D4. A deployable agent is one who is uncritically obedient to directives perceived as charismatically legitimate (Selznick 1960). A deployable agent can be relied on to continue to carry out the wishes of the collectivity regardless of his own hedonic interests and in the absence of any external controls. Deployability can be operationalized as the likelihood that the individual will continue to comply with hitherto ego-dystonic demands of the collectivity (e.g., mending, ironing, mowing the lawn, smuggling, rape, child abuse, murder) when not under surveillance.

D5. Brainwashing is an observable set of transactions between a charismatically structured collectively and an isolated agent of the collectivity, with the goal of transforming the agent into a deployable agent. Brainwashing is thus a process of ideological resocialization carried out within a structure of charismatic authority.

The brainwashing process may be operationalized as a sequence of well-defined and potentially observable phases. These hypothesized phases are (1) identity stripping, (2) identification, and (3) symbolic death/rebirth. The operational definition of brainwashing refers to the specific activities attempted, whether or not they are successful, as they are either observed directly by the ethnographer or reported in official or unofficial accounts by members or ex-members. Although the exact order of phases and specific steps within phases may vary from group to group, we should always expect to see the following features, or their functional equivalents, in any brainwashing system: (1) the constant fluctuation between assault and leniency; and (2) the seemingly endless process of confession, re-education, and refinement of confession.

D6. Hyper-credulity is defined as a disposition to accept uncritically all charismatically ordained beliefs. All lovers of literature and poetry are familiar with ‘that willing suspension of disbelief for the moment, which constitutes poetic faith’ (Coleridge 1970: 147). Hyper-credulity occurs when this state of mind, which in most of us is occasional and transitory, is transformed into a stable disposition. Hyper-credulity falls between hyper-suggestibility on the one hand and stable conversion of belief on the other.2 Its operational hallmark is plasticity in the assumption of deeply held convictions at the behest of an external authority. This is an other-directed form of what Robert Lifton (1968) has called the protean identity state.

D7. Relational Enmeshment is a state of being in which self-esteem  depends upon belonging to a particular collectivity (Bion 1959; Bowen 1972; Sirkin and Wynne 1990). It may be operationalized as immersion in a relational network with the following characteristics: exclusivity (high ratio of in-group to out-group bonds), interchangeability (low level of differentiation in affective ties between one alter and another), and dependency (reluctance to sever or weaken ties for any reason). In a developmental context, something similar to this has been referred to by Bowlby (1969) as anxious attachment.
D8. Exit Costs are the subjective costs experienced by an individual who is contemplating leaving a collectivity. Obviously, the higher the perceived exit costs, the greater will be the reluctance to leave. Exit costs may be operationalized as the magnitude of the bribe necessary to overcome them. A person who is willing to leave if we pay him $1,000 experiences lower exit costs than one who is not willing to leave for any payment less than $1,000,000. With regard to cults, the exit costs are most often spiritual and emotional rather than material, which makes measurement in this way more difficult but not impossible.


Not all charismatic organizations engage in brainwashing. We therefore need a set of hypotheses that will allow us to test empirically whether any particular charismatic system attempts to practise brainwashing and with what effect. The brainwashing model asserts twelve hypotheses concerning the role of brainwashing in the production of uncritical obedience. These hypotheses are all empirically testable. A schematic diagram of the model I propose may be found in Figure 1.

p. 186This model begins with an assumption that charismatic leaders are capable of creating organizations that are easy and attractive to enter (even though they may later turn out to be difficult and painful to leave). There are no hypotheses, therefore, to account for how charismatic cults obtain members. It is assumed that an abundant pool of potential recruits to such groups is always available. The model assumes charismatic leaders, using nothing more than their own intrinsic attractiveness and persuasiveness, are initially able to gather around them a corps of disciples sufficient for the creation of an attractive social movement. Many ethnographies (Lofland 1996; Lucas 1995) have shown how easy it is for such charismatic movement organizations to attract new members from the general pool of anomic ‘seekers’ that can always be found within the population of an urbanized mobile society.

Hieromonk Ephraim &amp; Kids

The model does attempt to account for how some percentage of these ordinary members are turned into deployable agents. The initial attractiveness of the group, its vision of the future, and/or its capacity to bestow seemingly limitless amounts of love and esteem on the new member are sufficient inducements in some cases to motivate a new member to voluntarily undergo this difficult and painful process of resocialization.

H1. Ideological totalism is a necessary but not sufficient condition for the brainwashing process. Brainwashing will be attempted only in groups that are structures totalistically. However, not all ideologically totalist groups will attempt to brainwash their members. It should be remembered that brainwashing is merely a mechanism for producing deployable agents. Some cults may not want deployable agents or have other ways of producing them. Others may want them but feel uncomfortable about using brainwashing methods to obtain them, or they may not have discovered the existence of brainwashing methods.

H2. The exact nature of this resocialization process will differ from group to group, but, in general, will be similar to the resocialization process that Robert Lifton (1989) and Edgar Schein (1961) observed in Communist re-education centres in the 1950s. For whatever reasons, these methods seem to come fairly intuitively to charismatic leaders and their staffs. Although the specific steps and their exact ordering differ from group to group, their common elements involve a stripping away of the vestiges of an old identity, the requirement that repeated confessions be made either orally or in writing, and a somewhat random and ultimately debilitating alternation of the giving and the withholding of ‘unconditional’ love and approval. H2 further states that the maintenance of this program involves the expenditure of a measurable quantity of the collectivity’s resources. This quantity is known as C, where C equals the cost of the program and should be measurable at least at an ordinal level.

The resocialization process has baffled many observers, in my opinion because it proceeds simultaneously along two distinct but parallel tracks, one involving cognitive functioning and the other involving emotional networking. These two tracks lead to the attainment of states of hyper-credulity and relational enmeshment, respectively. The group member learns to accept with suspended critical judgement the often shifting beliefs espoused by the charismatic leader. At the same time, the group member becomes strongly attached to and emotionally dependent upon the charismatic leader and (often especially) the other group members, and cannot bear to be shunned by them.

Hidden in the dark

H3. Those who go through the process will be more likely than those who do not to reach a state of hyper-credulity. This involves the shedding of old convictions and the assumption of a zealous loyalty to these beliefs of the moment, uncritically seized upon, so that all such beliefs become not mere ‘beliefs’ but deeply held convictions.

Under normal circumstances, it is not easy to get people to disown their core convictions. Convictions, once developed, are generally treated not as hypotheses to test empirically but as possessions to value and cherish. There are often substantial subjective costs to the individual in giving them up. Abelson (1986: 230) has provided convincing linguistic evidence that most people treat convictions more as valued possessions than as ways of testing reality. Cognitive dissonance theory predicts with accuracy that when subject to frontal attack, attachment to convictions tends to harden (Festinger, Riechen et. al. 1956; O’Leary 1994). Therefore, a frontal attack on convictions, without first undermining the self-image foundation of these convictions, is doomed to failure. An indirect approach through brainwashing is often more effective.

Scott Nevins.

When the state of hyper-credulity is achieved, it leaves the individual strongly committed to the charismatic belief of the moment but with little or no critical inclination to resist charismatically approved new or contradictory beliefs in the future and little motivation to attempt to form accurate independent judgments of the consequences of assuming new beliefs. The cognitive track of the resocialization process begins by stripping away the old convictions and associating them with guilt, evil, or befuddlement. Next, there is a traumatic exhaustion of the habit of subjecting right-brain convictions to left-brain rational scrutiny. This goes along with an increase in what Snyder (1974) has called self-monitoring, implying a shift from central route to peripheral route processing of information in which the source rather than the content of the message becomes all important.

H4. As an individual goes through the brainwashing process, there will be an increase in relational enmeshment with measurable increases occurring at the completion of each of the three stages. The purging of convictions is a painful process and it is reasonable to ask why anybody would go through it voluntarily. The payoff is the opportunity to feel more connected with the charismatic relational network. These people have also been through it, and only they really understand what you are going through. So cognitive purging leads one to seek relational comfort, and this confort becomes enmeshing. The credulity process and the enmeshing process depend on each other.

The next three hypotheses are concerned with the fact that each of the three phases of brainwashing achieves plateaus in both of these processes. The stripping phase creates the vulnerability to this sort of transformation. The identification phase creates realignment, and the rebirth phase breaks down the barrier between the two so that convictions can be emotionally energized and held with zeal, while emotional attachments can be sacralized in terms of the charismatic ideology. The full brainwashing model actually provides far more detailed hypotheses concerning the various steps within each phase of the process. Space constraints make it impossible to discuss these here. An adequate technical discussion of the manipulation of language in brainwashing, for example, would require a chapter at least the length of this one. Figure 2 provides a sketch of the steps within each phase. Readers desiring more information about these steps are referred to Lifton (1989: chap. 5).

P. 190 (scrprnt)
The Stages of Brainwashing & Their Effect on Hyper-credulity and Emotional Enmeshment

page 191

H5. The stripping phase. The cognitive goal of the stripping phase is to destroy prior convictions and prior relationships of belonging. The emotional goal of the stripping phase is to create the need for attachments. Overall, at the completion of the stripping phase, the situation is such that the individual is hungry for convictions and attachments and dependent upon the collectivity to supply them. This sort of credulity and attachment behaviour is widespread among prisoners and hospital patients (Goffman 1961).

H6. The identification phase. The cognitive goal of the identification phase is to establish imitative search for conviction and bring about the erosion of the habit of incredulity. The emotional goal of the identification phase is to instill the habit of acting out through attachment. Overall, at the completion of the identification phase of the individual has begun the practice of relying on the collectivity for beliefs and for a cyclic emotional pattern of arousal and comfort. But, at this point this reliance is just one highly valued form of existence. It is not yet viewed as an existential necessity.

H7. The symbolic death and rebirth phase. In the death and rebirth phase, the cognitive and emotional tracks come together and mutually support each other. This often gives the individual a sense of having emerged from a tunnel and an experience of spiritual rebirth.3 The cognitive goal of this phase is to establish a sense of ownership of (and pride of ownership in) the new convictions. The emotional goal is to make a full commitment to the new self that is no longer directly dependent upon hope of attachment or fear of separation. Overall, at the completion of the rebirth phase we may say that the person has become a fully deployable agent of the charismatic leader. The brainwashing process is complete.

H8 states that the brainwashing process results in a state of subjectivity elevated exit costs. These exit costs cannot, of course, be observed directly. But they can be inferred from the behavioural state of panic or terror that arises in the individual at the possibility of having his or her ties to the group discontinued. The cognitive and emotional states produced by the brainwashing process together bring about a situation in which the perceived exit costs for the individual increase sharply. This closes the trap for all but the most highly motivated individuals, and induces in many a state of uncritical obedience. As soon as exit from a group (or even from its good graces) ceases to be a subjectively palatable option, it makes sense for the individual to comply with almost anything the group demands–even to the point of suicide in some instances. Borrowing from Sartre’s insightful play of that name, I refer to this situation as the ‘no exit’ syndrome. When demands for compliance are particularly harsh, the hyper-credulity aspect of the process sweetens the pill somewhat by allowing the individual to accept uncritically the justifications offered by the charismatic leader and/or charismatic organization for making these demands, however far-fetched these justifications might appear to an outside observer.

H9 states that the brainwashing process results in a state of ideological obedience in which the individual has a strong tendency to comply with any behavioural demands made by the collectivity, especially if motivated by the carrot of approval and the stick of threatened expulsion, no matter how life-threatening these demands may be and no matter how repugnant such demands might have been to the individual in his or her pre-brainwashed state.

H10 states that the ‘brainwashing process results in increased deployability. Deployability extends the range of ideological obedience in the temporal dimension. It states that the response continues after the stimulus is removed. This hypothesis will be disconfirmed in any cult within which members are uncritically obedient only while they are being brainwashed but not thereafter. The effect need not be permanent, but it does need to result in some measurable increase in deployability over time.

H11 states that the ability of the collectivity to rely on obedience without surveillance will result in a measurable decrease in surveillance. Since surveillance involves costs, this decrease will lead to a quantity S, where S equals the savings to the collectivity due to diminished surveillance needs and should be measurable at least to an ordinal level.

H12 states that S will be greater than C. In other words, the savings to the collectivity due to decreased surveillance needs is greater than the cost of maintaining the brainwashing program. Only where S is greater than C does it make sense to maintain a brainwashing program. Cults with initially high surveillance costs, and therefore high potential savings due to decreased surveillance needs [S], will tend to be more likely to brainwash, as will cults structured so that the cost of maintaining the brainwashing system [C] are relatively low.

Holy Archangel Monks socializing over wine

Characteristics of a Good Theory

There is consensus in the social sciences that a good inductive qualitative theory is one that is falsifiable, internally consistent, concrete, potentially generalizable, and has a well-defined dependent variable (king, Keohane et. al. 1994). I think it should be clear from the foregoing that this theory meets all of these conditions according to prevailing standards in the social and behavioural sciences. However, since brainwashing theory has received much unjustified criticism for its lack of falsifiability and its lack of generalizability, I will briefly discuss the theory from these two points of view.

The criterion of falsifiability, as formulated primarily by Popper (1968), is the essence of what separates theory from dogma in science. Every theory must be able to provide an answer to the question of what evidence would falsify it. If the answer is that there is no possible evidence that would lead us to reject a so-called theory, we should conclude that it is not really a theory at all but just a piece of dogma.

Although Dawson (1998) and Richardson (1993) have included the falsifiability problem in their critiques of brainwashing; this criticism is associated mainly with the work of Dick Anthony (1996). Anthony’s claim that brainwashing theory is unfalsifiable is based upon  two related misunderstandings. First, he argues that it is impossible to prove that a person is acting with free will so, to the extent that brainwashing theory rests on the overthrow of free will, no evidence can ever disprove it. Second, he applies Popper’s criterion to cults in a way more appropriate for a highly developed deductive theoretical system. He requires that either brainwashing explain all ego-dystonic behaviour in cults or acknowledge that it can explain none of it. But, as we have seen, brainwashing is part of an inductive multifactorial approach to the study of obedience in cults and should be expected to explain only some of the obedience produced in some cults.

With regard to generalizability, cultic brainwashing is part of an important general class of phenomena whose common element is what Anthony Giddens has called ‘disturbance of ontological security’ in which habits and routines cease to function as guidelines for survival (Cohen 1989: 53). This class of phenomena includes the battered spouse syndrome (Barnett and LaViolette 1993), the behaviour of concentration camp inmates (Chodoff 1966), the Stockholm Syndrome (Kuleshnyk 1984; Powell 1986), and, most importantly, behaviour within prisoner of war camps and Communist Chinese re-education centres and ‘revolutionary universities’ (Lifton 1989; Sargant 1957;  Schein 1961). There exist striking homologies in observed responses across all of these types of events, and it is right that our attention be drawn to trying to understand what common theme underlies them all. As Oliver Wendell Holmes (1891: 325) attempted to teach us more than a century ago, the interest of the scientist should be guided, when applicable, by ‘the plain law of homology which declares that like must be compared with like.’

Cats of St. Nektarios


  1. Because of space limitations, I cannot give this important subject the attention it deserves in this chapter. Readers not familiar with the concept are referred to the much fuller discussion of this subject in the book by Robert Lifton as cited.
  2. Students of cults have sometimes been misled into confusing this state of hyper vrdulity with either hyper suggestibility on the one hand or a rigid ‘true belief’ system on the other. But at least one study has shown that neither the hyper-suggestible, easily hypnotized person nor the structural true believer are good candidates for encapsulation in a totalist cult system (Solomon 1981: 111-112). True believers (often fundamentalists who see in the cult a purer manifestation of their own worldview than they have seen before) do not do well in cults and neither do dye-in-the-wool sceptics who are comfortable with their scepticism. Rather it is those lacking convictions but hungering for them that are the best candidates.
  3. Hopefully, no reader will think that I am affirming the consequent by stating that all experiences of spiritual rebirth must be caused by brainwashing. This model is completely compatible with the assumption that most spiritual rebirth experiences have nothing to do with brainwashing. The reasoning here is identical to that connecting epilepsy with visions of the holy. The empirical finding that seizures can be accompanied by visions of the holy does not in any way imply that such visions are always a sign of epilepsy.

Cultural Contention over the Concept of Brainwashing (Benjamin Zablocki, 2001)

NOTE: The following article is taken from the 5th chapter of Misunderstanding Cults: Searching for Objectivity in a Controversial Field, entitled, Towards a Demystified and Disinterested Scientific Theory of Brainwashing.

That Word ‘Brainwashing’

The word brainwashing is, in itself, controversial and arouses hostile feelings. Since there is no scientific advantage in using one word rather than another for any concept, it may be reasonable in the future to hunt around for another word that is less polemical. We need a universally recognized term for a concept that stands for a form of influence manifested in a deliberately and systematically applied traumatizing and obedience-producing process of ideological resocialization.


Currently, brainwashing is the generally accepted term for this process, but I see no objection to finding another to take its place. There are in fact other terms, historically, that have been used instead, like ‘thought reform’ and ‘coercive persuasion.’ Ironically, it has been those scholars who complain the most about ‘the B-word’ who have also been the most insistent that none of the alternatives is any better. As long as others in the field insist on treating all possible substitute constructions as nothing more than gussied-up synonyms for a mystified concept of brainwashing (see, for example, Introvigne 1998: 2), there is no point as yet in trying to introduce a more congenial term.

An overly literal reading of the word brainwashing (merely a literal translation of the accepted Chinese term shi nao) could be misleading, as it seems to imply the ability to apply some mysterious biochemical cleanser to people’s brains. However, the word has never been intended as a literal designator but as a metaphor. It would be wise to heed Clifford Geertz’s (1973: 210) warning in this connection, to avoid such a ‘flattened view of other people’s mentalities [that] more complex meanings than [a] literal reading suggests [are] not even considered.’

Thus, please don’t allow yourself to become prejudiced by a visceral reaction to the word instead of attending to the underlying concept. There is a linguistic tendency, as the postmodernist critics have taught us, for the signified to disappear beneath the signifier. But the empirically based social sciences must resist this tendency by defining terms precisely. The influence of media-driven vulgarizations of concepts should be resisted. This chapter argues for the scientific validity of a concept, not a word. If you are interested in whether the concept has value, but you gag on the word, feel free to substitute a different word in its place. I myself have no particular attachment to the word brainwashing.

But if all we are talking about is an extreme form of influence, why do we need a special name for it at all? The name is assigned merely for convenience. This is a common and widely accepted practise in the social sciences. For example, in economics a recession is nothing more than a name we give to two consecutive quarters of economic contraction. There is nothing qualitatively distinctive about two such consecutive quarters as opposed to one or three. The label is assigned arbitrarily at a subjective point at which many economists begin to get seriously worried about economic performance. This label is nevertheless useful as long as we don’t reify it by imagining that it stands for some real ‘thing’ that happens to the economy when it experiences precisely two quarters of decline. Many other examples of useful definitions marking arbitrary points along a continuum could be cited. There is no objective way to determine the exact point at which ideological influence becomes severe and encompassing enough, and its effects long lasting enough, for it to be called brainwashing. Inevitably, there will be marginal instances that could be categorized either way. But despite the fact that the boundary is not precisely defined, it demarcates a class of events worthy of systematic study.

The Reciprocal Moral Panic

Study of brainwashing has been hampered by partisanship and tendentious writing on both sides of the conflict. In one camp, there are scholars who very badly don’t want there to be such a thing as brainwashing. Its non-existence, they believe, will help assure religious  liberty, which can only be procured by defending the liberty of the most unpopular religions. If only the non-existence of brainwashing can be proved, the public will have to face up to the hard truth that some citizens choose to follow spiritual paths that may lead them in radical directions. This camp has exerted its influence within academia. But, instead of using its academic skills to refute the  brainwashing conjecture, it has preferred to attack a caricature of brainwashing supplied by anti-cult groups for litigational rather than scientific purposes.

Ecological Fallacy

In the other camp, we find scholars who equally badly do want there to be such a thing as brainwashing. Its existence, they believe, will give them a rationale for opposition to groups they consider dangerous. A typical example of their reasoning can be found in the argument put forth by Margaret Singer that ‘Despite the myth that normal people don’t get sucked into cults, it has become clear over the years that everyone is susceptible to the lure of these master manipulators’ (Singer 1995: 17). Using a form of backward reasoning known as the ecological fallacy, she argues from the known fact that people of all ages, social classes, and ethnic backgrounds can be found in cults to the dubious conclusion that everyone must be susceptible. These scholars must also share some of the blame for tendentious scholarship. Lacking positions of leadership in academia, scholars on this side of the dispute have used their expertise to influence the mass media, and they have been successful because sensational allegations of mystical manipulative influence make good journalistic copy.

It’s funny in a dreary sort of way that both sides in this debate agree that it is a David and Goliath situation, but each side fancies itself to be the David courageously confronting  the awesome power of the opposition. Each side makes use of an exaggerated fear of the other’s influence to create the raw materials of a moral panic (Cohen 1972; Goode and Ben Yehudah 1994). Thus, a disinterested search for truth falls victim to the uncompromising hostility created by each side’s paranoid fear of the power of the other.

David with the Head of Goliath
David with the head of Goliath.

The ‘cult apologists’ picture themselves as fighting an underdog battle against hostile lords of the media backed by their armies of ‘cult-bashing’ experts. The ‘cult bashers’ picture themselves as fighting an underdog battle for a voice in academia in which apologists seem to hold all the gatekeeper positions. Each side justifies its rhetorical excesses and hyperbole by reference to the overwhelming advantages held by the opposing side within its own arena. But over the years a peculiar symbiosis has developed between these two camps. They have come to rely on each other to define their positions. Each finds it more convenient to attack the positions of the other than to do the hard work of finding out what is really going on in cults. Thomas Robbins (19888: 74) has noted that the proponents of these two models ‘tend to talk past each other since they employ differing interpretative frameworks, epistemological rules, definitions… and underlying assumptions.’ Most of the literature on the subject has been framed in terms of rhetorical disputes between these two extremist models. Data-based models have been all but crowded out.

Between these two noisy and contentious camps, we find the curious but disinterested scientist who wants to find out if there is such a thing as brainwashing but will be equally satisfied with a positive or negative answer. I believe that there can and should be a moderate position on the subject. Such a position would avoid the absurdity of denying any reality to what thousands of reputable ex-cult members claim to have experienced–turning this denial into a minor cousin of holocaust denial. At the same time, it would avoid the mystical concept of an irresistible and overwhelming force that was developed by the extremist wing of the anti-cult movement.

One of the most shameful aspects of this whole silly affair is the way pro-religion scholars have used their academic authority to foist off the myth that the concept of brainwashing needs no further research because it has already been thoroughly debunked. Misleadingly, it has been argued (Introvigne forthcoming; Melton forthcoming) that the disciplines of psychology and sociology, through their American scholarly associations, have officially declared the concept of brainwashing to be so thoroughly discredited that no further research is needed. Introvigne, by playing fast and loose with terminology, attempts to parlay a rejection of a committee report into a rejection of the brainwashing concept by the American Psychological Association. He argues that ‘To state that a report “lacks scientific rigor” is tantamount to saying that it is not scientific’ (Introvigne 1998: 3), gliding over the question of whether the ‘it’ in question refers to the committee report or the brainwashing concept.2 Conveniently, for Introvigne, the report in question was written by a committee chaired by Margaret Singer, whose involuntarist theory of brainwashing is as much a distortion of the foundational concept as Introvigne’s parody of it.

The truth is that both of these scholarly associations (American Psychological Association and American Sociological Association) were under intense pressure by a consortium of pro-religious scholars (a.k.a. NRM scholars) to sign an amicus curiae brief alleging consensus within their fields that brainwashing theory had been found to be bunk. This was in regard to a case concerning Moonie brainwashing that was before the United States Supreme Court (Molko v Holly Spirit Ass’n., Supreme Court of Calif. SF 25038; Molko v Holly Spirit Ass’n, 762 p.2d 46 [Cal. 1988], cert. Denied, 490 U.S. 1084 [1989]). The bottom line is that both of the associations, after bitter debate, recognized that there was no such consensus and refused to get involved. Despite strenuous efforts of the NRM scholars to make it appear otherwise, neither professional association saw an overwhelming preponderance of evidence on either side. Both went on the record with a statement virtually identical to my argument in this chapter: that not nearly enough is known about this subject to be able to render a definitive scientific verdict, and that much more research is needed. A few years later, the Society for the Scientific Study of Religion went on record with a similar statement, affirming ‘the agnostic position’ on this subject and calling for more research (Zablocki 1997: 114).

Although NRM scholars have claimed to be opposed only to the most outrageously sensationalized versions of brainwashing theory, the result, perhaps unintended, of their campaign has been to bring an entire important area of social inquiry to a lengthy halt. Evidence of this can be seen in the fact that during the period of 1962 to 2000, a time when cults flourished, not a single article supportive of brainwashing has been published in the two leading American journals devoted to the sociology of religion, although a significant number of such articles have been submitted to those journals and more than a hundred such articles have appeared in journals marginal to the field (Zablocki 1998: 267)

Crime Scene Photo of Heaven's Gate Bodies Found in Rancho Santa Fe, CA (1)
Crime Scene Photo of Heaven’s Gate Bodies Found in Rancho Santa Fe, CA.


The erroneous contention that brainwashing theory has been debunked by social science research has been loudly and frequently repeated, and this ‘big lie’ has thus come to influence the thinking of neutral religion scholars. For example, even Winston Davis, in an excellent article on suicidal obedience in Heaven’s Gate, expresses characteristic ambivalence over the brainwashing concept:

‘Scholarship in general no longer accepts the traditional, simplistic theory of brainwashing… While the vernacular theory of brainwashing may no longer be scientifically viable, the general theory of social and psychological conditioning is still rather in good shape… I therefore find nothing objectionable [sic] in Benjamin Zablocki’s revised theory of brainwashing as ‘a set of transactions between a charismatically led collectivity and an isolated agent of the collectivity with the goal of transforming the agent into a deployable agent.’ The tale I have to tell actually fits nicely into several of Robert Lifton’s classical thought reform categories (Davis 2000: 241-2).

The problem with this all too typical way of looking at things is the fact that I am not presenting some new revised theory of brainwashing but simply a restatement of Robert Lifton’s (1989, 1999) careful and rigorous theory in sociological terms.

There are, I believe, six issues standing in the way of our ability to transcend this reciprocal moral panic. Let us look closely at each of these issues with an eye to recognizing that both sides in this conflict may have distorted the scientifically grounded theories of the foundational theorists–Lifton (1989), Sargant (1957), and Schein (1961)– as they apply to cults.

The Influence Continuum

The first issue has to do with the contention that brainwashing is a newly discovered form of social influence involving a hitherto unknown social force. There is nothing about charismatic influence and the obedience it instills that is mysterious or asks us to posit the existence of a new force. On the contrary, everything about brainwashing can be explained entirely in terms of well-understood scientific principles. As Richard Ofshe has argued: ‘Studying the reform process demonstrates that it is no more or less difficult to understand than any other complex social process and produces no results to suggest that something new has been discovered. The only aspect of the reform process that one might suggest is new, is the order in which the influence procedures are assembled and the degree to which the target’s environment is manipulated in the service of social control. This is at most an unusual arrangement of commonplace bits and pieces’ (1992: 221-2).

Would-be debunkers of the brainwashing concept have argued that brainwashing theory is not just a theory of ordinary social influence intensified under structural conditions of ideological totalism, but is rather a ‘special’ kind of influence theory that alleges that free will can be overwhelmed and individuals brought to a state of mind in which they will comply with charismatic directives involuntarily, having surrendered the capability of saying no. Of course, if a theory of brainwashing really did rely upon such an intrinsically untestable notion, it would be reasonable to reject it outright.

The attack on this so-called involuntarist theory of brainwashing figures prominently in the debunking efforts of a number of scholars (Barker 1989; Hexham and Poewe 1997; Melton forthcoming), but is most closely identified with the work of Dick Anthony (1996), for whom it is the linchpin of the debunking argument. Anthony argues, without a shred of evidence that I have been able to discover, that the foundational work of Lifton and Schein and the more recent theories of myself (1998), Richard Ofshe (1992), and Stephen Kent (Kent and Krebs 1998) are based upon what he calls the ‘involuntarism assumption.’ It is true that a number of prominent legal cases have hinged on the question of whether the plaintiff’s free will had been somehow overthrown (Richardson and Ginsburg 1998). But nowhere in the scientific literature has there been such a claim. Foundational brainwashing theory has not claimed that subjects were robbed of their free will. Neither the presence nor the absence of free will can ever be proved or disproved. The confusion stems from the difference between the word free as it is used in economics as an antonym for costly, and as it is used in philosophy as an antonym for deterministic. When brainwashing theory speaks of individuals losing the ability to freely decide to obey, the word is being used in the economic sense. Brainwashing imposes costs, and when a course of action has costs it is no longer free. The famous statement by Rousseau (1913, p.3) that ‘Man is born free, and everywhere he is in chains,’ succinctly expresses the view that socialization can impose severe constraints on human behaviour. Throughout the social sciences, this is accepted almost axiomatically. It is odd that only in the sociology of new religious movements is the importance of socialization’s ability to constrain largely ignored.    

Geronda Ephraim (AZ)      

Unidirectional versus Bi-directional Influence

The second issue has to do with controversy over whether there are particular personality types drawn to cults and whether members are better perceived as willing and active seekers or as helpless and victimized dupes, as if these were mutually exclusive alternatives. Those who focus on the importance of the particular traits that recruits bring to their cults tend to ignore the resocialization process (Anthony and Robbins 1994).3 Those who focus on the resocialization process often ignore personal predispositions (Singer and Ofshe 1990).

All this reminds me of being back in high school when people used to gossip about girls who ‘got themselves pregnant.’ Since that time, advances in biological theory have taught us to think more realistically of ‘getting pregnant’ as an interactive process involving influence in both directions. Similarly, as our understanding of totalistic influence in cults matures, I think we will abandon undirectional explanations of cultic obedience in favour of more realistic, interactive ones. When that happens, we will find ourselves able to ask more interesting questions than we do now. Rather than asking whether it is the predisposing trait or a manipulative process that produces high levels of uncritical obedience, we will ask just what predisposing traits of individuals interact with just what manipulative actions by cults to produce this outcome.

A number of the debunking authors use this artificial and incorrect split between resocialization and predisposing traits to create a divide between cult brainwashing theory and foundational brainwashing theory as an explanation for ideological influence in China and Korea in the mid-twentieth century. Dick Anthony attempts to show that the foundational literature really embodied two distinct theories. One, he claims, was a robotic control theory that was mystical and sensationalist. The other was a theory of totalitarian influence that was dependent for its success upon pre-existing totalitarian beliefs of the subject which the program was able to reinvoke (Anthony 1996: i). Anthony claims that even though cultic brainwashing theory is descendant from the former, it claims its legitimacy from its ties to the latter.

The problem with this distinction is that it is based upon a misreading of the foundational literature (Lifton1989; Schein 1961). Lifton devotes chapter 5 of his book to a description of the brainwashing process. In chapter 22 he describes the social structural conditions that have to be present for this process to be effective. Anthony misunderstands this scientific distinction. He interprets it instead as evidence that Lifton’s work embodies two distinct theories: one bad and one good (Anthony and Robbins 1994). The ‘bad’ Lifton, according to Anthony, is the chapter 5 Lifton who describes a brainwashing process that may have gone on in  Communist reindoctrination centres, but which, according to Anthony, has no applicability to contemporary cults. The ‘good’ Lifton, on the other hand, describes in chapter 22 a structural situation that Anthony splits off and calls a theory of thought reform. Anthony appears to like this ‘theory’ better because it does not involve anything that the cult actually does to the cult participant (Anthony and Robbins 1995). The cult merely creates a totalistic social structure that individuals with certain predisposing traits may decide that they want to be part of.

Unfortunately for Anthony, there are two problems with such splitting. One is that Lifton himself denies any such split in his theory (Lifton 1995, 1997). The second is that both  an influence process and the structural conditions conducive to that process are necessary for any theory of social influence. As Lifton demonstrates in his recent application of his theory to a Japanese terrorist cult (Lifton 1999), process cannot be split off from structure in any study of social influence.

Geronda Ephraim, Monks, Devotees

Condemnatory Label versus Contributory Factor

The third issue has to do with whether brainwashing is meant to replace other explanatory variables or work alongside them. Bainbridge (1997) and Richardson (1993) worry about the former, complaining that brainwashing explanations are intrinsically unifactoral, and thus inferior to the multifactoral explanations preferred by modern social science. But brainwashing theory has rarely, if ever, been used scientifically as a unifactoral explanation. Lifton (1999) does not attempt to explain all the obedience generated in Aum Shinrikyo by the brainwashing mechanism. My explanation of the obedience generated by the Nruderhof relies on numerous social mechanisms of which brainwashing is only one (Zablocki 1980). The same can be said for Ofshe’s explanation of social control in Synanon (1976). Far from being unifactoral, brainwashing is merely one essential element in a larger strategy for understanding how charismatic authority is channelled into obedience.

James Thurber once wrote a fable called The Wonderful (1957), which depicted the cultural collapse of a society that was free to express itself using twenty-five letters of the alphabet but was forbidden to use the letter O for any reason. The intellectual convolutions forced on Thurber’s imaginary society by this ‘slight’ restriction are reminiscent of the intellectual  convolutions forced on the NRM scholars by their refusal to include brainwashing in their models. It is not that these scholars don’t often have considerable insight into cult dynamics, but the poor mugs are, nevertheless, constantly getting overwhelmed by events that their theories are unable to predict or explain. You always find them busy playing catch-up as they scramble to account for each new cult crisis as it develops on an ad hoc basis. The inadequacy of their models cries out ‘specification error’ in the sense that a key variable has been left out.

The Thurberian approach just does not work. We have to use the whole alphabet of social influence concepts from Asch to Zimbardo (including the dreaded B-word) to understand cultic obedience. Cults are a complex social ecology of forces involving attenuation effects (Petty 1994), conformity (Asch 1951), crowd behaviour (Coleman 1990), decision elites (Wexler 1995), deindividuation (Festinger, Pepitone et. al. 1952), extended exchange (Stark 1999), groupthink (Janis 1982), ritual (Turner (1969), sacrifice and stigma (Iannaccone 1992), situational pressures (Zimbardo and Anderson 1993), social proof (Cialdini 1993), totalism (Lifton 1989), and many others. Personally, I have never seen a cult that was held together only by brainwashing and not also by other psychological factors, as well as genuine loyalty to ideology and leadership.

Arguments that brainwashing is really a term of moral condemnation masquerading as a scientific concept have emerged as a reaction to the efforts of some anti-cultists (not social scientists) to use brainwashing as a label to condemn cults rather than as a concept to understand them. Bromley (1998) has taken the position that brainwashing is not a variable at all but merely a peremptory label of stigmatization–a trope for an ideological bias, in our individualistic culture, against people who prefer to live and work more collectivistically. Others have focused on the observe danger of allowing brainwashing to be used as an all-purpose moral excuse (It wasn’t my fault. I was brainwashed!), offering blanket absolution for people who have been cult members–freeing them from the need to take any responsibility for their actions (Bainbridge 1997; Hexham and Poewe 1997; Introvigne forthcoming; Melton forthcoming). While these allegations represent legitimate concerns about potential abuse of the concept, neither is relevant to the scientific issue. A disinterested approach will first determine whether a phenomenon exists before worrying about whether its existence is politically convenient.

Geronda Ephraim baptism

Obtaining Members versus Retaining Members

The fourth issue has to do with a confusion over whether brainwashing explains how cults obtain members or how they retain them. Some cults have made use of manipulative practices like love-bombing and sleep deprivation (Galanti 1993), with some degrees of success, in order to obtain new members. A discussion of these manipulative practices for obtaining members is beyond the scope of this chapter. Some of these practices superficially resemble techniques used in the earliest phase of brainwashing. But these practices, themselves, are not brainwashing. This point must be emphasized because a false attribution of brainwashing to newly obtained cult recruits, rather than to those who have already made a substantial commitment to the cult, figures prominently in the ridicule of the concept by NRM scholars. A typical straw man representation of brainwashing as a self-evidently absurd concept is as follows: ‘The new convert is held mentally captive in a state of alternate consciousness due to “trance-induction techniques” such as meditation, chanting, speaking in tongues, self-hypnosis, visualization, and controlled breathing exercises … the cultist is [thus] reduced to performing religious duties in slavish obedience to the whims of the group and its authoritarian or maniacal leader’ (Wright 1998: 98).

Foundational brainwashing theory was not concerned with such Svengalian conceits, but only with ideological influence in the service of the retaining function. Why should the foundational theorists, concerned as they were with coercive state-run institutions like prisons, ‘re-education centres,’ and prisoner-of-war camps have any interest in explaining how participants were obtained? Participants were obtained at the point of a gun.4 The motive of these state enterprises was to retain the loyalties of these participants after intensive resocialization ceased. As George Orwell showed so well in his novel 1984, the only justification for the costly indoctrination process undergone by Winston Smith was not that he love Big Brother while Smith was in prison, but that Big Brother be able to retain that love after Smith was deployed back into society. Nevertheless, both ‘cult apologists’ and ‘cult bashers’ have found it more convenient to focus on the obtaining function.

Geronda Ephraim with child

If one asks why a cult would be motivated to invest resources in brainwashing, it should be clear that this can not be to obtain recruits, since these are a dime a dozen in the first place, and, as Barker (1984) has shown, they don’t tend to stick around long enough to repay the investment. Rather, it can only be to retain loyalty, and therefore decrease surveillance costs for valued members who are already committed. In small groups bound together only by normative solidarity, as Hechter (1987) has shown, the cost of surveillance of the individual by the group is one of the chief obstacles to success. Minimizing these surveillance costs is often the most important organizational problem such groups have to solve in order to survive and prosper. Brainwashing makes sense for a collectivity only to the extent that the resources saved through decreased surveillance costs exceed the resources invested in the brainwashing process. For this reason, only high-demand charismatic groups with totalistic social structures are ever in a position to benefit from brainwashing.5

This mistaken ascription of brainwashing to the obtaining to the obtaining function rather than the retaining function is directly responsible for two of the major arguments used by the ‘cult apologists’ in their attempt to debunk brainwashing. One has to do with a misunderstanding of the role of force and the other has to do with the mistaken belief that brainwashing can be studied with data on cult membership turnover.

The widespread belief that force is necessary for brainwashing is based upon a misreading of Lifton (1989) and Schein (1961). A number of authors (Dawson 1998; Melton forthcoming; Richardson 1993) have based their arguments, in part, on the contention that the works of foundational scholarship on brainwashing are irrelevant to the study of cults because the foundational literature studied only subjects who were forcibly incarcerated. However, Lifton and Schein have both gone on public record as explicitly denying that there is anything about their theories that requires the use of physical force or threat of force. Lifton has specifically argued (‘psychological manipulation is the heart of the matter, with or without the use of physical force’ [1995: xi]) that his theories are very much applicable to cults.6 The difference between the state-run institutions that Lifton and Schein studied in the 1950s and 1960s and the cults that Lifton and others study today is in the obtaining function not in the retaining function. In the Chinese and Korean situations, force was used for obtaining and brainwashing was used for retaining. In cults, charismatic appeal is used for obtaining and brainwashing is used, in some instances, for retaining.

A related misconception has to do with what conclusions to draw from the  very high rate of turnover among new and prospective recruits to cults. Bainbridge (1997), Barker (1989), Dawson (1998), Introvigne (forthcoming), and Richardson (1993) have correctly pointed out that in totalistic religious organizations very few prospective members go on to become long-term members. They argue that this proves that the resocialization process cannot be irresistible and therefore it cannot be brainwashing. But nothing in the brainwashing model predicts that it will be attempted with all members, let alone successfully attempted. In fact, the efficiency of brainwashing, operationalized as the expected yield of deployable agents7  per 100 members, is an unknown (but discoverable) parameter of any particular cultic system and may often be quite low. For the system to be able to perpetuate itself (Hechter 1987), the yield need only produce enough value for the system to compensate it for the resources required to maintain the brainwashing process.

Moreover, the high turnover rate in cults is more complex than it may seem. While it is true that the membership turnover is very high among recruits and new members, this changes after two or three years of membership when cultic commitment mechanisms begin to kick in. this transition from high to low membership turnover is known as the Bainbridge Shift, after the sociologist who first discovered it (Bainbridge 1997: 141-3). After about three years of membership, the annual rate of turnover sharply declines and begins to fit a commitment model rather than a random model.8

Membership turnover data is not the right sort of data to tell us whether a particular cult practises brainwashing. The recruitment strategy whereby many are called but few are chosen is a popular one among cults. In several groups in which I have observed  the brainwashing process, there was very high turnover among initial recruits. Brainwashing is too expensive to waste on raw recruits. Since brainwashing is a costly process, it generally will not pay for a group to even attempt to brainwash one of its members until that member has already demonstrated some degree of staying power on her own.9

Geronda Ephraim &amp; Geronda Paisios

Psychological Traces

The fifth issue has to do with the question of whether brainwashing leaves any long-lasting measurable psychological traces in those who have experienced it. Before we can ask this question in a systematic way, we have to be clear about what sort of traces we should be looking for. There is an extensive literature on cults and mental health. But whether cult involvement causes psychological problems is a much more general question than whether participation in a traumatic resocialization process leaves any measurable psychological traces.

There has been little consensus on what sort of traces to look for. Richardson and Kilbourne (1983: 30) assume that brainwashing should lead to insanity. Lewis (1983: 30) argues that brainwashing should lead to diminished IQ scores. Nothing in brainwashing theory would lead us to predict either of these outcomes. In fact, Schein points out that ‘The essence of coercive persuasion is to produce ideological and behavioral change in a fully conscious, mentally intact individual’ (1959: 437). Why in the world would brainwashers invest scarce resources to produce insanity and stupidity in their followers? However, these aforementioned authors (and others) have taken the absence of these debilitative effects as ‘proof’ that brainwashing doesn’t happen in cults. At the same time, those who oppose cults have had an interest, driven by litigation rather than science, in making exaggerated claims for mental impairment directly resulting from brainwashing. As Farrell has pointed out, ‘From the beginning, the idea of traumatic neurosis has been accompanied by concerns about compensation’ (1998: 7).

Studies of lingering emotional, cognitive, and physiological effects on ex-members have thus far shown inconsistent results (Katchen 1997; Solomon 1981; Ungerleider and Wellisch 1983). Researchers studying current members of religious groups have found no significant impairment or disorientation. Such results have erroneously been taken as evidence that the members of these groups could, therefore, not possibly have been brainwashed. However, these same researchers found these responses of current members contaminated by elevations on the ‘Lie’ scale, exemplifying ‘an intentional attempt to make a good impression and deny faults’ (Ungerleider and Wellisch 1983: 208). On the other hand, studies of ex-members have tended to show ‘serious mental and emotional dysfunctions that have been directly caused by cultic beliefs and practices (Saliba 1993: 106). The sampling methods of these latter studies have been challenged (Lewis and Bromley 1987; Solomon 1981), however, because they have tended to significantly over-sample respondents with anti-cult movement ties. With ingenious logic, this has led Dawson (1998: 121) to suggest in the same breath that cult brainwashing is a myth but that ex-member impairment may be a result of brainwashing done by deprogrammers.

All this controversy is not entirely relevant to our question, however, because there is no reason to assume that a brainwashed person is going to show elevated scores on standard psychiatric distress scales. In fact, for those for whom making choices is stressful, brainwashing may offer psychological relief. Galanter’s research has demonstrated that a cult ‘acts like a psychological pincer, promoting distress while, at the same time, providing relief’ (1989: 93). As we shall see below, the brainwashing model predicts impairment and disorientation only for people during some of the intermediate stages, not at the end state. The popular association of brainwashing with zombie or robot states comes out of a misattribution of the characteristics of people going through the traumatic brainwashing process to people going through the traumatic brainwashing process to people who have completed the process. The former really are, at times, so disoriented that they appear to resemble caricatures of zombies or robots. The glassy eyes, inability to complete sentences, and fixed eerie smiles are characteristics of disoriented people under randomly varying levels of psychological stress. The latter, however, are, if the process was successful, functioning and presentable deployable agents.

geron Efraim1

Establishing causal direction in the association between cult membership and mental health is extremely tricky, and little progress has been made thus far. In an excellent article reviewing the extensive literature in this area, Saliba (1993: 108) concludes: ‘The study of the relationship between new religious movements and mental health is in its infancy.’ Writing five years later, Dawson (1998: 122) agrees that this is still true, and argues that ‘the inconclusiveness results of the psychological study of members and ex-members of NRMs cannot conceivably be used to support either the case for or against brainwashing.’ Saliba calls for prospective studies that will establish baseline mental health measurements for individuals before they join cults, followed by repeated measures during and afterward. While this is methodologically sensible, it is impractical because joining a cult is both a rare and unexpected event. This makes the general question of how cults affect mental health very difficult to answer.

Fortunately, examining the specific issue of whether brainwashing leaves psychological traces may be easier. The key is recognizing that brainwashing is a traumatic process, and, therefore, those who have gone through it should experience an increasing likelihood in later years of post-traumatic stress disorder. The classic clinical symptoms of PTSD — avoidance, numbing, and increased arousal (American Psychiatric Association 1994: 427) — have been observed in many ex-cult members regardless of their mode of exit and current movement affiliations (Katchen 1997; Zablocki 1999). However, these soft and somewhat subjective symptoms should be viewed with some caution given recent controversies over the ease with which symptoms such as these can be iatrogenically implanted, as, for example, false memories (Loftus and Ketcham 1994).

In the future, avenues for more precise neurological tracking may become available. Judith Herman (1997: 238) has demonstrated convincingly that ‘traumatic exposure can produce lasting alterations in the endocrine, autonomic, and central nervous systems … and un the function and even the structure of specific areas of the brain.’ It is possible in the future that direct evidence of brainwashing may emerge from brain scanning using positron emission tomography. Some preliminary research in this area has suggested that, during flashbacks, specific areas of the brain involved with language and communication may be inactivated (Herman 1997: 240; Rauch van der Kolk, et. al. 1996). Another promising area of investigation of this sort would involve testing for what van der Kolk and McFarlene (1996) have clinically identified as ‘the black hole of trauma.’ It should be possible to determine, once measures have been validated, whether such traces appear more often in individuals who claim to have gone through brainwashing than in a sample of controls who have been non-brainwashed members of cults for equivalent periods of time.

elder ephraim pascha

Separating the Investigative Steps

The final issue is a procedural one. There are four sequential investigative steps required to resolve controversies like the one we have been discussing.these steps are concerned with attempt, existence, incidence, and consequence. A great deal of confusion comes from nothing more than a failure to recognize that these four steps need to be kept analytically distinct from one another.

To appreciate the importance of this point, apart from the heat of controversy, let us alter the scene for a moment and imagine that the scientific conflict we are trying to resolve is over something relatively innocuous — say, vegetarianism. Let us imagine that on one side we have a community of scholars arguing that vegetarianism is a myth, that nobody would voluntarily choose to live without eating meat and that anyone who tried would quickly succumb to an overpowering carnivorous urge. On the other side, we have another group of scholars arguing that they had actually seen vegetarians and observed their non-meat-eating behavior over long periods of time, and that, moreover, vegetarianism is a rapidly growing social problem with many new converts each year being seduced by this enervating and debilitating diet.

It should be clear that any attempt to resolve this debate scientifically would have to proceed through the four sequential steps mentioned above. First, we would have to find out if anybody ever deliberately attempts to be a vegetarian. Maybe those observed not eating meat were simply unable to obtain it. If nobody could be found voluntarily attempting to follow a vegetarian diet, we would next have to observe him carefully enough and long enough to find out whether he succeeds in abstaining from meat. If we observe even one person successfully abstaining from meat, we would have to conclude that vegetarianism exists, increasing our confidence in the theory of the second group of researchers. But the first group could still argue, well, maybe you are right that a few eccentric people here and there do practise vegetarianism, but not enough to constitute a social phenomenon worth investigating. So, the next step would be to measure the incidence of vegetarianism in the population. Out of every million people, how many do we find following a vegetarian diet? If it turns out to be very few, we can conclude that, while vegetarianism may exist as a social oddity, it does not rise to the level of being a social phenomenon worthy of our interest. If, however, we find a sizeable number of vegetarians, we still need to ask, ‘So what?’ This is the fourth of our sequential steps. Does the practice of vegetarianism have any physical, psychological, or social consequences? If so, are these consequences worthy of our concern?

Each of these investigative steps requires attention focused on quite distinct sets of substantive evidence. For this reason, it is important that we not confuse them with one another as is so often done in ‘apologist’ writing about brainwashing, where the argument often seems to run as follows: Brainwashing doesn’t exist, or at least it shouldn’t exist, and even if it does the numbers involved are so few, and everybody in modern society gets brainwashed  to some extent, and the effects, if any, are impossible to measure. Such arguments jump around, not holding still long enough to allow for orderly and systematic confirmation or disconfirmation of each of the steps.

Once we recognize the importance of keeping the investigative steps methodologically distinct distinct from one another, it becomes apparent that the study of brainwashing is no more problematic (although undoubtedly much more difficult) than the study of an advertising campaign for a new household detergent. It is a straightforward question to ask whether or not some charismatic groups attempt to practise radical techniques of socialization designed to turn members into deployable agents. If the answer is no, we stop because there can be no brainwashing. If the answer is yes, we go on to a second question: Are these techniques at least sometimes effective in producing uncritical obedience? If the answer to this question is ye (even for a single person), we know that brainwashing exists, although it may be so rare as to be nothing more than a sociological oddity. therefore, we have to take a third step and ask. How frequently is it effective? What proportion of those who live in cults are subjected to brainwashing, and what proportion of these respond by becoming uncritically obedient? And, finally, we need to ask a fourth important question: How long do the effects last? Are the effects transitory, lasting only as long as the stimulus continues to be applied, or are they persistent for a period of time thereafter, and, if so, how long? Let us keep in mind the importance of distinguishing attempt from existence, from incidence, from consequences.

To be continued…



  1. When I speak of ego dystonic behaviour, I refer to behaviour that was ego dystonic to the person before joining the cult and after leaving the cult.
  2. I have no doubt that Introvigne, who is a European attorney, is sincere in his desire to stifle brainwashing research out of fear that any suggestion that brainwashing might possibly occur in cults will be seized on by semi-authoritarian government committees eager to suppress religious liberty. Personally, I applaud Introvigne’s efforts to protect the fragile tree of religious freedom of choice in the newly emerging democracies of Eastern Europe. But I don’t appreciate his doing so by (perhaps inadvertently) sticking his  thumb on the scales upon which social scientists attempt to weigh evidence.
  3. The Anthony and Robbins article cited demonstrates how little we really know about traits that may predispose people to join cults. They say ‘…some traditionally conservative religious groups attract people who score highly on various measures of totalitarianism, e.g., the F scale or Rokeach’s Dogmatism scale… It seems likely that these results upon certain Christian groups would generalize to alternative religious movements or cults, as many of them have theological and social beliefs that seem similar to those in some fundamentalist denominations’ (1994:470).. Perhaps, but perhaps not. No consensus has yet emerged from numerous attempts to find a cult personality type, but this seems like a promising area of research to continue.
  4. Some, it is true, were nominally volunteers into re-education programs. However, the power of the state to make their lives miserable if they did not volunteer cannot be ignored.
  5. Unfortunately, however, uncritical obedience can be wayward and dangerous. It can be useful to a cult leader when the cult is functioning well. But it often has been perverted to serve a destructive or self-destructive agenda in cults that have begun to disintegrate.
  6. Some confusion on this subject has emerged from the fact that Lifton has distanced himself from those attempting to litigate against cults because of alleged brainwashing. He has constantly argued (and I wholeheartedly agree) that brainwashing, in and of itself, where no force is involved, should not be a matter for the law courts.
  7. Formal definitions for this and other technical terms will be presented in the next section of this chapter.
  8. In other words, the probability of a person’s leaving is inversely dependent upon the amount of time he or she has already spent as a member.
  9. The ‘cult-basher’ version of brainwashing theory has played into this misunderstanding by confounding manipulative recruitment techniques (like sleep deprivation and ‘love-bombing’) with actual brainwashing. While there may be some overlap in the actual techniques used, the former is a method for obtaining new members, whereas brainwashing is a method for retaining old members.

Do We Need to Know Whether Cults Engage in Brainwashing? (Benjamin Zablocki, 2001)

NOTE: The following article is taken from the 5th chapter of Misunderstanding Cults: Searching for Objectivity in a Controversial Field, entitled, Towards a Demystified and Disinterested Scientific Theory of Brainwashing.


Nobody likes to lose a customer, but religions get more touchy than most when faced with the risk of losing devotees they have come to define as their own. Historically, many religions have gone to great lengths to prevent apostasy, believing virtually any means justified to prevent wavering parishioners from defecting and thus losing hope of eternal salvation. In recent centuries, religion in our society has evolved from a system of territorially based near-monopolies into a vigorous and highly competitive faith marketplace in which many churches, denominations, sects, and cults vie with one another for the allegiance of ‘customers’ who are free to pick and choose among competing faiths. Under such circumstances, we should expect to find that some of the more tight-knit and fanatical religions in this rough-and-tumble marketplace will have developed sophisticated persuasive techniques are known in the literature by the controversial term ‘brainwashing.’ This chapter is devoted to a search for a scientific definition of brainwashing and an examination of the evidence for the existence of brainwashing in cults. I believe that research on this neglected subject is important for a fuller understanding of religious market dynamics.1 And, ultimately, research on this subject may yield a wider dividend as well, assisting us in our quest for a fuller understanding of mass charismatic movements such as Fascism, Nazism, Stalinism, and Maoism.

Do We Need to Know Whether Cults Engage in Brainwashing?

The question of why people obey the sometimes bizarrely insane commands of charismatic leaders, even unto death, is one of the big unsolved mysteries of history and the social sciences. If there are deliberate techniques that charismatic leaders (and charismatically led organizations) use to induce high levels of uncritical loyalty and obedience in their followers, we should try to understand what these techniques are and under what circumstances and how well they work.

This chapter is about nothing other than the process of inducing ideological obedience in charismatic groups. Many people call this process brainwashing, but the label is unimportant. What is important is that those of us who want to understand cults develop models that recognize the importance that some cults give to strenuous techniques of socialization designed to induce uncritical obedience to ideological imperatives regardless of the cost to the individual.

Obedience is Life

The systematic study of obedience has slowed down considerably within the behavioural sciences. Early laboratory studies of obedience-inducing mechanisms got off to a promising start in the 1960s and 1970s, but were correctly criticized by human rights advocates for putting laboratory subjects under unacceptable levels of stress (Kelman and Hamilton 1989; Milgram 1975; Zimbardo 1973). Permission to do obedience-inducing experiments on naive experimental subjects became almost impossible to obtain and these sort of laboratory experiments virtually ceased. However, large numbers of charismatic cultic movements appeared on the scene just in time to fill this vacuum left by abandoned laboratory studies. Being naturally occurring social ‘experiments,’ obedience-induction in such groups could be studied ethnographically without raising the ethical objections that had been raised concerning laboratory studies.

Social theorists are well aware that an extremely high degree of obedience to authority is a reliably recurring feature of charismatic cult organizations (Lindholm 1990; Oakes 1997). But most social scientists interested in religion declined this opportunity. For reasons having more to do with political correctness than scientific curiosity, most of them refused to design research focused on obedience-induction. Many even deny that deliberate programs of obedience-induction ever occur in cults.

The existence of a highly atypical form of obedience to the dictates of charismatic leaders is not in question. Group suicides at the behest of a charismatic leader are probably the most puzzling of such acts of obedience (Hall 2000; Lalich 1999; Weightman 1983), but murder, incest, child abuse, and child molestation constitute other puzzling examples for which credible evidence is available (Bugliosi and Gentry 1974; Lifton 1999; Rochford 1998). However, agreement on these facts is not matched, as we shall see, by agreement on the causes of the obedience, its pervasiveness among cult populations, or the rate at which it decays after the influence stimuli are removed.

AZ - Monastic Procession with bishop

But given the fact that only a small proportion of the human population ever join cults, why should we care? The answer is that the sociological importance of cults extends far beyond their numerical significance. Many cults are harmless and fully deserving of protection of their religious and civil liberties. However, events of recent years have shown that some cults are capable of producing far more social harm than one might expect from the minuscule number of their adherents. The U.S. Department’s annual report on terrorism for the year 2000 concludes that ‘while Americans were once threatened primarily by terrorism sponsored states, today they face greater threats from loose networks of groups and individuals motivated more by religion or ideology than by politics’ (Miller 2000:1).

In his recent study of a Japanese apocalyptic cult, Robert Jay Lifton (1999: 343) has emphasized this point in the following terms:   


‘Consider Asahara’s experience with ultimate weapons…With a mad guru and a few hundred close followers, it is much easier to see how the very engagement with omnicidal weapons, once started upon, takes on a psychological momentum likely to lead either to self-implosion or to world explosion…Asahara and Aum have changed the world, and not for the better. A threshold has been crossed. Thanks to this guru, Aum stepped over a line that few had even known was there. Its members can claim the distinction of being the first group in history to combine ultimate fanaticism with ultimate weapons in a project to destroy the world. Fortunately, they were not up to the immodest task they assigned themselves. But whatever their bungling, they did cross that line, and the world will never quite be the same because, like it or not, they took the rest of us with them.’

Potentially fruitful scientific research on obedience in cultic settings has been stymied by the well-intentioned meddling of two bitterly opposed, but far from disinterested, scholarly factions. On the one hand, there has been an uncompromising outcry of fastidious naysaying by a tight-knit faction of pro-religion scholars. Out of a fear that evidence of powerful techniques for inducing obedience might be used by religion’s enemies to suppress the free expression of unpopular religions, the pro-religion faction has refused to notice the obvious and had engaged in a concerted (at times almost hysterical) effort to sweep under the rug any cultic-obedience studies not meeting impossibly rigorous controlled experimental standards (Zablocki 1997).On the other hand, those scholars who hate or fear cults have not been blameless in the pathetic enactment of this scientific farce. Some of them have tried their best to mystically transmute the obedience-inducing process that goes on in some cults from a severe and concentrated form of ordinary social influence into a magic spell that somehow allows gurus to snap the minds and enslave the wills of any innocent bystander unlucky enough to come into eye contact. By so doing, they have marginalized themselves academically and provided a perfect foil for the gibes of pro-religion scholars.

Brainwashing is the most commonly used word for the process whereby a charismatic group systematically induces high levels of ideological obedience. It would be naively reductionistic to try to explain cultic obedience entirely in terms of brainwashing. Other factors, such as simple conformity and ritual, induce cultic obedience as well. But it would be an equally serious specification error to leave deliberate cultic manipulation of personal convictions out of any model linking charismatic authority to ideological obedience.

EE &amp; Nuns in NA

However, the current climate of opinion, especially within the sociology of new religious movements, is not receptive to rational discussion of the concept of brainwashing, and still less to research in this area. Brainwashing has for too long been a mystified concept, and one that has been the subject of tendentious writing (thinly disguised as theory testing) by both its friends and enemies. My aim in this chapter is to rescue for social science a concept of brainwashing freed from both mystification and tendentiousness. I believe it is important and long overdue to restore some detachment and objectivity to this field of study.

The goal of achieving demystification will require some analysis of the concept’s highly freighted cultural connotations, with particular regard to how the very word brainwash became a shibboleth in the cult wars. It is easy to understand how frightening it may be to imagine that there exists some force that can influence one down to the core level of basic beliefs, values, and worldview. Movies like The Manchurian Candidate have established in the popular imagination the idea that there exists some mysterious technique, known only to a few that confers such power. Actually, as we will see, the real process of brainwashing involves only well-understood processes of social influence orchestrated in a particularly intense way. It still is, and should be, frightening in its intensity and capacity for extreme mischief, but there is no excuse for refusing to study something simply because its frightening.

EE Planting Trees

The goal of establishing scientific disinterest will require the repositioning of the concept more fully in the domain of behavioural and social science rather than its present domain, which is largely that of civil and criminal legal proceedings. It is in this domain that it has been held hostage and much abused for more than two decades. The maxim of scholarly disinterest requires the researcher to be professionally indifferent as to whether our confidence in any given theory (always tentative at best) is increased or decreased by research. But many scholarly writers on this subject have become involved as expert witnesses, on one side or the other, in various law cases involving allegations against cult leaders or members (where witnesses are paid to debate in an arena in which the only possible outcomes are victory or defeat). This has made it increasingly difficult for these paid experts to cling to a disinterested theoretical perspective.

In my opinion, the litigational needs of these court cases have come, over the years, to drive the scientific debate to an alarming degree. There is a long and not especially honourable history of interest groups that are better armed with lawyers than with scientific evidence, and that use the law to place unreasonable demands on science. One need only think of the school segregationists’ unreasonable demands, fifty years ago, that science prove that any specific child was harmed in a measurable way by a segregated classroom; or the tobacco companies’ demands, forty years ago, that science demonstrate the exact process at the molecular level by which tobacco causes lung cancer. Science can serve the technical needs of litigation, but, when litigation strategies set the agenda for science, both science and the law are poorer for it.

My own thirty-six years of experience doing research on new religious movements has convinced me beyond any doubt that brainwashing is practised by some cults some of the time on some of their members with some degrees of success. Even though the number of times I have used the vague term some in the previous sentence gives testimony to the fact that there remain many still-unanswered questions about this phenomenon, I do not personally have any doubt about brainwashing’s existence. But I have also observed many cults that do not practise brainwashing, and I have never observed a cult in which brainwashing could be reasonably described as the only force holding the group together. My research (Zablocki 1971; 1991; 1996; Zablocki and Aidala 1991) has been ethnographic, comparative, and longitudinal. I have lived among these people and watched the brainwashing process with my own eyes. I have also interviewed people who participated in the process (both as perpetrators and subjects). I have interviewed many of these respondents not just one time but repeatedly over a course of many years. My selection of both cults and individuals to interview has been determined by scientific sampling methods (Zablocki 1980: app A), not guided by convenience nor dictated by the conclusions I hoped to find. Indeed, I have never had an axe to grind in this field of inquiry. I didn’t begin to investigate cults in the hope of finding brainwashing. I was surprised when I first discovered it. I insist on attempting to demonstrate its existence not because I am either for or against cults but only because it seems to me to be an incontrovertible, empirical fact.

Although my own ethnographic experience leads me to believe that there is overwhelming evidence that brainwashing is practised in some cults, my goal in this chapter is not to ‘prove’ that brainwashing exists, but simply to rescue it from the world of bogus ideas to which it has been banished unfairly, and to reinstate it as a legitimate topic of social science inquiry. My attempt to do so in this chapter will involve three steps. First, I will analyse the cultural misunderstandings that have made brainwashing a bone of contention rather than a topic of inquiry. Second, I will reconstruct the concept in a scientifically useful and empirically testable form within the framework of social influence theory. Third, I will summarize the current state of evidence (which seems to me to be quite compelling) that some cults do in fact engage in brainwashing with some degrees of success.

To be continued…

St. Nektarios Xenonas


  1. Most of the examples in this chapter will be drawn from studies of religious cults because these are ones with which I am most familiar through my research. But it should be noted that cults need not be religious, and that there are plenty of examples of brainwashing in political and psychotherapeutic cults as well.

Greater Phoenix Tea Party Patriot Starts Petition to Save St. Anthony’s Monastery from being Rezoned

Seraphim Larsen has been a spiritual child of Geronda Paisios, abbot of St. Anthony’s Monastery, almost since the beginning of the monastery’s establishment in Florence; circa 1995. Geronda Paisios baptized him into the orthodox faith during a secret ceremony. Seraphim was the chairman of the Pinal County Republican party.1 Furthermore, he is also a member of the Greater Phoenix Tea Party Patriots.2 It is no secret that Geronda Paisios has openly supported Ron Paul and other right-wing candidates when asked by his spiritual children who they should vote for. Greek edition books concerning the New World Order and Globalism that are sold in St. Anthony’s bookstore contain photocopies from English pamphlets distributed by American neo-Nazi groups.



According to the PRAVSLAVIE:3

A developer is proposing drastic changes to the zoning and land usage for parcels of land IMMEDIATELY ADJACENT to Saint Anthony’s Monastery. These changes would have a dramatic and significant negative impact to Saint Anthony’s Greek Orthodox Monastery. It would also be a dramatic negative impact to all the local residents and out-of-state visitors who attend services there, to all the families who have relatives buried in the cemetery, and to everyone who currently enjoys the tranquil and graceful nature of the Monastery.

The proposed development is truly a drastic change:
• Over 200 new houses are to be developed on land north of the Monastery, including the area immediately adjacent to the Monastery Cemetery
• A new road is to be built extending from Paisano Road, then running in between the two Monastery hills, immediately at the base of the Prophet Elias Chapel, then very close to the cemetery. As this would be the only access road to the 200+ houses, traffic and noise would be greatly increased.
• The noise, traffic, housing, music, animals, etc. would bring a dramatic change to the whole community, disturbing not only the Monastery but the neighbors as well.
• The natural habitat and wildlife would be impacted by a big new development as well, disturbing several rare plant and animal species in the area.

YOUR URGENT ACTION IS REQUESTED! Public meetings are already scheduled for the coming week!! Please sign the petition immediately!

The petition simply asks the local authorities to leave the zoning of these properties unchanged. The purpose of zoning laws is to provide long-term expectations for property owners on how the land in their neighborhood is intended to be used. This allows people to plan and build in a manner appropriate to the existing neighborhood. The proposed new development represents a drastic change to current zoning usage and would have a strongly negative impact on the local community.
The petition also asks the Bureau of Land Management and the Arizona State Land Department simply to leave the usage of these properties unchanged. These agencies must approve any changes to the land usage where the developer wants to build the new road, since it passes through lands controlled by those agencies.

There is a private hearing early next week with Planning & Zoning and the local property owners (including the monastery). The monastery is hoping to bring as many signatures as possible (both paper petitions and printouts of the online petition) to that meeting.

If they don’t make a decision following that initial meeting, it’s likely they will have a public hearing sometime over the next few weeks.

Readers wishing to sign the petition can do so here: https://www.change.org/p/preserve-the-sacred-grounds-of-saint-anthony-s-monastery

Map showing the support from people all over the world.

Seraphim Larsen has also made a sample email for people to send:4

Send to: Pinal County Supervisors: Pete Rios <Pete.Rios@pinalcountyaz.gov>, Cheryl Chase <cheryl.chase@pinalcountyaz.gov>, Steve Miller <Steve.Miller@pinalcountyaz.gov>, Anthony Smith <Anthony.Smith@pinalcountyaz.gov>, Todd House <Todd.House@pinalcountyaz.gov>

Subject Line: Leave Zoning Unchanged near Saint Anthony’s Greek Orthodox Monastery

Email Body:

Dear Pinal County Supervisors,

Thank you for all the great things you are doing for Pinal County!

I am writing to ask you to leave the zoning unchanged near Saint Anthony’s Greek Orthodox Monastery at the end of Paisano Road in the Cactus Forest area immediately south of Florence.

Elmore/Brimhall LLC has submitted a rezoning application to the Pinal County Planning and Zoning Commission, asking that a large parcel immediately adjacent to the north border of the Monastery be rezoned from General Rural to a Residential classification for one-acre residential homes.

I am writing to you to express my strong opposition to this rezoning proposal. The Monastery is the spiritual home to thousands of Greek Orthodox faithful who regularly attend services and pray on the peaceful grounds of the Monastery. The Monastery is the actual residence to over fifty Greek Orthodox Monks who have devoted their lives to worship, prayer, and serving the community. The Monastery is world-renowned for its tranquil, beautiful, and holy grounds, a major tourist attraction where tens of thousands of people visit every year from Arizona and around the world. Many visitors have also expressed interest and concern about the natural habitat and wildlife, and the negative impact of significant development. Finally, the Cemetery at the Monastery is the resting place for many faithful people whose family and relatives would be greatly disturbed to have a busy new road and housing development immediately adjacent to the cemetery, with road noise, houses, animals, and music disturbing the peace of the holy grounds.

I strongly urge you to advise the Planning and Zoning Commission to deny this application for a zoning change.

Thank you for your time and consideration.




More than 5,000 signatures!

Seraphim Larsen

Florence, AZ

JUN 3, 2016 — We passed two significant milestones yesterday:

• We exceeded 5,000 signatures! The petition continues to build momentum!

• We had the meeting with the developer, with the Monastery and the other adjacent property owners.

The developer presented their plans and explained why they are asking for a change in zoning. They listened to the Monastery’s objections but gave no hint of conceding anything to the Monastery—they fully intend to go forward with their rezoning application. The one other neighbor who attended the meeting seemed sympathetic to the Monastery’s concerns, but their own concerns were more about how the development could impact access to their own property, which is at the far end of the development.

In other words, there weren’t any real surprises—this is the outcome we were expecting. The developer is not going to give up this project easily.

This meeting was required for the developer to complete the submission of their zoning application, which will all be turned over to the Planning and Zoning Commission in 1-2 weeks. A public hearing before the Planning & Zoning Commission will then be scheduled—we will be sure to let everyone know when that will happen and how you can speak at the hearing if you want.

It’s likely there will be more than one public hearing with Planning & Zoning, who will eventually make their recommendation to the County Board of Supervisors. The Supervisors will then address the question at another public meeting, at which they will approve or deny the rezoning application.

We will keep everyone informed every step of the way, and let you know what you can do to help. For now, please continue to email the Board of Supervisors—especially if you are local or in-state. See the link below for a sample message and all the contact information for the Board members.


Please also keep the Monastery in your prayers, and also the many people who are working to support the effort to block the rezoning.



  1. http://sonoranalliance.com/pinal-county-republican-chairman-endorses-robert-graham-for-azgop-chairman/
  2. http://phoenixteaparty.ning.com/profile/SeraphimLarsen
  3. http://www.pravoslavie.ru/english/93666.htm
  4. https://www.facebook.com/groups/PreserveSaintAnthonys/

St. Anthony’s EVOO (Gwen Ashley Walters, 2016)

NOTE: The following article is taken from Phoenix Mag, Issue: January 2016:


South of the Arizona State Prison Complex in Florence, smack dab in the middle of the desert with distant mountain vistas, is a serene Orthodox Greek monastery. Flagstone pathways meander around meticulously manicured grounds and storybook chapels.

St. Anthony’s Monastery, established in 1995, funds itself in part by selling olive oil its monks press and bottle on-site from an estate grove. The oil is exquisite, and a fine example of terroir in action: Olive trees flourish in the Sonoran desert. In the fall, monks handpick olives from Mission, Manzanilla and Sevillano olive trees, pressing them all together to produce a cold-pressed, extra virgin olive oil that is raw, robust and a clear expression of the harsh ground from where it comes.

Visitors are welcome at the monastery, but check the website for hours and dress code. Women are required to wear loose-fitting long skirts and long-sleeve shirts, headscarves and closed-toe shoes. The bookstore sells the olive oil in addition to whole olives, hot sauces, and jams and marmalades made from the property’s orchards. You don’t have to travel to Florence to buy a bottle of this exceptional, unfiltered olive oil. Find the oil, sold by volunteers, at select farmers’ markets including Old Town Scottsdale, Roadrunner and Singh Farms, as well as at The Bodega at FnB and Sphinx Date Co. Palm & Pantry. $9 for 250 ml; $24 for 750 ml.

St. Anthony’s Monastery


I Bought Psychiatric Drugs for 40 [Athonite] Monks (Fr. Porphyrios, 2001)

NOTE: The following are excerpts taken from “Radioparagka” show—the Radio Station of the Church Greece. The show’s theme was “Obedience and Freedom” and it was broadcast on November 10, 2002. Fr. Constantine interviews Fr. Porphyrios, a monk from the Hilander Cell of St. Isaac the Syrian. Since the interview has a lot of repetitiveness and superfluous banter, this article has been condensed and contains only the pertinent information.

Since the turn of the century, a few Hagiorite monks have spoken out about the abuses that take place in various monasteries on Mount Athos. These individuals were from different monasteries, under different Gerondas, and yet tell the same story: authoritarianism, sick forms of obedience being imposed, individuals developing mental illnesses after years in these environments, compulsion of monastics to see psychiatrists and take psychiatric drugs, monastics attempting suicides and monastics committing suicide. This “authentic” Athonite monasticism has been prevalent since at least the 70s and these irregularities are now accepted as a normal part of the genuine 21st century monastic experience.

MP3: https://web.archive.org/web/20070113233012/http://www.dorcas.gr/syn/porf.mp3

Text: https://web.archive.org/web/20060623070151/http://www.dorcas.gr/syn/porf.htm#40

Fr. Constantine: Do sick forms of obedience exist?

Fr. Porphyrios: I will disillusion you, Fr. Constantine. Yes, only sick forms of obedience exist today. I’m talking about Mount Athos; I’m speaking purely about Athonite matters and I’ll be speaking from my own experience.

An Abbot yells at his monks, “You don’t listen to me! I lost 500,000 [dr.] on Coca Colas and 7 Up! You’re going to make me explode! Does my money grow on trees?”1 Contemporary Elders are tilling and organizing over Greece towards wild novices. And the finale of these relationships is authoritarianism, which reaches schizophrenia. The difference with the Great Anthony is that he said, “My child, in order to sit near me, certain conditions must be met and above all, love Christ.” Whereas what we seek today on Mount Athos, dozens of times, is what is said and sucked like candy from Athonite abbots who, as you know, 90% of them didn’t do one hour of obedience.

Patriarch Bartholomew & Archbishop Demetrios with the The Coca-Cola Polar Bear ( The World of Coca-Cola, Atlanta, Georgia, USA. October 29, 2009)

Fr. Constantine: They weren’t subordinates beforehand?

Fr. Porphyrios: Never! Only the elders who came from Joseph the Cave-dweller previously did obedience.2 All the others, from the office or Brussels, with their collars and their trousers, became abbots on Mount Athos. Forgive me, but this is the reality. But, exceptions exist. Nothing is 100% absolute. Even on gold we write 999.99.3 Understand? And so, St. Anthony said, “You must love Christ.” Do you know what contemporary elders say? “If you leave my side, you will go to hell my evil child. You will be lost!” Understand?

Once, I was at a monastery and a student came to stay for fifteen days. After fifteen days, they placed him as my assistant. I saw him with a new pair of shoes. Hm! He smelt like gun powder to me. I asked him, “Are you engaged, will you go, this and that.” He said, “I thought that I would stay here.” Today, he is a schizophrenic! And the monastery boasts 120-130 monks. I don’t know how many monks they have but 40 of them are taking psychiatric drugs out of obedience…bad obedience!

ιερομόναχος π. Λουκάς
Hieromonk Lukas Gregoriatis (one of the monk-doctors who manages the use of psychotropic drugs in the monastery)

Fr. Constantine: Of course, when you say psychotropic drugs, there is a book written by a Grigoriatis monk who responds to such allegations. He mentions that psychotropic drugs are bought for Mount Athos. He also says that he buys them for the pilgrims that visit from abroad. Visitors come and say they need psychiatric drugs.” http://www.rel.gr/index.php?rpage=meletes&rpage2=showkeimeno.php&link_id=5

Fr. Porphyrios: Excuse me! I myself bought psychotropic drugs for 40 monks! Understand? I personally bought psychiatric medication for 40 monks when I went out of Mount Athos to run errands. I don’t know how many monks there are now who take psychiatric drugs.

I tread differently. I went asked someone once—one with no taste, from the bench, the university—and I asked him. We chatted a bit and he told me, “You will do this!” I said, “I cannot.” He replied, “You will do this, it is obedience!” I was coerced; the priest said it. Fr. Constantine, everyone has been trampled by disobedience. I was pressed by obedience. I did obedience and suffered AF [atrial fibrillation] and was put on a pacemaker.5

Fr. Constantine: This priest gave you that medical advice?

Fr. Porphyrios: No! He told me to do something that my conscience couldn’t allow me to do. I told him, “I cannot do that thing.”

Fr. Constantine: Was the obedience a sin?

Fr. Porphyrios: The obedience was a spiritual sin, not a carnal sin. My conscience would not allow it.

Fr. Constantine: You had the right not to do this obedience since it was a spiritual sin.

Fr. Porphyrios: Yes, of course. Also, this priest wasn’t my spiritual father. We chatted, I told him one issue. What else do you want me to tell you?

Fr. Constantine: Everything! You lived the events.

Fr. Porphyrios: A certain monk in a large monastery fell into some small offence. He was ashamed to tell the abbot and he told some transient, worldly spiritual father. All of Mount Athos to Daphne learned about it. And his geronda—the great and holy geronda—called him to the Assembly and told him, “Because you didn’t tell me this, for one year you will not receive a blessing and I will not give you my blessing.” Fr. Constantine, before the new elders, spiritual fathers, such as the ascetic Geronda Sofronius, came to the monastery every month to confess people. The old elders wisely decided that a spiritual father from another monastery could come every month so that we could say something that we couldn’t tell the abbot so as not to create a disagreement or discord. Understand? Today, this has been cut off. Now if you confess to someone other than the abbot, you’re excommunicated and expelled from the community.6

And I’ll tell you something else. I saw a “very pleasant” abbot beating a subordinate and I lost it! I got my stuff and left. The abbot told me, “Come here! Do you know what I’m doing? What Abba Dorotheos did!” Eh! I couldn’t bear it and told him, “Excuse me, Geronda, but as soon as you become St. Dorotheos and Abba Seridos, then you can hit children.”7 It broke my heart. Now this monk is in the wind. I went and he told me, “Porphyrios, I beg you to discharge me.” We both cried. This is obedience today, Fr. Constantine. But when we haven’t done obedience ourselves, how can we impose someone else to do obedience? Naturally, obedience is salvation!

“But if one does not sincerely wish to do the will of God, then though go to a prophet, God might place in the heart of that prophet an answer corresponding to the man’s corrupt heart, as the Scripture says, And if a prophet should cause to err and should speak, I the Lord have caused that prophet to err (Ezekiel 14:9).”

I remember Geronda Gabriel told me the following, “Do you see that man? He came from Cambridge with a transistor in the armpit. And I said to him, ‘What did you come here to do?’ He replied, ‘I came here to be a monk.’ I asked him, ‘Do you know what it means to be a monk?’ I tell you to fall from the balcony and you fall.’ The man replied, ‘Be careful, Geronda. If you’re telling me a joke, you catch me by my feet because I will fall.’ Geronda Gabriel told me, ‘He is an angel.’

And Geronda Gabriel Dionysiatis placed someone as the supervisor and tells him after three-four sessions; the monk disagreed, ‘Geronda, I will resign.’ The Geronda asked, ‘Why, my child?’ ‘I cannot disagree with my Geronda.’ Geronda Gabriel told him, ‘Ah, pay attention. I placed you there for this reason. You will say your opinion in governing. You will do obedience in spiritual matters. Also, I want there to be a contrary voice.’

Today, whoever disagrees about the Hegumen council departs. And they’ve gathered all the idiots; those who never tread in the monastery, they are from outside and they’ve placed them, how do you say…

Elder Gabriel Dionysiatis a
Elder Gabriel Dionysiatis (†1983)

Fr. Constantine: This is sick. If there is no voice to be heard…

Fr. Porphyrios: No, no, no. Let me tell you something else—Speechless!—about an abbot who went to a new monastery that had an old brotherhood. ‘Oh well,’ he says, ‘are they dumb? Do they not have any other opinion than that from the Geronda?’ And Geronda Gabriel told Fr. Kallinikos, ‘You will stay, my child! Your objection is needed so we can find the right one.

Today, Fr. Constantine, the Geronda is the path (ο-δο) of salvation. And he is also our obstacle to arrive at Christ. ‘I am here (‘δω)!’ Understand? Things are tragic!



  1. When Geronda Ephraim of Arizona was the abbot of Filotheou Monastery, on the Feast Day of Pascha he would traditionally bless the monks to have a can of Coca Cola for trapeza. He carried this tradition over to St. Anthony’s Monastery in Florence, AZ. In some of Geronda Ephraim’s North American monasteries, the monastics have soda frequently throughout the year.
  2. The following is a pilgrim’s impression of his visit to Elder Joseph the Hesychast and his disciples in the 50s:

“Many times the Greeks have a weakness to make up stories, especially about people‘s alleged holiness, and as is often the case, once a story starts, it gets repeated and magnified and blown completely out of proportion. Others hear it and relate it in a different way, and it becomes another story, and soon someone has a reputation of being one of the greatest wonderworkers the Church has ever known.

“To some extent, this is what happened with this Elder Ephraim of Katounakia, and it‘s most definitely what happened to the other Fr. Ephraim, who took control of Philotheou Monastery, and who could not remain on the Holy Mountain, but resides in America. The three pilgrims came to Katounakia for a visit, and the meeting with Fr. Ephraim of Katounakia was quite eventless, other than the fact that Fr. Ephraim seemed like a reluctant prisoner, not one rejoicing in obedience. There was quiet talking, all in Greek, greetings and goodbyes.

“The outcome of this occasion was in stark contrast with the words of Fr. Ephraim in his book which recently appeared. Ironically, the book was named Obedience is Life: Elder Ephraim of Katounakia, and in this very book this alleged Elder Ephraim records that he later forced his elder to begin once again commemorating the Ecumenical Patriarch at a time when the Ecumenical Patriarch had openly espoused the heresy of Ecumenism. He thus recorded his own blatant disobedience in a book supposedly teaching obedience. The Elder Nikephoros, it is said, grieved very much that his disciple, Fr. Ephraim, had forced him to do something against his conscience. Three years before he died, however, the Elder Nikephoros returned to the church. http://www.archbishopgregory.info/chapter_06_first_visit_to_mount_athos.shtml

  1. 999.99 (five nines fine) The purest type of gold currently produced.
  2. In Geronda Ephraim’s monasteries, it is taught that monastics lose their salvation if they leave the monastery–whether a novice, rassaphore, or great-schema. Unless, of course, they get a blessing. In 2000/2001 a monk from St. Nektarios Monastery was visiting St. Anthony’s once for an immigration appointment–when new monasteries were opened the immigrant monks who were transferred did not update their address and would travel back to Arizona when they had INS appointments. This monk did not want to return to St. Nektarios Monastery due to the angry and oppressive atmosphere that existed there at that time (as one of the Athonite Fathers said, “Do you want to destroy a monk’s spirituality? Send him to help establish a new monastery). Geronda Ephraim gave him a blessing to be a monk at St. Anthony’s but due to the fact that he didn’t first consult his Geronda [Joseph] and did it of his own volition, it was frowned upon by many of the other monastics in various monasteries. Some even believed that this act would greatly hinder his chances for salvation and that although Geronda Ephraim gave a “blessing,” he was actually doing obedience to this young monk, and the whole affair was really just an act of self-will not really covered by a blessing. In another case, a novice from Arizona received a blessing to go live at Filotheou Monastery on Mount Athos. Many of his brother monastics gossiped that in reality, Geronda didn’t actually bless it but rather this novice ceaselessly begged and harassed Geronda to allow him to go and in the end Geronda Ephraim did obedience to his will and “blessed” it. Later, rumors went around the monastery that he became deluded, was re-baptized at an old calendarist monastery, and was ordained a deacon. Afterwards, he left the monastic life and returned to the world. These stories are used as cautionary tales to scare young novices and remind them that if they don’t stay with their elder no matter what, and do absolute blind obedience, they will become deluded, leave, and lose their soul.
  3. In Geronda Ephraim’s monasteries, the only possible sin in obedience is disobedience; i.e. not obeying the command the elder has given. Both physical and spiritual obedience are expected, and it must be blind, without questioning, back-talk, murmuring, judging, etc. Thus, if the obedience is a “spiritual” sin (i.e. lying, perjury, falsifying records, gaslighting pilgrims to cover-up a scandal, etc.), the disciple is not sinning if he/she obeys the command. The Geronda or Gerondissa are responsible before God for the commands given (thus, if the order they give is a sin, it is their burden and sin). The disciple is only responsible before God for the obedience they did or did not do, regardless of whether it is a sin or breaking of the commandments. In doing obedience they have not sinned and will not have to give an account before God on why they broke His commandments. If they disobey the command, they will have to give an account for their disobedience, and possibly lose their salvation for this crime.
  4. In Geronda Ephraim’s monasteries, a monastic does not have a blessing to confess to other spiritual fathers other than their own. I.e., a monk from Michigan wouldn’t ask the hieromonks in TX or NY if he could see them for confession. The only other person monastics can confess to other than their own spiritual father/mother in the monasteries is Geronda Ephraim in Arizona. A blessing is usually still required before this happens, though.
  5. Geronda Ephraim uses the story of Akkakios a lot in his homilies to his monastics. This story is meant to encourage moanstics to endure everything their superiors put them through with patience and without murmuring. “The humble monk distinguished himself by his patient and unquestioning obedience to his Elder, a harsh and dissolute man. He forced his disciple to toil excessively, starved him with hunger, and beat him without mercy. Despite such treatment, St Acacius meekly endured the affliction and thanked God for everything. St Acacius died after suffering these torments for nine years. Five days after Acacius was buried, his Elder told another Elder about the death of his disciple. The second Elder did not believe that the young monk was dead. They went to the grave of Acacius and the second Elder called out: “Brother Acacius, are you dead?” From the grave a voice replied, “No, Father, how is it possible for an obedient man to die?” [St. Akakios is celebrated in the Orthodox Church on July 7th].
  6. This is the kind of obedience required in Geronda Ephraim’s monasteries. Similar to Abraham who didn’t question God when told to murder his son, Isaac, nor did he think twice about sacrificing him, so to should the disciple monk do everything Geronda asks without hesitation or a double mind.
  7. AKAKIOS111