Hostname: page-component-8448b6f56d-xtgtn Total loading time: 0 Render date: 2024-04-15T12:43:14.838Z Has data issue: false hasContentIssue false

Understanding Process Tracing

Published online by Cambridge University Press:  18 October 2011

David Collier
Affiliation:
University of California, Berkeley
Rights & Permissions [Opens in a new window]

Abstract

Process tracing is a fundamental tool of qualitative analysis. This method is often invoked by scholars who carry out within-case analysis based on qualitative data, yet frequently it is neither adequately understood nor rigorously applied. This deficit motivates this article, which offers a new framework for carrying out process tracing. The reformulation integrates discussions of process tracing and causal-process observations, gives greater attention to description as a key contribution, and emphasizes the causal sequence in which process-tracing observations can be situated. In the current period of major innovation in quantitative tools for causal inference, this reformulation is part of a wider, parallel effort to achieve greater systematization of qualitative methods. A key point here is that these methods can add inferential leverage that is often lacking in quantitative analysis. This article is accompanied by online teaching exercises, focused on four examples from American politics, two from comparative politics, three from international relations, and one from public health/epidemiology.

Type
The Teacher
Copyright
Copyright © American Political Science Association 2011

Process tracing is a fundamental tool of qualitative analysis. In the framework presented here,Footnote 1 it is defined as the systematic examination of diagnostic evidence selected and analyzed in light of research questions and hypotheses posed by the investigator. Process tracing can contribute decisively both to describing political and social phenomena and to evaluating causal claims. George and Bennett have played the leading role in developing this method as an essential form of within-case analysis,Footnote 2 and Fenno's “soaking and poking” is a kindred research procedure.Footnote 3

Although the idea of process tracing is often invoked by scholars as they examine qualitative data, too often this tool is neither well understood nor rigorously applied. Relatedly, the field of qualitative methods in political science—in sharp contrast to quantitative methods—is inadequately equipped with procedures for teaching basic research tools, including process tracing.

This two-fold deficit motivates this article, which offers a new framework for understanding, applying, and teaching process tracing. The approach is distinctive in three ways.

  • Process Tracing vis-à-vis CPOs. The evidence on which process tracing focuses corresponds to what Collier, Brady, and Seawright (Reference Collier, Brady, Seawright, Brady and Collier2010a) call causal-process observations, or CPOs. The idea of CPOs highlights the contrast between (a) the empirical foundation of qualitative research, and (b) the data matrices analyzed by quantitative researchers, which may be called data-set observations (DSOs). Some of the literature on which this article draws (e.g., Brady Reference Brady, Brady and Collier2010; Freedman Reference Freedman, Brady and Collier2010a; Mahoney Reference Mahoney2010) formulates arguments in terms of CPOs, rather than in terms of process tracing per se. The present article treats these methodological tools as two facets of the same research procedure. Throughout, the article consistently refers to “process tracing” to avoid applying two labels to what is basically the same method.

  • Description. Careful description is a foundation of process tracing, a perspective emphasized by Mahoney (Reference Mahoney2010, 125–31). Process tracing inherently analyzes trajectories of change and causation, but the analysis fails if the phenomena observed at each step in this trajectory are not adequately described. Hence, what in a sense is “static” description is a crucial building block in analyzing the processes being studied.

  • Sequence. Process tracing gives close attention to sequences of independent, dependent, and intervening variables. Again, we follow Mahoney, who has productively advanced this approach.

TEACHING EXERCISES

This new formulation of process tracing is accompanied by online teaching exercisesFootnote 4 that encompass diverse substantive areas.

a. American Politics: Fenno (Reference Fenno1977) on members of Congress; Brady (Reference Brady, Brady and Collier2010) on the 2000 Presidential election; Skocpol et al. (Reference Skocpol, Gans and Munson2000) on civic associations; and Weaver (Reference Weaver2007) on crime policy.

b. Comparative Politics: Lerner (Reference Lerner1958) on social change in a Turkish village; and Rogowski (Reference Rogowski, Brady and Collier2010) on the interaction of theory and case studies.

c. International Relations: Tannenwald (Reference Tannenwald1999) on the US “nuclear taboo” after World War II; Bennett (Reference Bennett, Brady and Collier2010) on the Fashoda crisis, Germany's expansion of military goals during World War I, and Soviet nonintervention in Eastern Europe in 1989; and Schultz (Reference Schultz2001) on democracy and coercive diplomacy.

d. Public Health: Freedman (Reference Freedman, Brady and Collier2010a) on major breakthroughs in the history of epidemiology.

e. Detective Fiction: The Sherlock Holmes story “Silver Blaze” (posted online with the exercises) serves as the basis for an exercise, and also as a running example in parts of the following presentation. This story is not social science, yet it provides vivid illustrations of process tracing and is an engaging text for teaching.

PROCESS TRACING, PRIOR KNOWLEDGE, AND DIAGNOSTIC EVIDENCE

Process tracing, to reiterate, is an analytic tool for drawing descriptive and causal inferences from diagnostic pieces of evidence—often understood as part of a temporal sequence of events or phenomena. Given the close engagement with cases and the centrality of fine-grained case knowledge, process tracing can make decisive contributions to diverse research objectives, including: (a) identifying novel political and social phenomena and systematically describing them; (b) evaluating prior explanatory hypotheses, discovering new hypotheses, and assessing these new causal claims; (c) gaining insight into causal mechanisms; and (d) providing an alternative means—compared with conventional regression analysis and inference based on statistical models—of addressing challenging problems such as reciprocal causation, spuriousness, and selection bias. Thus, qualitative tools can add leverage in quantitative analysis. They can also strengthen causal inference in small-N designs based on the matching and contrasting of cases—designs which have great value, but whose contribution to causal inference urgently needs to be supplemented by within-case analysis.Footnote 5

Process tracing requires finding diagnostic evidence that provides the basis for descriptive and causal inference. How does the researcher establish that a given piece of evidence is diagnostic?Footnote 6

Identifying evidence that can be interpreted as diagnostic depends centrally on prior knowledge.Footnote 7 For the purpose of the online exercises, we distinguish four interrelated types of knowledge—extending distinctions offered by Waltz (Reference Waltz1979), whose ideas are important in the international relations examples included here.

  • Conceptual Frameworks. A first type of prior knowledge involves sets of interrelated concepts, often accompanied by general ideas of how the concepts can be operationalized. These frameworks thereby identify and link the topics seen as meriting analytic attention. The framework often points to the counterfactuals that conceptually establish what it means for a given phenomenon to be absent, that is, the “contrast space” (Garfinkel Reference Garfinkel1981) that organizes the analysis.

  • Recurring Empirical Regularities. These are established patternsFootnote 8 in the relationships among two or more phenomena. Waltz (Reference Waltz1979, 1) states that this is “not simply … a relationship that has been found, but … one that has been found repeatedly.” The corresponding “if a, then b” (Reference Waltz1979, 1) connection may be viewed as causal, or it may be understood descriptively.

  • Theory-I. This builds on these recurring regularities by more tightly connecting them as a set of insights into “a particular behavior or phenomenon” (Waltz Reference Waltz1979, 2). Thus, many social scientists seek to build theory “by collecting carefully verified, interconnected hypotheses.”

  • Theory-II. A final type of prior knowledge entails not only interconnected empirical regularities (Theory-I), but also a set of statements that explain them, that is, offering explanations of why these regularities occur (Waltz Reference Waltz1979, 5). Theory-II may also be called an explanatory model.

As is clear in the exercises, some studies are explicit and precise about the prior knowledge that frames the research, whereas for other studies it is necessary to consult a wider literature to understand the theoretical background. Unfortunately, as investigators write up their research, they may overstate the coherence of the findings vis-à-vis prior knowledge—sometimes making it hard to identify the theoretical starting point. Reconstructing this starting point can require detective work—which is sometimes needed in evaluating diagnostic evidence in some of the exercises.

Against this backdrop, we consider the contribution of process tracing to descriptive and causal inference.

DESCRIPTIVE INFERENCE

Careful description is fundamental in all research, and causal inference—whether assessed with qualitative or quantitative tools—depends on it.Footnote 9 Close engagement with case knowledge in process tracing can provide a good foundation for addressing this task.

A key point must be underscored again. As a tool of causal inference, process tracing focuses on the unfolding of events or situations over time. Yet grasping this unfolding is impossible if one cannot adequately describe an event or situation at one point in time. Hence, the descriptive component of process tracing begins not with observing change or sequence, but rather with taking good snapshots at a series of specific moments. To characterize a process, we must be able to characterize key steps in the process, which in turn permits good analysis of change and sequence.Footnote 10

Mahoney (Reference Mahoney2010, 127–28) illustrates descriptive inference in process tracing with Tannenwald's (Reference Tannenwald1999) study of the “Nuclear Taboo.”Footnote 11 Tannenwald argues that the horrified reaction to the use of nuclear weapons at the end of the World War II created a nuclear taboo that strongly influenced later US nuclear policy, specifically decisions about the non-use of nuclear weapons during subsequent military crises. Whereas this taboo grew out of the reaction at the level of public opinion, it evolved into a normative mandate embraced by policy makers (Reference Tannenwald1999, 462). A crucial task in Tannenwald's study is to establish empirically (a) that this horrified reaction did in fact occur; (b) how widespread it was; and (c) that the elements of this reaction did indeed add up to a nuclear taboo. Process tracing focuses on finding and interpreting diagnostic evidence that addresses these descriptive tasks. This nuclear taboo, in turn, is the key independent variable in the study that is evaluated vis-à-vis rival explanations of the non-use of nuclear weapons.

Lerner's (Reference Lerner1958) analysis of rapid “modernization” in a Turkish village likewise illustrates the intensive description that should be a foundation of process tracing.Footnote 12 This transformation results from the election of a new national governing party and the subsequent introduction of infrastructure that includes electricity and a modern road to Ankara. The transformation of the village is the dependent variable, and the author's goal is to describe change in this variable over time. The analysis focuses on dozens of specific observations of social attributes and interactions; demographic characteristics; and material objects, physical infrastructure, and commercial establishments.

These two examples of description differ in important ways. Tannenwald relies on diverse primary and secondary sources—including official documents, memoirs, and biographies—that shed light on the politics of nuclear policy making. By contrast, Lerner's study depends on intensive interviewing, carried out by his field assistants. Further, as noted, the phenomenon described through process tracing is Tannenwald's main independent variable; thus, the nuclear taboo is her hypothesized explanation for the US post-World War II non-use of nuclear weapons. By contrast, for Lerner the posited modernization of the village is the dependent variable triggered by an electoral shift and the initiative of the victorious party to build new infrastructure. Finally, while both Lerner and Tannenwald offer a rich and detailed description of a key variable, Tannenwald also gives substantial attention to rival explanations.

The two examples also illustrate another point: The qualitative researcher should recognize that the fine-grained description in process tracing sometimes relies on quantitative data. This is certainly reasonable, given that—in the spirit of pursuing multi-method research—the boundary between qualitative and quantitative should not be rigid.Footnote 13 For Lerner, some of the information is demographic, involving numerical data. As seen in the exercises, Brady's (Reference Brady, Brady and Collier2010) process tracing study employs quantitative data on elections and voting. In parallel, Tannenwald could have assessed the pervasiveness of horrified reactions by counting their overall frequency, different types of horrified reactions, and change in these counts over time. Process tracing does indeed focus on single “nuggets” of information, yet sometimes this information involves counts and not just single actions or occurrences.

A different form of description, based on counterfactuals, is illustrated by the Sherlock Holmes story “Silver Blaze.” Here the central puzzleFootnote 14 is to explain the murder of John Straker, trainer of the racehorse Silver Blaze. The focus is on a singular event that cannot be disaggregated—a focus also common in process-tracing research in international relations. With singular events, description may be based on comparison of the observed value of a given variable with one or more hypothetical—i.e., counterfactual Footnote 15—values that are seen as plausible alternatives, but that do not occur in the case being studied. The comparison depends on the contrast space noted above, which builds on the researcher's background knowledge. Counterfactuals are important in diverse areas of research (e.g., King, Keohane, and Verba Reference King, Keohane and Verba1994, 77–78, 88–89), and they play a particularly visible role here.

CAUSAL INFERENCE

Basic ideas about applying process tracing to causal inference may be summarized in terms of four empirical tests. Slightly adapting the formulation of Bennett (Reference Bennett, Brady and Collier2010), who builds on the work of Van Evera (Reference Van Evera1997), the tests are classified according to whether passing the test is necessary and/or sufficient for accepting the inference. Based on these criteria, table 1 presents the four tests: straw-in-the-wind, hoop, smoking-gun, and doubly decisive. The table also notes the implications for rival hypotheses of passing each test. If a given hypothesis passes a straw-in-the-wind test, it only slightly weakens rival hypotheses; with hoop tests it somewhat weakens them; with smoking-gun tests it substantially weakens them; and with doubly decisive tests passing eliminates them—of course, with the usual caveat that the definitive elimination of a hypothesis is often hard to achieve in social science.

Table 1 Process Tracing Tests for Causal Inference

Source: Adapted from Bennett (Reference Bennett, Brady and Collier2010, 210), who builds on categories formulated by Van Evera (Reference Van Evera1997, 31–32).

Before we introduce causal inference, it is useful to reiterate two ideas discussed above: process tracing can focus either on recurring events or on a singular event; and although it is reasonable to think of process tracing as a qualitative method, it sometimes relies on quantitative information. Three other points should also be emphasized:

  • Specification of Hypotheses. Careful, analytically informed specification of hypotheses is essential both in selecting and interpreting pieces of evidence, and in weighing them against one another. Background knowledge is fundamental here.

  • Distinctions among Tests. The distinctions in table 1 are a useful heuristic, but should not be taken rigidly. The decision to treat a given piece of evidence as the basis for one of the four tests can depend on the researcher's prior knowledge, the assumptions that underlie the study, and the specific formulation of the hypothesis. Although in general the appropriate test is clear, sometimes a piece of evidence treated as a straw-in-the-wind might instead be viewed as the basis for a hoop test or a smoking-gun test (see tables 4 and 5 below). Alternatively, it might simply be viewed as an “intermediate” test, with corresponding implications for rival hypotheses.

  • Assumptions and Interpretations. The decision about which test is appropriate to a particular piece of evidence thus involves different assumptions and interpretations.Footnote 16 For example, if researchers make the weaker assumption that a given event (or other piece of evidence) may be a coincidence, they should and will be more cautious. Alternatively, if they make the stronger assumption—based on prior knowledge—that it is probably not a coincidence, they may arrive at a different conclusion about accepting or rejecting the hypothesis.Footnote 17

Against this backdrop, we discuss the four process-tracing tests in table 1, using the Sherlock Holmes story “Silver Blaze” as a running example. At this point it may be useful for readers to examine the story itself.

Mapping the Holmes story onto the framework presented earlier, we might say that the suspects are in effect the hypotheses, and the clues are causal process observations (CPOs). Table 2 provides an overview of the story and presents the hypotheses used in illustrating the four tests, organized according to whether they concern an independent, intervening, or dependent variable. As indicated in the table, the mystery contains two causal puzzles: explaining the murder of John Straker and the disappearance of the horse. The following examples concentrate on the murder, and references to the horse's disappearance are considered when crucial to the murder itself.

Table 2 Overview of “Silver Blaze”

Straw-in-the-Wind Tests

These tests, illustrated in table 3, can increase the plausibility of a given hypothesis or raise doubts about it, but are not decisive by themselves. Straw-in-the-wind tests thus provide neither a necessary nor a sufficient criterion for accepting or rejecting a hypothesis, and they only slightly weaken rival hypotheses. Of the four tests, these are the weakest and place the least demand on the researcher's knowledge and assumptions. Yet they provide valuable benchmarks in an investigation by giving an initial assessment of a hypothesis. Furthermore, if a given hypothesis passes multiple straw-in-the-wind tests, it adds up to important affirmative evidence.

Table 3 Straw-in-the-Wind Tests

In “Silver Blaze,” one straw-in-the-wind is based on the clues about the bill for expensive women's clothing found in Straker's pocket and Straker's wife's ignorance of the costly dress that had been purchased. This lends weight to Holmes's suspicion about Straker's role (H1) and to the idea that Straker might have had a financial motive for throwing the race, but is not by itself a decisive piece of evidence. Another straw-in-the-wind is one of the most famous clues in all of detective fiction: that the dog presumably guarding the horse's stable “did nothing in the night,” an observation that points to the possibility that someone known to the dog—i.e., Straker—abducted the horse (H3). Yet it certainly does not confirm this hypothesis.

Hoop Tests

Hoop tests (table 4) set a more demanding standard. The hypothesis must “jump through the hoop” to remain under consideration, but passing the test does not by itself affirm the hypothesis. Although not yielding a sufficient criterion for accepting the explanation, it establishes a necessary criterion. Hoop tests do not confirm a hypothesis, but they can eliminate it. Compared to the straw-in-the-wind tests, passing hoop tests has stronger implications for rival hypotheses: it somewhat weakens their plausibility, without precluding the possibility that alternative hypotheses may be relevant.

Table 4 Hoop Tests

Table 4 presents two hoop tests, both focused on the hypothesis that Simpson killed Straker (H6). Simpson carried a potential murder weapon, so that he passes the corresponding hoop test and is therefore not precluded as a suspect. However, his timid, non-menacing appearance might seem to preclude his being a murderer who “shattered” Straker's head with a “savage” blow. Hence, he fails this second hoop test. Table 4 also illustrates alternative interpretations of the same piece of evidence. One could assume, as just stated, that a timid individual such as Simpson would never commit a savage murder—thereby eliminating him as a suspect through the hoop test. Alternatively, it may be unlikely, but definitely not impossible, that a timid and nonmenacing person would commit such a murder. This would yield a straw-in-the-wind that casts doubt on the idea that that he is the murderer—yet he remains a possible suspect.

Smoking-Gun Tests

The metaphor of a “smoking gun” conveys the idea that a suspect who is caught holding a smoking gun is presumed guilty. However, those with no smoking gun may not be innocent. In other words, this provides a sufficient but not necessary criterion for accepting the causal inference. It can strongly support a given hypothesis, but failure to pass does not reject it. If a given hypothesis passes, it substantially weakens rival hypotheses.

In “Silver Blaze,” the first smoking-gun test (table 5) is straightforward. The fact that the maid brought the curried mutton to the stable shows that the initiative to drug the stable boy—a key step in the chain of events—had to begin in Straker's household (H2). By contrast, the hypothesis that Straker planned to cause harm (H4) is ambiguous and illustrates the importance of prior knowledge and assumptions. Depending on such knowledge and assumptions, the knife found with Straker can be viewed as extraordinarily odd and suspicious, or only somewhat unusual. Accordingly, the knife is alternatively a smoking-gun or a straw-in-the-wind.

Table 5 Smoking-Gun Tests

Doubly Decisive Tests

These tests provide strong inferential leverage that confirms one hypothesis and eliminates all others. They meet both the necessary and sufficient standard for establishing causation. As Bennett (Reference Bennett, Brady and Collier2010, 211) notes, single tests that accomplish this are rare in social science, but this leverage may be achieved by combining multiple tests, which together support one explanation and eliminate all others.

Turning again to the Sherlock Holmes example (table 6), we see that Simpson, Straker, and the horse are suspects in Straker's death. Simpson (H6) and Straker (H7) are removed from suspicion by hoop tests. Based on straw-in-the-wind tests, one of which might be interpreted as a smoking-gun test, Holmes infers that the horse kicked Straker, thereby inflicting the grievous blow that shattered his head. The conjunction of these diverse tests serves to eliminate other suspects and establish the horse's guilt (H8), thereby meeting the standard of both necessity and sufficiency.

Table 6 Building a Doubly Decisive Test by Evaluating Alternative Hypotheses

Combining tests in this way poses important challenges. Central here is Holmes's method of elimination, a strategy invoked various times in Doyle's stories. Put simply, when the investigator has eliminated all plausible alternatives, the remaining scenario must be the correct one. Variants of this method are widely recognized, as with eliminative induction in Bayesian analysis (Vineberg Reference Vineberg1996) and J.S. Mill's (Reference Mill1974, 397–98) method of residues.

The method of elimination is especially relevant here because although two suspects are definitely eliminated through hoop tests, the guilt of the horse is established primarily on the basis of weaker straw-in-the-wind tests. The procedure of elimination is valuable because it relies centrally on the definitive exoneration of the first two suspects, and only secondarily on explicit demonstration of the horse's guilt.

Further, the method of elimination has special relevance to the case of Silver Blaze, given that both Mill (Reference Mill1974, 398) and Holmes emphasize the value of this method for discovering unusual or even bizarre explanations—such as the kick of the horse as a murder weapon (see table 6, H8, inference d). Mill states that among his methods, “this is the most fertile in unexpected results” (p. 398), and as Holmes puts it, “when you have eliminated the impossible, whatever remains, however improbable, must be the truth… .”Footnote 18 In contemporary political science, a recurring concern is that the discipline needs tools for discovering unexpected and unusual explanations. The method of elimination merits attention as precisely this kind of tool.Footnote 19

Causal-Sequence Framework: Auxiliary Outcome Test

Another test is suggested by Mahoney's (Reference Mahoney2010, 125–31) causal-sequence framework. He introduces the idea of auxiliary outcomes, which are not part of the main causal sequence yet provide valuable inferential leverage.Footnote 20 Fruitful theories generate multiple observable implications (King, Keohane, and Verba Reference King, Keohane and Verba1994, passim), and a particular independent variable or mechanism hypothesized to influence the dependent variable may also affect an auxiliary outcome. The inference thereby derived may further support the causal importance of the independent variable or intervening mechanism. In “Silver Blaze,” the lame sheep are an example of an auxiliary outcome (table 7). Holmes infers that Straker wished to practice the delicate operation required to injure the horse (H5). This does not directly injure the horse; it is a secondary outcome that makes Holmes's ideas about the central causal process more plausible. Auxiliary outcome tests generally yield straws-in-the-wind, as occurs in this example.

Table 7 Auxiliary Outcome

Mahoney further illustrates the auxiliary outcome test with a social science example: Luebbert's (Reference Luebbert1991) famous book, Liberalism, Fascism, or Social Democracy. Luebbert's central argument is that a “red-green” coalition of socialist parties and the middle peasantry was a key factor in the formation of national-political economies in interwar Europe. Mahoney shows that while this claim is partly developed through small-N comparative research and partly through a focus on mechanisms, Luebbert also builds his case by arguing that if a red-green alliance really did foster social democracy, it should have left behind other markers, including the reluctance of socialists to challenge the distribution of wealth in the countryside (Mahoney Reference Mahoney2010, 130). The discovery of such auxiliary outcomes suggests that the red-green alliance had a key impact on other domains of national politics. This finding reinforces the idea that the alliance was highly influential, yielding stronger grounds for inferring that it also shaped the national political-economic regime (Mahoney Reference Mahoney2010, 130).

CONCLUSION

This article seeks to improve the practice of process tracing as a strategy of qualitative analysis, a strategy that can also contribute to quantitative research. The discussion is accompanied by the online exercises focused on ten empirical studies, from diverse subfields, aimed at encouraging careful thinking—and productive teaching—about process tracing.

Three concluding points merit emphasis. First, as Brady, Collier, and Seawright (Reference Brady, Collier, Seawright, Brady and Collier2010, 22) note, “both qualitative and quantitative research are hard to do well.” Qualitative tools such as process tracing can address some challenges faced in quantitative analysis, but process tracing faces serious issues in its own right.Footnote 21 Doubts may arise as to which causal-inference test is appropriate. The analysis may face standard problems of missing variables. Measurement error can be an issue, and probabilistic relationships are harder to address than in quantitative research. This article is intended as one step in developing and refining tools for process tracing—and it is urgent that it not be the last step. More work must be done.

Second, in a given study, how does one begin to carry out process tracing? It is certainly valuable to approach process tracing with the expectation of using the causal inference tests presented in table 1, yet these tests are not always easy to apply. It can therefore be productive to start with a good narrative or with a timeline that lists the sequence of events. One can then explore the causal ideas embedded in the narratives, consider the kinds of evidence that may confirm or disconfirm these ideas, and identify the tests appropriate for evaluating this evidence.

Finally, along with the value per se of refining process tracing, this discussion is important in wider debates on political methodology. Political science is in a period of major innovation in refining tools for quantitative analysis, and in particular, quantitative tools for causal inference. This trend has produced some worries among qualitative researchers about the adequacy of their own tools, and perhaps it has intensified the skepticism of some quantitative researchers about causal inference in qualitative studies. This skepticism led the eminent statistician David Freedman (Reference Freedman, Brady and Collier2010a) to counter with the argument that the kind of qualitative analysis involved in process tracing is indeed a type of scientific inquiry in its own right. In that spirit, the goal here is to take steps toward placing this form of inquiry on a more rigorous foundation.

Footnotes

1 The approach discussed here differs from other research traditions that can be linked to the idea of process tracing—for example, the work on mechanisms of Tilly (Reference Tilly2001) and McAdam, Tarrow, and Tilly (Reference McAdam, Tarrow and Tilly2001).

2 Within-case analysis can become multi-case analysis if different facets of the initial “case” are analyzed. The key idea here is that the point of departure is a single case, when viewed from the perspective of a wider comparative analysis focused on a larger N.

3 George (Reference George and Lauren1979); George and McKeown (Reference George, McKeown, Coulam and Smith1985); George and Bennett (Reference George and Bennett2005); Bennett (Reference Bennett, Box-Steffensmeier, Brady and Collier2008, Reference Bennett, Brady and Collier2010). On “soaking and poking,” see Fenno (Reference Fenno1977, 884; Reference Fenno1978, xiv; Reference Fenno1998, v). Process tracing has much in common with Lazarsfeld's (Reference Lazarsfeld and Komarovsky1940, preface) procedure of “discerning”; Campbell's (Reference Campbell1975, 181–82) “pattern matching,” which is also advocated by Yin (1984/Reference Yin2008); Sewell's (Reference Sewell and McDonald1996, 261) “causal narrative”; Bates et al.'s (Reference Bates, Greif, Levi, Rosenthal and Weingast1998) “analytic narratives”; and Hall's (Reference Hall, Mahoney and Rueschemeyer2003, 391–95) “systematic process analysis.”

5 Ideas about these designs based on a matching of cases are often drawn from J.S. Mill (Reference Mill1974) and Przeworski and Teune (Reference Przeworski and Teune1970). For a comment on the weakness of these designs for causal inference, see Collier, Brady, and Seawright (Reference Collier, Brady, Seawright, Brady and Collier2010a, 10).

6 Addressing this question raises issues about the logic of inquiry and the form of social scientific knowledge that are well beyond the scope of this discussion. Only a few basic points are addressed here that are salient for the accompanying exercises.

7 Obviously, such prior knowledge is essential in all research, both qualitative and quantitative.

8 Waltz calls claims about these regularities “law-like statements” (p. 1). We prefer the alternative label used here.

9 The expressions “descriptive inference” and “causal inference” are employed here in the sense of King, Keohane, and Verba (Reference King, Keohane and Verba1994, 7–8, chaps. 2–3). Their usage can be seen as approximating an ordinary language meaning of “description” and “causation”; and by “inference” they mean that researchers have “the goal of making inferences that go beyond the particular observations collected,” that is, they are analyzed within the larger framework used by the investigator. This usage contrasts with ideas of “descriptive inference” and “statistical inference” that are standard in the work of statisticians (e.g., Berk Reference Berk2004, chap. 11).

10 Achieving good description in this sense, and developing fruitful ideas about the unfolding of the process, may of course interact in an iterative manner.

11 Tannenwald's study is also discussed in Collier, Brady, and Seawright (Reference Collier, Brady, Seawright, Brady and Collier2010a, 189–90; Reference Collier, Brady and Seawright2010b, 509).

12 Lerner's analysis—which is the focus of one of the exercises—is closely tied to modernization theory, which might concern some readers; and at certain points the presentation seems condescending. Further, the female interviewer is presented in a sexist way (although in survey research, selecting interviewers in light of characteristics such as these is widely recognized as important). However, these drawbacks are outweighed by the opportunity presented by the chapter to illustrate the practice of making careful observations, and also to see how they can be integrated into a complex picture of social change.

13 For a framing of qualitative vis-à-vis quantitative in terms of four dimensions, see Collier, Brady, and Seawright (Reference Collier, Brady, Seawright, Brady and Collier2010a, 177–82).

14 Another puzzle is explaining the disappearance of the horse, but as Holmes himself emphasizes, that is a secondary issue (see p. 11 in the accompanying online version of the story).

15 See Levy's (Reference Levy, Box-Steffensmeier, Brady and Collier2008) excellent discussion of counterfactuals and case studies.

16 There is a parallel here to the idea in statistical work that the test does not stand on its own, but rather is shaped by prior assumptions. In quantitative analysis, the construction of the statistical model depends heavily on such assumptions, and in general the statistical test does not directly evaluate these assumptions. Rather, it estimates the relationship based on the supposition that the model assumptions, as well as the underlying assumption of causality, are true. See, for example, Freedman (Reference Freedman2010b).

17 In one story, Sherlock Holmes takes a strong stand on coincidences (“Adventure of the Second Stain”; in Doyle Reference Doyle1960, this is on p. 655). Watson refers to the juxtaposition of two key events as “an amazing coincidence.” Holmes replies: “A coincidence! The odds are enormous against its being a coincidence. No figures can express them. No, my dear Watson, the two events are connected—must be connected. It is for us to find the connection.” Ironically, it turns out that these two events are only tangentially connected, so Watson's statement was closer to the truth than Holmes's, and the weaker assumption was more appropriate. In another story, Holmes is initially more cautious about inferences and coincidences, but then he backtracks and insists on the certitude of his inferences (“The Sign of Four,” chap. 1; in Doyle Reference Doyle1960, this is on p. 93).

18 “The Sign of Four,” chap. 6 (in Doyle Reference Doyle1960, this is on p. 111).

19 A further perspective on unusual or bizarre explanations (see again table 6, H8, inference d) derives from William James's famous dictum that “every difference must make a difference.” To put this in a less extreme form, it might be said that some differences make a difference. In this instance, the form of the murder was so distinctive that it called for a distinctive explanation—which turned out to be the kick of a horse. On William James, see Copi (Reference Copi1953, 331–32).

20 Andrew Bennett (personal communication) has underscored the parallel here with diagnostic tests in medicine.

21 Waldner (Reference Waldner2011) offers an interesting discussion of such issues.

References

Bates, Robert H., Greif, Avner, Levi, Margaret, Rosenthal, Jean-Laurent, and Weingast, Barry. 1998. Analytic Narratives. Princeton, NJ: Princeton University Press.Google Scholar
Bennett, Andrew. 2008. “Process Tracing: A Bayesian Perspective.” In The Oxford Handbook of Political Methodology, ed. Box-Steffensmeier, Janet M., Brady, Henry E., and Collier, David, 702–21. New York: Oxford University Press.Google Scholar
Bennett, Andrew. 2010. “Process Tracing and Causal Inference.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 207–19. Lanham, MD. Rowman and Littlefield.Google Scholar
Berk, Richard A. 2004. Regression Analysis: A Constructive Critique. Newbury Park, CA: Sage.CrossRefGoogle Scholar
Brady, Henry E. 2010. “Data-Set Observations versus Causal-Process Observations: The 2000 U.S. Presidential Election.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 237–42. Lanham, MD: Rowman and Littlefield.Google Scholar
Brady, Henry E., Collier, David, and Seawright, Jason. 2010. “Refocusing the Discussion of Methodology.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 1531. Lanham, MD: Rowman and Littlefield.Google Scholar
Campbell, Donald T. 1975. “‘Degrees of Freedom’ and the Case Study.” Comparative Political Studies 8 (2) : 178–93.CrossRefGoogle Scholar
Collier, David, Brady, Henry E., and Seawright, Jason. 2010a. “Sources of Leverage in Causal Inference: Toward an Alternative View of Methodology.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 161–99. Lanham, MD: Rowman and Littlefield.Google Scholar
Collier, David, Brady, Henry E., and Seawright, Jason. 2010b. “Outdated Views of Qualitative Methods: Time to Move On.” Political Analysis 18 (4) : 506–13.CrossRefGoogle Scholar
Copi, Irving M. 1953. Introduction to Logic. New York: Macmillan.Google Scholar
Doyle, Arthur Conan. 1960/1986. The Complete Sherlock Holmes, Vol. 1 and 2. New York: Doubleday. The 1986 edition (New York: Bantam Classics) is available in paperback, and all the stories are readily available online.Google Scholar
Fenno, Richard F. Jr. 1977. “U.S. House Members in Their Constituencies: An Exploration.” American Political Science Review 71 (3): 883917.CrossRefGoogle Scholar
Fenno, Richard F. Jr. 1978. Home Style: House Members in their Districts. Boston: Little, Brown.Google Scholar
Fenno, Richard F. Jr. 1998. Senators on the Campaign Trail: The Politics of Representation. Norman: University of Oklahoma Press.Google Scholar
Freedman, David A. 2010a. “On Types of Scientific Inquiry: The Role of Qualitative Reasoning.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 221–36. Lanham, MD: Rowman and Littlefield.Google Scholar
Freedman, David A. 2010b. Statistical Models and Causal Inference: A Dialogue with the Social Sciences. New York: Cambridge University Press.Google Scholar
Garfinkel, Harold. 1981. Forms of Explanation. Rethinking the Questions in Social Theory. New Haven, CT: Yale University Press.Google Scholar
George, Alexander L. 1979. “Case Studies and Theory Development: The Method of Structured, Focused Comparison.” In Diplomacy: New Approaches in History, Theory and Policy, ed. Lauren, Paul Gordon, 4368. New York: The Free Press.Google Scholar
George, Alexander L., and Bennett, Andrew. 2005. Case Studies and Theory Development in the Social Sciences. Cambridge, MA: MIT Press.Google Scholar
George, Alexander L., and McKeown, Timothy J.. 1985. “Case Studies and Theories of Organizational Decision Making.” In Advances in Information Processing in Organizations, Vol. 2. ed. Coulam, Robert F. and Smith, Richard A., 2158. Greenwich, CT: JAI Press.Google Scholar
Hall, Peter A. 2003. “Aligning Ontology and Methodology in Comparative Research.” In Comparative Historical Analysis in the Social Sciences, ed. Mahoney, James and Rueschemeyer, Dietrich, 373404. New York: Cambridge University Press.CrossRefGoogle Scholar
King, Gary, Keohane, Robert O., and Verba, Sidney. 1994. Designing Social Inquiry: Scientific Inference in Qualitative Research. Princeton, NJ: Princeton University Press.CrossRefGoogle Scholar
Lazarsfeld, Paul F. 1940. “Introduction.” In The Unemployed Man and His Family, ed. Komarovsky, Mirra, ix–xii. New York: Dryden.Google Scholar
Lerner, Daniel. 1958. The Passing of Traditional Society: Modernizing the Middle East. New York: The Free Press. A somewhat abridged version, which unfortunately omits some key details, is reprinted in Development: A Cultural Studies Reader, ed. Susanne Schech, and Jane Haggis (Oxford: Blackwell, 2002). This abridged version is available online in Google Books at: http://tinyurl.com/Lerner-Grocer-Chief.Google Scholar
Levy, Jack S. 2008. “Counterfactuals and Case Studies.” In The Oxford Handbook of Political Methodology, ed. Box-Steffensmeier, Janet M., Brady, Henry E., and Collier, David, 627–44. New York: Oxford University Press.Google Scholar
Luebbert, Gregory M. 1991. Liberalism, Fascism, or Social Democracy: Social Classes and the Political Origins of Regimes in Interwar Europe. New York: Oxford University Press.CrossRefGoogle Scholar
Mahoney, James. 2010. “After KKV: The New Methodology of Qualitative Research.” World Politics 62 (1): 120–47.CrossRefGoogle Scholar
McAdam, Doug, Tarrow, Sidney, and Tilly, Charles. 2001. Dynamics of Contention. New York: Cambridge University Press.CrossRefGoogle Scholar
Mill, John Stuart. 1974. A System of Logic. Toronto: University of Toronto Press.Google Scholar
Przeworski, Adam, and Teune, Henry E.. 1970. The Logic of Comparative Social Inquiry. New York: Wiley.Google Scholar
Rogowski, Ronald. 2010. “How Inference in the Social (but Not the Physical) Sciences Neglects Theoretical Anomaly.” In Rethinking Social Inquiry: Diverse Tools, Shared Standards, 2nd ed., ed. Brady, Henry E. and Collier, David, 8997. Lanham, MD: Rowman and Littlefield.Google Scholar
Schultz, Kenneth A. 2001. Democracy and Coercive Diplomacy. New York: Cambridge University Press.CrossRefGoogle Scholar
Sewell, William. 1996. “Three Temporalities: Toward an Eventful Sociology.” In The Historic Turn in the Human Sciences, ed. McDonald, Terrence J., 245–80. Ann Arbor: University of Michigan Press.Google Scholar
Skocpol, Theda, Gans, Herbert, and Munson, Ziad. 2000. “A Nation of Organizers: The Institutional Origins of American Civic Life.” American Political Science Review 94 (3) : 527–46.CrossRefGoogle Scholar
Tannenwald, Nina. 1999. “The Nuclear Taboo: The United States and the Normative Basis of Nuclear Non-Use.” International Organization 53 (3): 433–68.CrossRefGoogle Scholar
Tilly, Charles. 2001. “Mechanisms in Political Processes.” Annual Review of Political Science 4: 2141.CrossRefGoogle Scholar
Van Evera, Stephen. 1997. Guide to Methods for Students of Political Science. Ithaca, NY: Cornell University Press.Google Scholar
Vineberg, Susan. 1996. “Eliminative Induction and Bayesian Confirmation Theory.” Canadian Journal of Philosophy 26 (June): 257–66.CrossRefGoogle Scholar
Waldner, David. 2011. “Process Tracing: Its Promise and Its Problems.” Paper presented at the Institute for Qualitative and Multi-Method Research, Syracuse University.Google Scholar
Waltz, Kenneth N. 1979. Theory of International Politics. New York: McGraw-Hill.Google Scholar
Weaver, Vesla M. 2007. “Frontlash: Race and the Development of Punitive Crime Policy.” Studies in American Political Development 21 (Fall): 230–65.CrossRefGoogle Scholar
Yin, Robert K. 1984/2008. Case Study Research: Design and Methods, 1st and 4th ed. Beverly Hills, CA: Sage.Google Scholar
Figure 0

Table 1 Process Tracing Tests for Causal Inference

Figure 1

Table 2 Overview of “Silver Blaze”

Figure 2

Table 3 Straw-in-the-Wind Tests

Figure 3

Table 4 Hoop Tests

Figure 4

Table 5 Smoking-Gun Tests

Figure 5

Table 6 Building a Doubly Decisive Test by Evaluating Alternative Hypotheses

Figure 6

Table 7 Auxiliary Outcome

Supplementary material: PDF

Collier supplementary material

Examples and Exercises

Download Collier supplementary material(PDF)
PDF 265.1 KB