Academia.eduAcademia.edu

Social Mechanisms and Explanatory Relevance

2011, Analytical Sociology and Social Mechanisms (edited by P. Demeulenaere)

In this paper I will discuss mechanistic explanation in the social sciences from the point of view of the philosophical theory of explanation. My aim is to show that the current accounts of mechanistic explanation do not serve the agenda of analytical social science as well as they should. I will not challenge the idea that causal explanations in the social sciences involves mechanisms or that social scientists should seek causal explanations and a mechanistic understanding of social phenomena, but will argue that to improve explanatory practices in the social sciences analytical social scientists should employ tools more substantial than the metaphor of a mechanism.

Social Mechanisms and Explanatory Relevance Petri Ylikoski Department of History and Philosophy University of Tampere & Philosophy of Science Group University of Helsinki [email protected] In this paper I will discuss mechanistic explanation in the social sciences from the point of view of the philosophical theory of explanation. My aim is to show that the current accounts of mechanistic explanation do not serve the agenda of analytical social science as well as they should. I will not challenge the idea that causal explanations in the social sciences involves mechanisms or that social scientists should seek causal explanations and a mechanistic understanding of social phenomena, but will argue that to improve explanatory practices in the social sciences analytical social scientists should employ tools more substantial than the metaphor of a mechanism. I will begin by presenting the basics of the erotetic approach to explanation and the notion of explanatory understanding. Section 2 argues that mechanisms do have many important roles related to explanation, but that they do not provide a solution to the problem of explanatory relevance. Section 3 introduces the idea of a contrastive explanandum and argues that paying attention to the identity of the explanandum is a necessary condition for proper assessment of explanatory claims. Section 4 argues that explanatory relevance should be understood as counterfactual dependence and that a manipulationist account of causation provides a fruitful framework for making sense of causal explanations. Section 5 discusses some of the consequences of these ideas for mechanistic explanation in the social sciences. 1 The purpose of the theory of explanation 1 The aim of the theory of explanation is to make sense of explanations. It addresses questions such as following: What kind of an activity is explanation and how is it related to other epistemic (and practical) activities? What kinds of explanations are there and what are their relationships to each other? By which criteria are explanations evaluated by scientists and by which criteria should they be evaluated? All of these questions circle around the central question: What constitutes an explanation? A convenient way to approach this question is to think about explanatory relevance. The problem of explanatory relevance is one of the major challenges for the covering-law account of explanation. The examples of men taking contraceptive pills explaining them not becoming pregnant, and hexing explaining why salt dissolves in water, are counterexamples of the covering-law account because they involve intuitively explanatorily irrelevant factors while fulfilling all the criteria of a satisfactory covering-law explanation. These counterexamples and other arguments have led philosophers of science to conclude that the covering-law account is a failure. It does not solve the central problem for any theory of explanation: the problem of explanatory relevance. This problem is not easy to solve. For example, Wesley Salmon’s causal theory of explanation falls victim to exactly the same counterexamples – it is not enough that we demand that an explanation only provides some information about the causal process, we want to have relevant information (Hitchcock 1995). The problem of explanatory relevance can be understood as a problem of explanatory selection (Hesslow 1983). Causal explanation provides information about causal history, but not all information about that history is regarded as explanatory. We have to pick the right aspects of the causal process to be included in the explanation. That is to say, how far in the causal history should we reach? How do we choose the events to be included in the explanation? How do we choose the right level of abstraction for describing these events? In how much detail should the events be described and which of their details should be included in the description? Apparently 2 we somehow manage to solve these problems intuitively when we are constructing explanations, but it would be much better if we could make the principles governing these judgments explicit. This is the main challenge for the theory of explanation. The theory of explanation should not be a purely theoretical enterprise; the ultimate motive is to develop conceptual tools for improving explanatory practices in the sciences (and in everyday life). The theory of explanation should have practical relevance and its success should be judged by the improvements it makes possible in explanatory practices. Such contributions are especially important in the social sciences where controversies about explanation are common. The social mechanisms movement is motivated by similar practical goals. While it has provided philosophical arguments for mechanistic explanation, the aims have been ultimately practical. The improvements in social science explanatory practice are the final evaluation criteria for philosophical arguments about explanation and causation. In this paper I will outline an account of explanation that in my judgment best advances the aims of both the philosophical theory of explanation and the social mechanisms movement. I call it the contrastive counterfactual account of explanation. The first element of this account is the erotetic approach to explanatory inquiry. Explanations as answers to questions Most philosophers of explanation agree that explanations are answers to questions. Some, like Hempel (1965), have used it as an informal starting point for their discussion, whereas others (van Fraassen 1980, Achinstein 1983) have built their theories of explanation around it. The latter approaches are commonly called erotetic approaches to explanation and are often associated with the broader question-theoretical account of scientific research. In the erotetic approach, scientific enquiry is regarded as a process of answering and elaborating questions. Empirical research is oriented towards fact-finding questions: it aims to answer what, 3 when, where, and how much questions. The relevance of these questions (and the required precision of the answers) is determined by practical how to questions and by explanatory why and how questions. The insight provided by the erotetic approach is based on the analysis of interrelations between different questions and its description of the research process as an organized series of questions. For example, the research process often requires that big explanatory questions be sliced to a series of smaller ones that can be addressed through empirical inquiry. Explanation and understanding In the erotetic approach, explanation is an answer to an explanation-seeking question. The explanation is regarded as complete when it fully answers the given question. This makes the notion of explanation quite narrow. To capture the broader goal of epistemic activities, we need another notion. I suggest that this be understanding (Ylikoski 2009). We are interested in finding answers to explanation-seeking questions because we wish to have knowledge about the dependencies governing the world. In other words, the goal of explanation is to understand these dependencies. Wittgenstein argued that understanding should not be understood as a sensation, an experience, or a state of mind. Understanding is not a special moment or phase, but a more permanent attribute. It is an ability (Wittgenstein 1953, §§ 143-159, 179-184, 321-324). I agree: when a person understands something, she is able to do certain things, which does not mean that understanding is some sort of special skill. Understanding consists of knowledge about relations of dependence. When one understands something, one can make all kinds of correct inferences about it. Many of these inferences are counterfactual: What would have happened if certain things had been different? What will happen if things were to be changed in a certain manner? Thus the fundamental criterion according to which understanding is attributed is the ability to make inferences to counterfactual situations, the ability to answer contrastive what-if-things-had-been-different questions 4 (what if questions, for short) by relating possible values of the explanans variables to possible values of the explanandum variable (Ylikoski 2009). The present account ties together theoretical and practical knowledge: they are not completely different notions. For example, in the case of causal explanation, explanatory understanding is crucial to our pragmatic interests, since answers to what if questions concerning the effects of possible interventions enable us to predict the effects of manipulation. Whereas the DN model and the associated epistemic conception of explanation conceive of the possessor of understanding as a passive observer of external events, the contrastive counterfactual theory links our theoretical practices to our roles as active agents (Woodward 2003). The degree of understanding conveyed by an explanation can be defined as the number and importance of counterfactual inferences that the explanatory information makes possible. Why do many people think that understanding is a mental state or an experience? The reason is that there exists a mental experience that is closely related to understanding: the sense of understanding. It is a feeling that tells us when we have understood or grasped something. This sense of confidence (and the feeling that often comes with it) can be easily confused with what we think it indicates: understanding. Ideally these two things would go hand in hand, and assimilating them should not create any trouble. However, real life is different. The sense of understanding is a highly fallible indicator of understanding. Sometimes one has a false sense of understanding and sometimes one understands without having any associated feelings or experiences. The sense of understanding does not give us direct access to knowledge that is the basis of our understanding, so it would be highly surprising if the sense of understanding would turn out to be perfectly calibrated to our understanding. The fallibility of the sense of understanding can be demonstrated experimentally. People often overestimate the detail, coherence, and depth of their understanding (Rozenblit and Keil 2002; Ylikoski 2009). 5 The existence of the sense of understanding should not be regarded as any kind of oddity. It plays an important metacognitive role in our cognitive life. It gives us confidence to try things, and when it is lacking we can sensibly abstain from the activity in question. It also guides the search for new knowledge, and tells us when to stop the search for new information; it signals when we know enough. In addition, the sensation associated with the sense of understanding can have a motivational role. Satisfying curiosity is highly rewarding (Schwitzgebel 1999, Gopnik 2000). It provides motivation for learning and other cognitive activities, and for this reason has an important role in human cognition. The desire to satisfy one’s curiosity also provides important psychological motivation for conducting scientific research (Ylikoski 2009). The phenomenology of the sense of understanding can mislead one into thinking that understanding is an on-off phenomenon (‘Now I got it!’). This is not the case. First, the understanding can be about different aspects of the phenomenon. Second, these aspects may be understood in various degrees. Consider an ordinary object like a personal computer. Different individuals understand to varying degrees how their computer works. Some might know about the software, or some specific piece of software, and others the hardware. Most people just use the software without any understanding of a computer’s internal workings. Despite these differences, they all understand something about their PC. The crucial question is which aspects of it they understand. A comparison of their understanding is possible, but strict assessment of overall understanding is complicated: there are many dimensions to compare. However, the important point is that by asking what has been understood, the extent of understanding can always be specified (Ylikoski 2009). This notion of understanding can be used to make sense of social science explanations. For example, we can ask what kind of what if questions does a certain account of social mechanism answer. Answering this question can be difficult, but this difficulty only emphasizes the contrast 6 between our confidence that it provides explanatory insight and our inability to articulate (or to agree about) what it explains. Surely this kind of exercise would benefit the aims of analytical social science. We can also raise the more general question of whether the social theorist’s sense of understanding is well calibrated to his abilities to make correct counterfactual inferences about social phenomena he is studying. I would be highly surprised if it were to turn out that widespread explanatory overconfidence didn’t exist among social theorists. 2 Mechanisms and explanatory relevance The idea of a mechanism has many uses in the philosophy of science. Most of these ideas also figure in social science discussions about explanation (Hedström & Ylikoski 2010). Here are four different ideas about the contribution of mechanisms to explanatory understanding. The first idea is about heuristics. According to it, existing mechanistic explanations can serve as templates, schemes or recipes for the search of causes. The knowledge about a possible mechanism tells the researcher what to look for and where. This simplifies the search for causes, especially in situations where one can be confident that the menu of possible explanatory mechanisms covers all the plausible alternatives. The second idea is related to justification of causal claims. It posits that the knowledge about possible mechanisms can provide support for causal claims. Causal claims without an account of the underlying mechanisms are possible and in principle fully legitimate, but knowledge of a mechanism makes them much more secure. This idea originates from everyday thinking, but it is also accepted in scientific contexts. Of course, the idea can sometimes be misleading: an imagined causal mechanism can give false credence to spurious causal claims. The third idea is about the presentation of explanatory information. A mechanism scheme provides useful means for presenting and organizing explanatory information. A narrative form 7 makes the explanatory information more digestible for humans, and mechanism schemes can be regarded as templates for such narratives. They outline the central features of the explanatory narrative and help people to focus on the right pieces of information. The fourth idea concerns the organization of social scientific knowledge. According to an old empiricist view, general knowledge in science consists of empirical generalizations and more abstract theoretical principles from which these generalizations can (ideally) be deduced. The mechanistic account of knowledge challenges this picture on two counts. First, the locus of generality (and explanatory power) in social scientific knowledge is considered to lie in the mechanism schemes. The social sciences do not have that many valid empirical generalizations and those that they have are not very explanatory. On the contrary, they are among the things that require explanation (Cummins 2000). Explanatory power – and general applicability – comes from knowledge of possible causal mechanisms. When social scientific knowledge expands, it does not do so by formulating empirical generalizations that have broader application, but by adding or improving items in its toolbox of possible causal mechanisms. This brings us to the second challenge to the traditional picture of social science knowledge: the ideal of knowledge in no longer an axiomatic system, but a much looser idea of an expanding theoretical toolbox. The expectation is that mature social science would be more like a textbook of modern cell biology than a treatise in elementary geometry. These are important ideas – and I subscribe to all of them – but they do not address the issue of explanatory relevance. They do not tell us what is the explanatory import of mechanisms. Nor do they tell us how to construct a good or satisfactory mechanistic explanation. Furthermore, they provide no guidance for the evaluation of suggested mechanistic explanations. The insufficiency of mechanistic ideas 8 To see why the idea of mechanisms is not very helpful in dealing with the problem of explanatory relevance, we have to understand how this notion works. Let us begin by distinguishing two ways to talk about mechanisms. Examples of both are plentiful in the literature and most of the time people seem to assume that they incorporate the same notion of mechanism. The first (let us call it A-mechanism) regards mechanisms as a particular causal chain. The mechanism is whatever connects the cause and effect. No matter how long or complicated the causal process is, it can be called a mechanism if its description answers the question how did the cause bring about the effect? The second way (B-mechanism) to talk about a mechanism regards it as a theoretical building block. Here the mechanism is an object of theoretical interest, and it often applies only to simplified and idealized explananda. The above mechanistic idea about the organization of general knowledge is based on this notion of mechanism. The differences between these two notions can be seen more easily if we consider their interrelations. The first thing to observe is that a single A-mechanism can involve many B-mechanisms. A number of different B-mechanisms can work serially or simultaneously in the same causal process. Furthermore, nothing prevents B-mechanisms from working in opposite directions. For example, if we are explaining the rise in the murder rate after the collapse of a corrupt central government, some of the B-mechanisms could work towards a reduction of crime, whereas others would increase it. The explanandum of an A-mechanism is the total (or net) effect, whereas the explanandum of a B-mechanism is more like a component effect. Clearly we should not use these two notions of mechanism interchangeably. The idea of an A-mechanism builds upon the idea that the knowledge of the details of a causal process makes the explanation better. This idea feels intuitively correct. People prefer detailed explanations to mere sketches of explanation. This preference shows that detail is regarded as a virtue in explanatory contexts. The crucial question here is what do we mean by more detailed? 9 An explanation is more detailed when it omits less of the relevant information. The key word here is relevant. The details of the causal process can be quite harmful for the explanation if they are irrelevant from the point of view of the explanandum. There has been some discussion in the philosophy of science about whether the addition of irrelevant details makes the explanation completely nonexplanatory or whether it just makes an explanation worse (see Salmon 1998). We do not need to take a stance on this partly verbal issue; it is sufficient to point out that irrelevant details at least decrease the cognitive salience of the explanation. Irrelevant details can mislead: in such cases we identify wrong factors as explanatory. A more detailed causal story might also be more difficult to grasp. This is a simple fact of human cognition: our memory and ability to focus are limited and burdening them restricts our inferential performance. In both cases our ability to answer what if questions decreases. In other words, we will understand less. Although the notion of an A-mechanism is important, it does not provide much insight into the nature of explanation. It is a placeholder notion: whatever explains the fact that c caused e is the mechanism. In this sense it is of limited analytical value: it names the challenge, but does not provide tools for dealing with the problem of explanatory relevance. Nor does the notion of an Amechanism give any guidance for constructing mechanistic explanations. It does not tell us which level of organization we should focus on, which elements of the process should be incorporated in to the mechanism, or how detailed the description of these items should be. The identity criteria for B-mechanisms are stricter, but this notion faces the opposite problem. Bmechanisms are in many ways analogical to component causes. Just like component causes, they can be involved in cases of overdetermination, in which more than one mechanism works to guarantee a certain outcome. (Sometimes they could work simultaneously, sometimes one mechanism could preempt another.) Similarly, they could be involved in the cases of counteracting causes, in which the mechanisms compete by having opposite causal effects. Finally, as with 10 an individual component cause, a B-mechanism can be simply explanatorily irrelevant. In the case of an A-mechanism the connection to the explanandum is guaranteed by definition, but such a guarantee is lacking in the case of a B-mechanism. Any causal process involves many Bmechanisms at various levels of organization (physical, chemical, etc.); it is an open question whether they are in any way relevant to the explanandum we are interested in. Here again we find the problem of explanatory relevance: we need some guidance with B-mechanisms. What is the appropriate level of organization to focus on? What makes a given B-mechanism relevant to the explanandum? In how detailed a manner we should describe the mechanism? When people construct and evaluate mechanistic explanations they are intuitively making judgments about explanatory relevance. Everybody agrees that good mechanistic explanations capture the relevant aspects of the causal process. However, nothing in the idea of a mechanism helps us in making and evaluating these judgments – we have to trust our intuitions. The trouble with intuitive assessments of explanations is that they are based on unreliable metacognitive processes. The sense of understanding is highly unreliable, and this shows in the disagreements people have in their judgments about explanatory value. Given the intuitive appeal of mechanistic storytelling in everyday reasoning, the critics of mechanistic theories of explanation might be right in being skeptical about their value. Especially when we are working with highly abstract sketches of mechanisms (as is typical in the social sciences), the danger of the illusion of depth of understanding is a real one. Clearly, the notion of a mechanism is not a sufficient tool for improving explanatory practices in the social sciences. It needs to be supplemented with additional ideas from the theory of explanation. I will now turn to these ideas and try to show how they can be used to make sense of mechanistic explanation. 11 3 The importance of the explanandum The problem of explanatory relevance in causal explanation is that the causal history of an event includes a vast number of elements and aspects that are not explanatorily relevant to the explanation-seeking question we are addressing. A natural way to start sorting out this problem is to take a closer look at the explanandum. A contrastive account of explanation is helpful here. According to this view, explanations are answers to questions in the following form: Why the fact rather than the foil happened? We want to know why things are one way rather than some other way. An anecdote about the famous bank robber Willie Sutton illustrates the basic idea of contrastive explanation. When Sutton was in prison, a journalist asked him why he robbed banks. Sutton answered, “Well, that’s where the money is.” The journalist had in his mind the question: “Why do you rob banks, instead of leading an honest life?,” whereas Sutton answered the question: “Why do you rob banks, rather than gas stations or grocery stores?” This is the basic insight of the contrastive approach. We do not explain simply ‘Why f?,’ rather, our explanations are answers to the contrastive question ‘Why f rather than c?’ (Garfinkel 1981, 21-22). Instead of explaining plain facts, we are explaining contrastive facts (Ylikoski 2007). Sometimes the contrast is imagined: we ask why an object has a particular property rather than a different property we expected it to have or our theory predicted it would have. Sometimes the contrast arises from a comparison: we ask why an object has a certain property rather than being like an otherwise similar object that has a different property. In both cases we are explaining a difference: why a fact is the case rather than its exclusive alternative. I will adopt the convention of expressing the contrast in the following manner: fact [foil], which should be read as fact rather than foil. The number of foils is not limited; there can be more than one. The contrastive idea does not put strict limitations on kinds of facts that can serve as the ex- 12 plananda: they can belong to different ontological categories. They can be properties, events, quantities or qualities. The crucial thing is that the fact and its foil should be exclusive alternatives to each other. To cover the wide variety of possible ontological categories, I will here simply speak about variables. This terminology does not commit us to any specific ontology, but allows us to make our points more generally and to avoid messy philosophical debates about the nature of events and other ontological categories. The basic idea is that whatever the real relata of explanation are, they can be represented as variables. The idea of contrastive explanandum is a practical tool for analyzing explananda and for making the intended explanandum more explicit. Although it is not always apparent from the surface appearance of explanations, all explanation-seeking questions can be analyzed and further explicated by articulating the implicit contrastive structure of the explanandum. This explication makes the aims of explanation-seeking questions more clear and makes the evaluation of the adequacy of the proposed explanations easier. Articulating the contrastive structure forces one to be specific about the intended object of the explanation and about the respects in which we think the object could have been different (Ylikoski 2007). Are we explaining a singular event or a more general phenomenon or regularity? Are we addressing properties of individuals or of populations? What is the appropriate level of description: are we after micro-level details or patterns found at the macro level? Quite often the original question turns out to be a whole set of related contrastive questions. This is a good thing: smaller questions are something that we can realistically hope to answer by means of empirical enquiry. Contrastive articulation is also useful in controversies over apparently conflicting explanations. Often the apparently competing explanations turn out to be addressing complementary or completely independent questions. This is as it should be: we can be 13 pluralists about explanation-seeking questions, but whether the answers are correct is still an objective matter. The contrastive approach does not only help to clarify the explanandum, it can also be used to evaluate explanations. Rather than asking what the intended explanandum is, we can also ask what is the contrast that the given explanation (or piece of explanatory information) can explain? This is very useful approach, particularly, in social science controversies about explanation. Rather than getting entangled with difficult interpretive problems about what certain theorists are really attempting to explain, we can take a look at the explanations they provide and consider what they in fact explain (Ylikoski 2007). The contrastive idea underlies our preferred form of causal inquiry: the scientific experiment. We contrast the control group with the experimental group or the process after the intervention with the process before the intervention. In both of these cases, we are trying to account for the differences between the outcomes. The basic idea is to keep the causal background constant and to bring about changes in the outcomes by carefully controlled interventions. A similar contrastive setup motivates comparative research. In general, the idea that explanations are contrastive is natural if one thinks that the aim of explanation is to trace relations of dependency. Our goal is to understand how changes in the cause variable bring about changes in the effect variables. We want to know what makes the difference and then leave out the factors that do not. The contrastive idea is based on a very intuitive feature of our explanatory cognition. 4 Making a difference The starting point of the contrastive-counterfactual approach to explanation is realistic: explanations attempt to trace objective relations of dependence. These dependences can either be causal or constitutive. Dependencies are objective in the sense that they are independent of our 14 ways of perceiving and theorizing about them. They are also separate from our means of describing them. That is to say, the real relata of explanation are facts, not sentences. If the explanandum is contrastive, as I have suggested, it is natural to think that the explanans is also contrastive. Following this idea, philosophers of causation are increasingly thinking that causation is doubly contrastive (Woodward 2003; Schaffer 2005). The same idea applies also to causal explanation. In the simplest form, causal explanatory claims have the following structure: c [c*] explains e [e*] In plain language, the cause variable having the value c rather than c* explains why the explanandum variable has the value e rather than e*. A natural way to understand this relation is to regard it as a counterfactual dependence: if the cause variable had had the value c*, the value of the explanandum variable would have been e* rather than e. Note that although the claim is about counterfactual situation, there is nothing counter to the facts in the relation of dependence itself. Also of importance is that counterfactual dependence is a modal notion: explanation is not about subsumption under empirical regularities, but about counterfactual dependence. The idea of a counterfactual theory of causation is very old. We can avoid some of the historical problems of counterfactual theories of causation if we do not regard it as a reductive theory of causation and if we limit its application to causal explanation. The new idea is to combine it with the idea of contrastive explanandum. Together with a sophisticated manipulation account of causation this idea helps us to solve most of the problems that counterfactual theories of causation have faced (See Ylikoski 2001). An important problem for the counterfactual theories has been the specification of the truth conditions for the counterfactuals. A manipulation account of causation is helpful here. In the manipulation account of causation c [c*] causes e [e*] if we can bring about e* [e] by bringing about c* [c] 15 This cannot serve as a reductive analysis of causation, as ‘bringing about’ is already a causal notion. However, it can serve as a starting point for an adequate descriptive analysis of the notion of causation. An advantage of the manipulation account is that it provides a natural way of distinguishing between real causal relations and mere correlations. Real causal relations can be used as bases for effective interventions, whereas mere correlations do not allow this. Experiments and other causal interventions in the real world are often quite messy. To make sense of the meaning of causal claims we need the notion of an ideal intervention, developed by Woodward (2003, 94-151). An ideal intervention I changes the value of effect variable Y only via changing the value of cause variable X as shown below. (The formulation and the graph are adapted from Craver 2007.) U 1. I does not change Y directly 4 2. I does not change the value of any causal inX S Y termediate S between X and Y except by chang2 ing the value of X that is a cause of Y 4. I acts as a switch that controls the value of X C I 3. I is not correlated with some other variable C 1 3 causal relation absence of causal relation absence of correlation irrespective of X’s other causes U. The notion of ideal intervention gives us straightforward semantics for causal counterfactuals. It also makes it possible to avoid anthropocentric features of earlier manipulation accounts of causation: the definition does not refer to a human agent. Nor must we assume that the intervention is humanly possible – we are only interested in possible interventions. For us the main advantage of this account is that it provides us with a natural way of making sense of counterfactuals in 16 causal explanation, meaning that an explanatory claim is correct if ideal intervention on the explanans variable had brought about the appropriate change in the explanandum variable. An interesting feature of Woodward’s account is his view of explanatory generalizations about effects of ideal interventions (Woodward 2003, 239-314). These invariances lack many traditional features of laws of nature. They usually hold only for a limited range of possible interventions (and changes in background factors); they can refer to particular objects, places and times; and they might contain exceptions. In short, they might hold only in a certain domain and break up outside of it. However, in contrast to the traditional empiricist accounts of laws of nature, invariance is a modal notion: it makes a claim that the invariant relationship between cause and effect variables will hold under a set of possible interventions. According to Woodward, invariances capture what is essential from the point of view of explanation, whereas most of the traditional attributes are superficial from this angle. Consequently, in his account all explanatory generalizations are not laws. This is good news for the social sciences (and other sciences outside of fundamental physics): the generalizations satisfying the traditional criteria of lawhood are rare or nonexistent. It is plausible to think that our contrastive explanatory preferences stem from our nature as active interveners in natural, psychological, and social processes. We desire to know where to intervene to produce the changes we want, and this knowledge often presupposes answers to some why and how questions. Without this knowledge one would not know when the circumstances are suitable for an intervention and one would not be able to predict the results of the intervention. I do not wish to claim that the notion of explanation can be reduced to its origins in agency. After all, we are also interested in explaining things that cannot be humanly manipulated. However, our instrumental orientation might still explain why our explanatory preferences are as they are. 17 5 The explanatory import of mechanisms In the contrastive counterfactual account causal claims are based on change-relating invariances, meaning that no mechanism is needed for making a causal claim. This helps us to avoid the regress problem that an account which assumes that mechanisms are an integral part of all causal claims would face. However, it does not challenge the legitimacy of the assumption that outside of fundamental physics there will always be a causal mechanism mediating between the variables. Nor does it imply that we should not try to have knowledge about the mechanisms (Hedström & Ylikoski 2010). However, there is one important consequence: macro variables are explanatory in the same sense as micro variables. In other words, the explanatory factors (for explananda at individual or social levels) might be found at the social level. The crucial issue is whether the right kind of invariance exits between the explanans and the explanandum variables. If by changing the macro-variable we can bring about a relevant change in the explanandum variable, we have identified the cause. The intervention account does not privilege any specific level of description; it is a purely empirical matter which level provides the variables that inform us about the relevant invariances (Steel 2006). The crucial factor is the contrast on the explanandum side: depending on the contrast, a micro or macro variable might turn out to be more relevant (Ylikoski 2001, 77-100). The relevance is determined by the nature of invariances that the variables are involved in. This blocks the simple argument for methodological individualism through causal mechanisms. Causal claims do not have to incorporate mechanisms and they can refer to macro variables. Clearly, if we want to stick with methodological individualism, we need some independent arguments. If causal claims can be made without referring to mechanisms, the question arises what is their function in causal explanation? Is it a misguided idea that adds nothing to the process of causal explanation, as some critics have suggested? This would be an overreaction. It is important to recognize the limited scope of the above claims. These are claims about singular causal state18 ments that answer to simple explanation-seeking questions. However, when we turn our attention to the broader issue of understanding, the true contribution of mechanisms becomes apparent. A singular causal claim might be based on a simple claim about counterfactual dependence, but for theoretical understanding we need more. In the erotetic approach, explanation-seeking questions are not treated separately as independent entities. The questions come in clusters of closely related what if questions. The individual questions are related to each other in many ways. When we find that the value of variable Y is dependent on the value of variable X in a manner that allows (at least an ideal) intervention, this raises a whole series of questions. First, why does the counterfactual dependence hold? Second, what is the range of interventions that this invariance allows without breaking apart? Third, what are the background conditions for this invariance and how sensitive is the invariance to changes in these conditions? Fourth, are there other alternative interventions that can bring about the same changes in Y? Fifth, is it possible to account for a more precise explanandum? And sixth, is it possible to find a generalization that stays invariant in a broader range of interventions and background conditions? The knowledge of mechanisms is involved in answering all these questions. The questions are related to each other via chains of presupposition. As we can see from the first question, underlying every why question is a possible how question. Individual claims about causal influence are of limited interest. We want to know why X can cause changes in Y. Answering this question requires knowledge about causal mechanisms: we have to know how the changes in X are transmitted to changes in Y. The fourth question about alternative interventions is closely related and knowledge about possible mechanisms greatly facilitates answering it. Answers to both of these questions expand the range of what if questions we can answer about the phenomenon – they increase our understanding. 19 An important manner in which knowledge about mechanisms advances our understanding is integration. An answer to the first question integrates the causal claim with other pieces of knowledge. This is a very important consideration from the point of view of understanding. When an explanation is integrated into a larger theoretical framework, the theoretical connections can expand the range of answers to different what if questions in two ways. First, the integration allows for inferential connections to an already existing body of knowledge, and this might make it possible to find unforeseen dimensions in which what if questions concerning the explanandum phenomenon can be answered. Second, the explanation itself may bridge previous gaps within the existing theory and thus enable answers to new what if questions that do not directly concern the original explanandum phenomenon. (Ylikoski & Kuorikoski 2009.) The second and third questions arise from the fact that we want to know how broadly our causal knowledge can be exploited; in other words, we wish to know how many other what if questions can we answer with the same information. When we know more about the mechanism transmitting the causal influence, we have a better idea of what kinds of factors can disrupt the causal link and how. With this information, we can answer more what if questions concerning situations in which the background assumptions or conditions are different, for example in situations where some parts of the mechanism were altered or when the background factors not included in the explanatory generalization change. In this way, knowledge of mechanisms contributes to understanding the domain of application of the knowledge about the invariance. Without knowledge about the mechanisms we would have trouble evaluating the range of what if questions that can be answered with our piece of causal knowledge. Knowing the limits of one’s knowledge is an important ingredient of understanding. Finally, the fifth and sixth questions suggest that knowledge about the mechanisms might help us to improve the explanatory generalization. A common way to search for more general formula- 20 tions is to look at the underlying mechanisms and explore the possibility of integrating them into the explanatory generalization. This might make it possible to formulate the explanatory invariance in such a manner that it allows us to explain either a sharper explananda or a broader range of explananda than the original (Ylikoski & Kuorikoski 2009). It might also make it possible to formulate a more general statement of invariance that holds for a broader set of interventions and/ or background conditions. More general invariances are preferable, as they also make it possible to answer a broader set of what if questions. In both of these cases the road to better explanatory generalizations goes through mechanisms: they suggest variables that could be incorporated into the explanatory generalization (Ylikoski & Kuorikoski 2009). If these observations are correct, the call for mechanisms really makes sense. Knowledge about mechanisms means an ability to answer more what if questions, i.e. more understanding. When this explanatory contribution is combined with the four other functions of mechanisms mentioned earlier, we see that the social mechanisms movement has been motivated by the right kind of intuitions. However, it might be that some of the advocates of social mechanisms have had the wrong ideas about the source of legitimacy of these intuitions: the necessity of mechanisms does not derive from ontological arguments for methodological individualism or from semantics of individual causal claims. 6 Conclusion In this paper I have suggested that the current mechanistic ideas about explanation are insufficient for the purposes of analytical social science. The aim is to improve explanatory practices, but the notion of a mechanism is not very helpful. I have also suggested a number of ideas about explanation that can be used in improving social science explanations. The first is the erotetic approach to explanations combined with the idea that all explanation is contrastive. This idea is use21 ful in making explanations more explicit and by allowing one to be more precise about the intended explanandum, as it permits a much sharper evaluation of explanatory claims. It also makes it possible to think about interrelations between explanations by considering chains of presupposition. The second idea is the counterfactual criterion of explanatory relevance. When explanatory claims are understood as claims about possible (ideal) interventions, we get a notion of explanatory relevance that is both intuitive and powerful. Among other things, this approach allows for a novel way of understanding explanatory generalizations as claims about invariances. Third, I have suggested that we employ the notion of understanding to make sense of the contribution of the knowledge about mechanisms to our explanatory enterprise. A singular causal claim does not require knowledge about the mechanisms, but once we begin to consider the broader set of what if questions, the importance of mechanisms becomes apparent. The notion of understanding also helps us to see the unreliability of the intuitive way of evaluating the relevance of explanatory information. As the sense of understanding can be highly misleading, we have a further motivation for seeking a more explicit theory of explanation. Finally, I have suggested that we should decouple the idea of mechanistic explanation from the doctrine of methodological individualism. These two ideas should be argued separately, not as grounds for each other. 22 References Achinstein, Peter 1983: The Nature of Explanation. Oxford University Press. Oxford. Bechtel, William 2008: Mental Mechanism. Philosophical Perspectives on Cognitive Neuroscience. London: Routledge. Craver, Carl 2007: Explaining the Brain: Mechanisms and the Mosaic Unity of Neuroscience. Oxford: Clarendon Press. Cummins, Robert 2000: ”How Does It Work?” versus “What Are the Laws?”: Two Conceptions of Psychological Explanation, in Keil & Wilson (eds.) Explanation and Cognition, Cambridge: MIT Press: 117-144. Elster, Jon 1989: Nuts and Bolts for the Social Sciences. Cambridge: Cambridge University Press. Garfinkel, Alan 1981: Forms of Explanation. Yale University Press. New Haven. Gopnik, Alison 2000: Explanation as Orgasm and the Drive for Causal Knowledge: The Function, Evolution, and Phenomenology of the Theory Formation System, in Keil and Wilson (eds.) Explanation and Cognition. Cambridge: MIT Press: 299-324. Hedström, Peter & Richard Swedberg (eds.) 1998: Social mechanisms: An analytic approach to social theory. Cambridge: Cambridge University Press. Hedström, Peter 2005: Dissecting the Social. On the Principles of Analytical Sociology. Cambridge: Cambridge University Press. Hedström, Peter & Petri Ylikoski 2010: Causal Mechanisms in the Social Sciences, Annual Review of Sociology, forthcoming. Hempel, Carl 1965: Aspects of Scientific Explanation. New York: The Free Press. Hesslow, Germund 1983: Explaining differences and weighting causes, Theoria 49: 87-111. Machamer, Peter, Lindley Darden & Carl F. Craver 2000: Thinking About Mechanisms, Philosophy of Science 67: 1-25. Mayntz, Renate 2004: Mechanisms in the Analysis of Social Macro-Phenomena, Philosophy of the Social Science 34: 237-259. Pearl, Judea 2000: Causality – Models, Reasoning and Inference, Cambridge: Cambridge University Press. Rozenblit, Leonid, and Frank C. Keil 2002: The misunderstood limits of folk science: an illusion of explanatory depth, Cognitive Science 26: 521-562. Salmon, Wesley 1998: Causality and Explanation. Oxford: Oxford University Press. Schaffer, Jonathan 2005: Contrastive Causation, The Philosophical Review 114: 297-328. Steel, Daniel 2006: Methodological Individualism, Explanation, and Invariance, Philosophy of the Social Sciences 36: 440-463. Steel, Daniel 2008: Across the Boundaries. Extrapolation in Biology and Social Sciences. Oxford: Oxford University Press. Schwitzgebel, Eric 1999: Children’s Theories and the Drive to Explain, Science & Education 8: 457-488. 23 van Fraassen, Bas 1980: Scientific Image. Oxford: Oxford University Press. Wittgenstein, Ludwig 1953: Philosophical Investigations. Oxford: Basil Blackwell. Woodward, James 2003: Making Things Happen. A Theory of Causal Explanation. Oxford: Oxford University Press. Ylikoski, Petri 2001: Understanding Interests and Causal Explanation. Ph.D. thesis. University of Helsinki, May 2001. [http://ethesis.helsinki.fi/julkaisut/val/kayta/vk/ylikoski/] Ylikoski, Petri 2005: The Third Dogma Revisited, Foundations of Science 10, 395-419. Ylikoski, Petri 2007: The Idea of Contrastive Explanandum, in Persson and Ylikoski (eds.): Rethinking Explanation. Dordrecht: Springer, 27-42. Ylikoski, Petri 2009: Illusions in Scientific Understanding, in De Regt, Leonelli & Eigner (eds.) Scientific Understanding: Philosophical Perspectives, Pittsburgh University Press, 100119. Ylikoski, Petri & Jaakko Kuorikoski 2009: Dissecting Explanatory Power, Philosophical Studies, forthcoming. 24