Academia.eduAcademia.edu

Linking Policy Design and Implementation Styles

Howlett, Michael and Mukherjee, Ishani, eds. (2018) Routledge Handbook of Policy Design. London: Routledge, pp. 316-330

In this contribution, we bring together two strands of research that have so far existed side by side: the scholarship on policy design and the literature on policy implementation. Designing policies involves identifying goals and selecting policy instruments with which the goals can be reached. Policy implementation is about putting these policies into practice and comprises the definition of an implementation structure, decision-making within agencies, target group behavior, and policy results. We propose to connect the stages of policy design and policy implementation by means of the concept of policy feedbacks. Based on the two traditional approaches to studying policy implementation, we distinguish two ideal typical implementation styles. Policy delivery can be organized in a centrally controlled hierarchical manner, where fixed policy intentions are expected to be handed down to addressees with as little deviation as possible (centralized implementation), or it can be organized in a more decentralized manner which leaves implementing actors more leeway for adapting the policy to local circumstances (decentralized implementation). We argue that while decentralized implementation offers more insights into what makes policy work than centralized implementation, the latter is more likely to produce policy feedbacks that inform future policymaking and potential redesign of the policy concerned.

Linking Policy Design and Implementation Styles Jale Tosun and Oliver Treib Published in: Howlett, Michael and Mukherjee, Ishani, eds. (2018) Routledge Handbook of Policy Design. London: Routledge, pp. 316–330. Abstract In this contribution, we bring together two strands of research that have so far existed side by side: the scholarship on policy design and the literature on policy implementation. Designing policies involves identifying goals and selecting policy instruments with which the goals can be reached. Policy implementation is about putting these policies into practice and comprises the definition of an implementation structure, decision-making within agencies, target group behavior, and policy results. We propose to connect the stages of policy design and policy implementation by means of the concept of policy feedbacks. Based on the two traditional approaches to studying policy implementation, we distinguish two ideal typical implementation styles. Policy delivery can be organized in a centrally controlled hierarchical manner, where fixed policy intentions are expected to be handed down to addressees with as little deviation as possible (centralized implementation), or it can be organized in a more decentralized manner which leaves implementing actors more leeway for adapting the policy to local circumstances (decentralized implementation). We argue that while decentralized implementation offers more insights into what makes policy work than centralized implementation, the latter is more likely to produce policy feedbacks that inform future policymaking and potential redesign of the policy concerned. 1. Introduction Policy studies address a remarkably wide range of topics and cover a multitude of levels of analysis. The most micro-level research explores the behavior of individuals (e.g. John et al. 2009; Sager et al. 2014; Weaver 2015), whereas the most macro-level research searches for country-comparative patterns of policy decisions (e.g. Falkner et al. 2007; Falkner and Treib 2008; Damonte et al. 2014; Tosun 2013; Mortensen and Green-Pedersen 2015). At first glance, these two extreme ends of public policy do not seem to be conformable with one another, but upon closer inspection, the differences between them start to vanish. What connects these two research perspectives is the mutual interest in how the behavior of individual and collective actors in society can be steered to attain a given goal. An alternative viewpoint that connects the micro- and the macro-level is how past policies’ feedbacks shape future policies (e.g. Mettler and SoRelle 2017). The literature on policy feedbacks typically takes policy outcomes or impacts – here jointly referred to as policy results – as the starting point, but pays limited attention to how decisions regarding policy design or implementation 1 style may matter for producing such feedbacks. This gap is the point of departure for this chapter, in which we seek to connect the stages of policy design and policy implementation by means of the concept of policy feedbacks. Policy design involves two sets of decisions. First, decisions regarding the substance of a policy in terms of what it aims to achieve, and which instruments are conducive to the policy’s goal. Second, the procedure of a policy in terms of which level of government and agency is responsible for implementing the policy concerned (e.g., DeLeon 1988; Howlett and Rayner 2007; Howlett 2009; Howlett et al. 2015; Capano et al. 2016; Chindarkar et al. 2017). The latter also includes considerations regarding the extent to which the individual implementing agencies concerned have some leeway in adapting the policy to local-level conditions. A constellation where policymakers expect fixed policy intentions to be handed down to addressees with as little deviation as possible is referred to as centralized implementation. The alternative constellation is decentralized implementation, which is characterized by leaving the implementing actors more leeway for adapting the policy to local circumstances. The analytical considerations underlying these two implementation styles and their effects for policy delivery have been discussed widely in the literature (e.g. Pülzl and Treib 2007; Damonte et al. 2014; Saetren 2014; Hupe and Sætren 2015; Hupe and Hill 2016). We seek to make a novel contribution to the literature by contending that the choice between centralized and decentralized implementation has implications for the production of policy feedbacks and therefore the dynamics and potential results of subsequent decisions on policy design. This chapter unfolds as follows. First, we introduce the concepts on which we elaborate in this chapter. Next, we introduce our integrated conceptual model, which is followed by suggestions for research designs suitable to test the empirical implications of the model. In the concluding section, we summarize our main arguments and offer some ideas for future research. 2. Clarifications on the key concepts In this section, we offer definitions of concepts that are integral elements of our conceptual model. The concepts concerned are policy tools, policy design, policy implementation, and policy feedbacks. We discuss each of the concepts in turn. Policy tools Recent developments in policy studies make it necessary to position ourselves with regard to the types of policy tools we analyze here. In the last few years, a literature on ‘new’ policy tools 2 based on the insights of behavioral economics has emerged. Commonly referred to as ‘nudge policies’, from the title of the influential book by Thaler and Sunstein (2008), the research program underlying this type of policies corresponds to that of heuristics and biases as put forth by the path-breaking work of Kahnemann et al. (1982). Nudge policies aim to bring about behavioral change by capitalizing on systematic cognitive and behavioral biases and thereby directing individual behavior toward a more beneficial outcome (Grüne-Yanoff and Hertwig 2016: 152). The appeal of nudge policies for policymakers derives from the fact that they promise to bring about effective behavioral change without using the classical, heavy-handed repertoire of policy tools such as coercion or financial incentives. This promise, however, may only hold for a narrow set of policy problems. Often, incentives for non-compliance may be too high and citizens or economic actors may have other reasons than a lack of the appropriate choice architecture for behaving the wrong way, such as a lack of resources or deeply held beliefs (Weaver 2015). There are thus good reasons to hold on to the classical tools of government action. Even though nudge policies have been applied in several contexts (e.g., John et al. 2009), many problems may still be tackled effectively only by resorting to the coercive force of the law, to financial incentives or disincentives, or to direct public service provision (e.g. Hood 1986; Howlett and Ramesh 1993). Most policies still employ these classical techniques, which is why we limit our analytical purview to this type of policy tools (but see Strassheim and Beck 2019). There are several analytical schemes for classifying classical policy tools, among which the socalled NATO scheme is particularly popular. This scheme differentiates between policy instruments based on the principal governing resource they use: policies using information to convince norm addressees of a particular goal (Nodality); policies that employ the coercive force of the law (Authority); policies that provide financial incentives or disincentives (Treasure); and policies that build or reform institutions to achieve their objectives (Organization) (Hood 1986; Hood and Margetts 2007). Another way of classifying policy tools is to differentiate between substantive and procedural policies. Substantive policy tools are intended to directly affect the quality and the quantity of the goods and services provided in society. Procedural policy tools are more diverse than substantive policy instruments and include the provision of information as well as the creation and reform of institutions (see Howlett 2005: 35-36). 3 Our decision to stick to classical policy tools has important consequences for the subsequent analysis. With nudge policies, policy implementation predominantly requires the policy addressees to change their behavior. Consequently, the implementation process is rather straightforward and simple. With classical policies, the implementation process is more complex, involving behavioral changes by the target groups, but also decisions with regard to the overall implementation structure and the involvement of governmental and nongovernmental organizations. Policy design Research on policy design has concentrated on (bundles of) policy tools that public actors choose in order to bring about behavioral changes (see, e.g., DeLeon 1988; Howlett and Rayner 2007; Howlett 2009; Howlett et al. 2015; Capano et al. 2016; Chindarkar et al. 2017; Howlett and Mukherjee 2017). The main question underlying the relevant scholarship is how policies can be constructed in such a way that they produce the desired policy results (Howlett et al. 2015; Colebatch 2017). From this, it follows that research on policy design is tightly linked to both questions of policy success and implementation. In this context, Howlett (2011) argues that the birth of this specific literature in the mid-1980s was motivated by policy scholars’ wish to improve their knowledge of how the choice of policy tools affects the choice of implementation structures and tools (but see Sager and Rielle 2013). While there is a close relationship between implementation and the success of a policy, the literature on policy design also emphasizes that a policy may fail due to an invalid theory connecting policy actions and desired outcomes. Consequently, according to this perspective, poor implementation is not necessarily the reason for poor policy outcomes (see Capano and Woo 2017). More recently, the perspective of research on policy design has moved from single tools to multiple tools and tool mixtures used to address policy problems. The main argument of this analytical perspective is that “efforts of policy makers often have failed due to poor designs which have failed adequately to incorporate this complexity into policy formulation” (Howlett et al., 2015, p. 300). The broadened perspective of tool mixtures opens up an entirely new perspective since the individual elements of the tool boxes selected may conflict with each other. Thus the process of formulating and adopting policy tool mixtures needs to focus more on the actors involved, their preferred policy designs, and their role in decision-making as defined by the respective rules and institutions. Policy tool mixtures are not only more difficult to design than single tools, but also more difficult to implement. 4 Policy implementation The literature on policy implementation deals with how policies are put into practice by administrative actors and to what extent they affect the behavior of societal target groups (for an overview, see Pülzl and Treib 2007). Following Winter (2012) and Vancoppenolle et al. (2015), policy implementation consists of four dimensions: defining the implementation structure, agency decision-making, target group behavior, and policy results. Implementation structure refers to the number and types of organizations involved in the implementation process. It is often the case that two or more ministries or agencies are responsible for implementing a policy. This is especially true with cross-sectoral policies, where the various organizations in charge of policy implementation need to coordinate their activities in what is known as horizontal coordination (e.g. Peters 2015). For example, in order to implement the Paris Agreement on Climate Change, the German government adopted the Climate Action Plan 2050, which stresses the need for horizontal coordination among several ministries at the federal level. In multilevel polities, policy implementation also often includes vertical coordination among organizations located at different levels of government (e.g. Bolleyer and Börzel 2010). Policy implementation may also require the involvement of private actors (e.g. Thomann et al. 2016; Tosun 2017; Tosun et al. 2017). In addition to the types of organizations or actors involved, the implementation structure includes policy-related rules and procedures and the allocation of resources. Agency decision-making refers to the process of making the legal stipulations more concrete and therefore implementable. This requires the members of one or several competent organizations to decide on these concretizations and to elaborate a procedure for the delivery of the services, which also includes considerations about the leeway for managerial discretion and the scope and nature of the involvement of street-level personnel (see, e.g., Sager et al. 2014). Sometimes street-level personnel benefits from great discretion in making decisions relevant for policy delivery such as in social work (e.g. Ellis 2007), whereas in other cases, their leeway for discretion is limited, as is often the case in policing (e.g. Rowe 2007). Target group behavior refers to the role the policy addressees play in the implementation process—their actions as well as their needs (see Pierce et al. 2014). To obtain social benefits, for example, the target group’s involvement in the implementation process consists of filing a request and complying with the conditions attached to the benefits (e.g., regular appointments with case managers). Implementation of other policies may entail more substantive changes to the behavior of the target groups (see Howlett 2005: 47). In the case of promoting biofuels, for 5 instance, policy implementation crucially depends on the behavior of the fuel producers and suppliers as well as on consumers’ buying choices (see Tosun 2017). Policy results are the outcome of the implementation process, which can correspond to the intentions of the policy concerned or deviate from them. For example, if a government seeks to reduce air pollution, it can adopt stricter limit values for the emission of air pollutants. The policy result would correspond to the intentions if levels of air pollution were diminished. The policy result would deviate from the intention of the policy if there were no changes or an increase in air pollution levels. If the policy results deviate from the stipulated intentions, the reason for this can be problems encountered in the implementation process or flaws in the design of the policy. Of these four elements, policy studies have devoted most attention to implementation structure and agency decision-making. With regard to the implementation structure, Howlett’s (2004, 2005) work on ‘implementation styles’ has been particularly influential. Implementation styles denote typical combinations of substantive and procedural policy tools that are employed to achieve given policy goals. Howlett differentiates between the severity of state constraints (with regard to resources and legitimacy) and the nature of the policy target (which relates to exchange or policy actors) in order to identify four ideal-typical implementation styles. Institutionalized voluntarism is the implementation style of choice when the policy targets constitute a large group and state constraints are high. It corresponds to an exhortation-based manipulation of market actors and the institutionalization of networks. For example, in 1985, the Chemistry Industry Association of Canada launched the Responsible Care Initiative, which is a complementary scheme to regulation to facilitate the self-control of the industry and to increase public and political trust in its activities. Regulatory corporatism is chosen when state constraints are high, but the policy targets constitute a small group. This implementation styles regulates market actors and manipulates their interest-articulation system by means of financial incentives. It corresponds to a ‘corporatist’ style economic planning models in industrial policymaking (Howlett 2004: 10). Levi-Faur (2010) refers to regulatory corporatism as a system where strong civil and state regulation co-exist. Of the four forms identified by Howlett (2004, 2005), this one has the most coercive character. 6 The third implementation style, directed subsidization, is expected to be chosen when governmental actors face low constraints and interact with large groups of policy targets. Directed subsidization refers to the extensive use of financial instruments to steer the behavior of market actors, coupled with the use of authority to preferentially recognize networks of actors (Howlett 2004: 10). For example, the German government has decided several years ago to promote the transition to a bio-based economy, which it supports by means of subsidies for the industries concerned as well as investing heavily into research activities that could potentially support this process. Lastly, the fourth implementation style, public provision with oversight, corresponds to a mobilization model, where governmental organizations use resources to provide goods and services to small groups of policy targets. At the same time, the government seeks to manipulate actor networks through information. This implementation style is chosen when the number of policy addressees is small and so are the constraints on the state (see Howlett 2004: 10). The state-based provision of internet in Estonia can be regarded as a case in point for this implementation style. Turning to agency decision-making, we enter the realm of a long established strand of policy studies. This strand of research has long been occupied with controversies between two opposing schools of thought. Top-down scholars stress the role of centrally defined policy goals that were to be executed by a hierarchically ordered administrative structure with as little deviation from the original stipulations as possible (Pressman and Wildavsky 1973; Van Meter and Van Horn 1975; Sabatier and Mazmanian 1979). Proponents of the bottom-up camp, in contrast, argue that policies are often implemented in a much more decentralized way, leaving individual street-level bureaucrats significant leeway in adapting the policy to local circumstances (Elmore 1979; Lipsky 1980; Hjern and Porter 1981). To overcome the often unfruitful clash between the opposing normative assumptions about ‘good’ policy delivery underlying these two schools of thought, a third group of scholars decided to develop integrated analytical models. The hybrid models bring together elements from both approaches and treat the question of whether actual implementation processes resembled more the bottom-up or the top-down approach as an empirical question (Goggin et al. 1990; Winter 1990, 2012; Matland 1995). Policy feedbacks The final concept we need to define before setting out our conceptual model refers to policy feedbacks. Most fundamentally, policy feedbacks are conceived as reactions by the policy 7 addressees or other parts of the society to policy decisions adopted in the past (e.g. Pierson 1993; Béland 2010). Mettler and SoRelle (2017) point to four dominant streams in the study of policy feedbacks. The first stream analyses how policies affect political agendas and the definition of policy problems. The second stream extends the logic of policy feedbacks to how they affect governance through their impact on the capacity of government and political learning by public officials. The third stream concentrates on how policy feedback influences the power of societal groups. The fourth stream concentrates on the study of individual political behavior by examining how policies shape the meaning of citizenship. To this framework, we add a fifth stream of research, which asks how necessary but unpopular policies can be designed to become ‘sticky’ and resilient towards attempts to dismantle them. For example, Jordan and Matt (2014) applied this perspective for the analysis of climate change policies. Of the five streams presented, the second stream on governance is most pertinent for our analysis. The corresponding literature argues that policies may affect future governance as they shape the feasible set of policy choices and the type of organizational arrangements assigned to new policies, and the broader scope of government action as well as political learning by public officials. According to Mettler and SoRelle (2017), such outcomes are likely to emerge if new policies enable governments to develop capacities they did not have previously, which they can then use for the delivery of policies developed subsequently. Following Pierson (1993), whether a policy is subject to redesign depends on whether the feedback is positive or negative. Positive feedback is expected to preserve the status quo, whereas negative feedback is likely to trigger attempts to redesign policies. Policy feedbacks can, in principle, stimulate or reinforce policy learning (Van der Knaap 1995), which, however, depends on the respective policymakers’ willingness and capacity to learn. From this, it becomes clear why policy feedback has been frequently discussed in the context of policy change (e.g. Béland 2010), while this concept has received limited attention from scholars working on policy implementation. 3. An integrated model of policy design, implementation style and policy feedback Our objective in this section is to develop a model that bridges the different strands of literature and integrates the concepts of policy design, implementation and policy feedback. We believe that such an integrated model is useful for, on the one hand, aligning the established literature 8 on policy implementation with the literature on policy design, and on the other hand, to connect policy studies with comparative politics by drawing on the concept of policy feedback. The starting point of the conceptual model is the stage of policy design, which results in two sets of decisions: the selection of the policy tools and the selection of the implementation structures. The selection of policy tools can be examined from the perspective of policy studies and comparative politics alike; both perspective offer important insights and highlight complementary aspect. Starting with policy studies, a recent conceptual note by Chindarkar et al. (2017) stresses the importance of capacity for the quality of policy design outcomes. The authors refer to two types of capacity: governance capacity and analytical capacity. When governance and analytical capacity are high, the authors expect capable design, while in the opposite constellation—where both capacities are low—poor design is the most likely outcome. When governance capacity is high but analytical capacity is low, the authors anticipate capable political design, where the policy is likely to be politically feasible but the quality of policy is low. The complementary scenario (low governance capacity, high analytical capacity) is likely to lead to poor political design, where the quality of the policy is likely to be good, but it may not be politically feasible. The literature on policy design also ties into research on path dependency and incrementalism. Drawing on historical institutionalism, path dependency models presume that the best predictor of future policy design is past policy design (see Pierce et al. 2014). This expectation is clearly at odds with the model put forward by Chindarkar et al. (2017), which rests on the assumption that policy design is the outcome of policy analysis. Nevertheless, the two perspectives can be reconciled: in situations where the analytical capacity of governments is low, the most rational decision to take could be to continue doing what they have done in the past. Comparative politics regards policy design as the outcome of institutional arrangements and the preferences of individual actors. The literature on veto players (Tsebelis 2002) has demonstrated that it makes a difference for the outcome of the policy design process which actors are involved in decision-making and what their policy preferences look like. Policy preferences stem, for example, from the ideologies of political parties that form the government (Tosun and Workman 2017). Governments that are composed of left-wing parties generally tend to display a greater propensity to intervene in the economy, whereas right-wing parties are more in favor of voluntary measures (e.g. Bakker et al. 2015). In parliamentary systems in Europe, political parties often need to form coalition governments that consist of two or more 9 political parties. In such constellations, not only the ideology of one party is likely to determine the selection of policy tools, but the ideologies of several parties and the way in which they reach agreements. Depending on the decision-making dynamics within coalition governments, policy design might be poor due to the need to reach a compromise solution. For the same reason, policy design can turn out to be of a good quality since the political parties in government need to convince one another and engage in an analysis of the existing policy options, which then leads to the selection of the alternative that appears most promising. In the previous section, we followed the conceptual work of Winter (2012) and Vancoppenolle et al. (2015) and differentiated between the implementation structure and agency decisionmaking. When looking at policy implementation as a complementary dimension of policy design, the differentiation between these two elements becomes clear. We conceive the selection of the implementation structure to include decisions on how much discretionary leeway agencies and bureaucrats inside the agencies should be granted. This does not mean that the need for subsequent decisions becomes obsolete, but the general decision about whether the implementation style is centralized or decentralized is likely to be made at the stage of policy design. Above we discussed the four ideal-typical identified by Howlett (2004, 2005) which are useful for explaining variations across different policy domains in one countries as well as across countries. However, for introducing an integrated conceptual model, we must abstract from the analytical considerations on which Howlett’s typology rests. Similarly, we cannot draw on the various models of policy implementation that approach this phenomenon from the top-down, bottom-up or a mixed perspective (see Pützl and Treib 2007) since these rest on specific assumptions and theoretical considerations as well. Yet we draw loosely on the perspective of top-down and bottom-up models of policy implementation as the selection of the implementation structure will most fundamentally set out whether the implementation process is centralized or decentralized. Our simplified definition of the implementation structure consists of two dimensions: first, the number of agencies involved in implementation; second, the degree to which the agencies involved have discretionary powers to adjust the policy design. Centralized implementation is characterized by a small number of implementing agencies that have little to no discretionary power. Resonating with the logic of top-down implementation, a centralized implementation structure seeks to ensure that there is no deviation from the original policy design. A decentralized implementation structure comprises multiple agencies and allows them to deviate 10 to a certain degree from the stipulations of the policy in order to take into account the local context, assuming that the policy results will benefit from these adjustments to the policy concerned. There are several explanations for why the policy design process may result in the selection of a centralized or a decentralized implementation structure. One of the explanations is the political system’s institutional features. For example, we can observe that federal polities tend to opt for decentralized implementation structures, whereas centralized polities are more likely to select centralized implementation structures (see, e.g., Wälti 2004). Another potential explanation is the nature of the policy problem underlying the policy design. For example, it would be virtually impossible to define a centralized implementation structure for a policy that aims to replace conventional fuels by fuel blend that contains a certain share of biofuels. The production and distribution of fuels is the task of private mineral oil companies and therefore policy implementation in this particular case requires these companies to be included in the implementation structure (see Tosun 2017). In such arrangements where public and private actors cooperate, the latter typically also have some degree of discretionary power in implementing the policy (see Thomann et al. 2016). Centralized implementation results in information about the effectiveness of a policy when it is implemented in an identical or at least similar manner. In more abstract terms, the implementation process – if done correctly – produces repeated observations of whether a policy in questions produces the intended policy results. Seen from this perspective, centralized implementation allows, in principle, for gathering information about the quality of the policy tools selected. This implementation structure also allows for inferring insights on how the policy targets and other societal groups perceive of the policy goal and the policy tool concerned. Put differently, policy-makers can gather information on whether the policy goal they seek to attain and the policy tools they employ are accepted and considered legitimate. Decentralized implementation potentially results in variation of the policy design originally adopted by policymakers, which increases the information basis for the evaluation of the policy tool concerned. The success of policy implementation depends on whether the process facilitated learning, capacity-building and support-building in order to address problems associated with it in a decentralized way, consistent with the interests of the actors involved (see Schneider and Ingram 1997; Pierce et al. 2014). Relating this to the literature on polycentric governance (Ostrom 2010), each adapted policy design can be regarded as a ‘policy experiment’ that could improve policymaking through processes of upscaling. Likewise, decentralized 11 implementation can produce insights into which variations of the policy design help to improve the societal acceptance and legitimacy of the policy tool concerned. To sum up, the degree to which the implementation structure is centralized is likely to affect which information governments may infer from the implementation process for future policy processes. For this information to have an impact on future policymaking, it needs to be transformed into policy feedback. To this end, policymakers need to collect this information, analyze it, and decide, in light of the insights gained, whether a new policy process is needed. In line with the literature, we expect positive feedback to support the policy status quo and negative feedback to start a new policy process that aims to improve the design of a policy adopted previously (see Mettler and SoRelle 2017). As we have seen above, decentralized implementation has a greater potential to offer insights into how policies must be designed to be effective and accepted. This richness in information, however, represents a potential impediment to improving future policy design by means of policy feedback. This expectation results from three observations. First, an analysis of variations of one policy requires a high level of analytical capacity, which not all governments possess and are willing to invest (see Wellstead et al. 2011). Second, the multitude of agencies involved in the implementation process may make it difficult to communicate with the actors on the governance level that is responsible for policy design. For example, it could be possible that information on variations of a central policy as it is implemented at the local level only reaches policymakers at the regional level, who are not responsible for the policy design and lack the competence to start a redesigning process. Third, information on the effectiveness of the adapted policy and how it is perceived by policy addressees and the broader society might be inconsistent and potentially force policymakers to decide whether they want to adopt redesigned policy that is likely to be less effective, but finds more acceptance among the target population, or the other way around. To be sure, this is a rather hypothetical case since it requires an even more advanced capacity and willingness to conduct policy analysis. Furthermore, it should be noted that while this model can be applied to individual policy tools and policy mixes alike (see Howlett et al. 2015), the latter entails an additional complication. Altogether, despite the greater level of useful information that may result from decentralized policy implementation, we postulate that this information is less likely to function as a policy feedback than the more sparse information resulting from centralized policy implementation. To be even clearer, we argue that the excess of information can possibly hamper the transformation of information into policy feedback and the initiation of a new policy process. 12 Information originating from centralized implementation is more likely to produce policy feedback, but the information available is more limited. While this can trigger a policy process to redesign the policy concerned, the changes in the policy design are likely to concentrate on one or a few aspects. This can result in a sequence of redesign efforts, of which each concentrates on different aspects. Figure 1 provides an overview of the structure of our conceptual model, including the stages of policy design, the selection of policy tools, the selection of the implementation structure, the outcomes of this selection process (centralized vs. decentralized), and the information resulting from the implementation structure selected. Figure 1: A Model of Policy Design, Implementation Style, and Policy Feedback (1) Low level of analytical capacity required (2) Information channels straightforward (3) Information less likely to be inconsistent Policy feedback more likely Policy design Policy tools Centralized Limited but clear informationon effectivenessand acceptance Decentralized Extensive but diverse information on effectiveness and acceptance Implementation structure Policy feedback less likely (1) High level of analytical capacity required (2) Information channels complex (3) Information more likely to be inconsistent Source: Own elaboration. 4. Toward the formulation and testing of hypotheses The conceptual model illustrated in Figure 1 deliberately omits a number of details in order to highlight the most important elements of the relationship between policy design, policy implementation, and policy feedback. It therefore does not include many of the explanatory factors discussed in the specific literatures on these concepts either. For example, the literature on policy design stresses the importance of analytical capacity (see Chindarkar et al. 2017), which our model only takes into account with regard to the likeliness of policy feedback. 13 Howlett’s (2004, 2005) treatise of implementation styles elaborates on the severity of state constraints and the nature of the policy targets, which would also have to be incorporated into an extended model. And the literature on policy feedback argues that welfare policies are particularly likely to produce policy feedback, which means that the type of policy represents another explanatory variable potentially to be included (see Soss and Schram 2007). The fact that our model does not incorporate theseand otherfactors does not suggest that they are irrelevant. In fact, quite the contrary is the case. We expect these explanatory factors to matter, but how exactly they matter needs to be determined in a next step of theorizing. In this next step, the macro-level model will have to be transformed into one or multiple meso- and microlevel models, which then allow for formulating and testing hypotheses. Before formulating empirically testable hypotheses that capture the role of a set of relevant explanatory variables, we suggest assessing the consistency of the macro-level model first. Only if the basic logic of the model holds is it worth engaging in theoretical refinements and sophisticated testing strategies. For this purpose, we propose adopting a comparative research design that allows for addressing the following three features: the implementation structure, the characteristics of the political system, and the analytical capacity. First of all, we need to determine how often and when policymakers choose a centralized or a decentralized implementation structure. Put differently: do we have variation in the implementation structures selected within and across countries? Drawing on the notion of implementation styles as put forward by Howlett (2004, 2005), one could expect that implementation structures tend to vary across countries, but less so across policy domains within countries. Following the conceptual reasoning by Winter (2012) and Vancoppenolle et al. (2015), the implementation structure is decided individually for each policy, which means that we could expect variation across countries, within countries and even within policy domains. Consequently, before proceeding with the conceptual discussion, we need to develop a clearer understanding of the variation in implementation structures and ideally we would do so for different types of variations: across countries, within countries, but across policy domains, and within countries and policy domains. Depending on the outcome of this exercise, we could extend our theoretical reasoning and add potential explanatory variables for the variation in implementation structure. A second but related research step is to pay attention to the features of political systems. Most importantly, we need to differentiate between unitary and federal polities since this may also determine whether implementation structures are always or mostly centralized or decentralized. 14 The institutional characteristics are also important for learning about possible constraints policymakers may face at the stage of policy design. As already explained above, it is plausible to expect that multi-party governments adopt a different approach to policy design than singleparty governments. Furthermore, it is not only important to differentiate between multi- and single-party governments, but also to take into account the ideological preferences of the governing parties with regard to the choice of policy tools. From this perspective, it appears important to observe whether there is variation in the process and outcome of policy design, depending on the specific institutional characteristics of polities such as the degree of federalism and the partisan composition of governments. The third line of empirical research should look into the varying analytical capacities of governments. Analytical capacity is a straightforward concept, but it is difficult to operationalize (but see Wellstead et al. 2011). Of course, there is information on the staff and budget of agencies, but these indicators do not tap into the concrete tools these agencies use to analyze policy outcomes and what they consider as the benchmark against which the outcomes are evaluated. The operationalization becomes even more challenging when we seek to carry out comparative research, which almost automatically leads to the need to reduce information in order to facilitate comparability. Therefore, with regard to this third dimension, we need to develop valid measurements, which we could then use to produce comparative insights. After mastering this task, a next step is to examine when information about the effectiveness and acceptance of policies actually produces policy feedbacks and whether the reasoning put forward above holds true. To this end, we need to explore variation in the production of policy feedback loops. Once we have a better understanding of whether there is any variation, we should test whether this variation is due to the complexity of information. It is conceivable that policy feedbacks do not or not only stem from the availability and characteristics of policyrelated information produced during the implementation process, but are triggered by other factors. Despite the richness of the literature on policy feedback (see Mettler and SoRelle 2017), our understanding of how (positive and negative) policy feedbacks come about is limited. Therefore, concentrating on cases where we can observe policy feedbacks and contrasting them with negative cases can produce insights that are not only interesting for the conceptual model proposed here, but also for the comparative politics literature on policy feedbacks in general. As the prior explanations have indicated, there are only few established operationalizations for measuring the concepts of importance here. Therefore, it makes sense to start with comparative case studies that address only one of the three dimensions. The question of when policy 15 feedbacks emerge requires even more advanced analytical skills and should be addressed after having examined the other factors. At later stages, the empirical investigation could move on to quantitative analyses. Regardless of whether the research logic is qualitative or quantitative, however, it is appears essential to adopt comparative designs since the questions raised above revolve around variation. Once we have an improved understanding of the variability in the concepts, we can apply more sophisticated methods to establish the causal relationships between the different concepts. Consequently, the exploration and systematic analysis of empirical variation appears to be the next steps to take before being able to further our theoretical and conceptual understanding. 5. Conclusion Public policies can only become effective once they are put into practice. A policy design may be perfect, but it can still fall short of delivering the intended results when the implementation is flawed. This was the key message of the path-breaking book by Pressman and Wildavsky published in 1973. Eager to further scholarship on policy implementation, we decided to bring together three concepts that have as yet existed side by side without being discussed jointly: policy design, implementation style, and policy feedback. We proposed to connect policy design and policy implementation by means of the concept of policy feedback. The latter has mostly been discussed by studies in the field of comparative politics and there are only few policy studies that embrace this concept and adapt it to the phenomena typically studied by scholarship on public policy (see, e.g., Béland 2010). This is a surprising finding since policy feedback theory is widely considered to belong to the group of policy process theories (see Mettler and SoRelle 2017). With this contribution, we offer a perspective for furthering our understanding of policy implementation as well as to bring the concept of policy feedback back into policy studies. In this regard, we are confident that our reasoning about when information derived from policy implementation becomes a policy feedback also contributes to the comparative politics literature on policy feedback. The literature on policy feedback typically assumes that a feedback is given. We challenge this automatism and argue that the availability of information does not necessarily correspond to a policy feedback. This contribution would not have been possible without the public policy literature on the analytical capacity of policymakers (e.g. Wellstead et al. 2011), which demonstrates the value to bringing these concepts together. 16 Policy delivery can be organized in a hierarchical manner, where fixed policy intentions are expected to be handed down to addressees with as little deviation as possible (centralized implementation structure), or it can be organized in a more decentralized manner which leaves implementing actors more leeway for adapting the policy to local circumstances. This chapter has argued that while decentralized implementation offers more insights into what makes policies work than centralized implementation, the latter is more likely to produce policy feedback that can inform future policymaking and potential redesign of the policy concerned. This expectation builds on the complexity of information that can potentially result from a decentralized implementation process. With decentralized implementation, we are likely to observe a policy to be adapted to the conditions at the local level. Seen from this perspective, the individual adapted policies represent policy implementation experiments, of which some will work and other will fail. This variation in the effectiveness of the (adapted) policy constitutes a valuable source for improving the quality of policy design in the future. However, transforming this information into a policy feedback requires a high level of analytical capacity. Even if the necessary analytical capacity is available, inconsistent or conflicting information supplied from the various agencies involved could still hamper the emergence of policy feedback. Consequently, future policy design is more likely to benefit from insights originating from a centralized implementation structure for this arrangements has a greater potential to produce a policy feedback and therefore to guide policy-making. We believe that this is a novel way of looking at policy implementation and its outcomes for the policy process. Most studies of policy implementation attribute decentralized structures a higher capacity to benefit from policy learning than centralized structures. In decentralized arrangements, implementers have flexibility and autonomy to adjust policies in light of particular local conditions and changes in the perception of policy problems (Nilsen et al. 2013). This results in more information about how a given policy can be made operational. The argument we put forward is that despite the lower level of information, centralized policy implementation is more likely to stimulate policy feedback and learning than decentralized implementation since the information is easier to assess and is more likely to reach the policymakers responsible for policy redesign. At this point, we could only present our conceptual model, but owe the readership empirical findings to substantiate our findings. The reason for this is that an empirical test of our claims warrants a complex research design, which requires making decisions on which aspects we would like to keep constant and which ones we want to vary across the cases analyzed. As we elaborated in the previous section, we are convinced that an empirical test of our model can 17 only be attained by adopting a comparative research design. Given the early phase in which our reasoning is, both comparative case study analysis and quantitative analysis can help to refine our argument. The comparison can be attained by looking at different policy domains within one countries or at policies adopted in one policy domain across countries. What is important is to select a research design that allows for varying the implementation structure, the analytical capacity of government, and features of the political system. Altogether, we can conclude that the joint study of policy design, policy implementation and policy feedback is still in its infancy and requires innovative conceptual thinking and theorizing. In addition, comparative empirical investigations are needed to create a continuous process of formulating and testing hypotheses. Perhaps the absence of systematic comparative research is currently the biggest challenge for the advancement of this field. Therefore, we welcome all endeavors to produce and analyze comparative data on policy design, policy implementation, and policy feedback. In so doing, public policy scholars are well advised to take into account the relevant comparative politics research on policy feedback. 6. References Bakker, R., De Vries, C., Edwards, E., Hooghe, L., Jolly, S., Marks, G., ... and Vachudova, M. A. (2015): Measuring party positions in Europe: The Chapel Hill expert survey trend file, 1999–2010. Party Politics, 21(1), 143-152. Béland, D. (2010): Reconsidering policy feedback: How policies affect politics. Administration & Society, 42(5), 568-590. Bolleyer, N., and Börzel, T. A. (2010): Non-hierarchical policy coordination in multilevel systems. European Political Science Review, 2(2), 157-185. Capano, G., Regini, M., and Turri, M. (2016): Changing Governance in Universities. Palgrave Macmillan UK. Capano, G., and Woo, J. J. (2017): Resilience and robustness in policy design: A critical appraisal. Policy Sciences, 1-28. Chindarkar, N., Howlett, M., and Ramesh, M. (2017): Introduction to the Special Issue: “Conceptualizing Effective Social Policy Design: Design Spaces and Capacity Challenges”. Public Administration and Development, 37(1), 3-14. 18 Colebatch, H. K. (2017): The idea of policy design: Intention, process, outcome, meaning and validity. Public Policy and Administration, 0952076717709525. Damonte, A., Dunlop, C. A., and Radaelli, C. M. (2014): Controlling bureaucracies with fire alarms: policy instruments and cross-country patterns. Journal of European Public Policy, 21(9), 1330-1349. DeLeon, P. (1988): The contextual burdens of policy design. Policy Studies Journal, 17(2), 297-309. Ellis, K. (2007): Direct payments and social work practice: The significance of ‘street-level bureaucracy’in determining eligibility. British Journal of Social Work, 37(3), 405-422. Elmore, R. F. (1979): Backward Mapping: Implementation Research and Policy Decisions. Political Science Quarterly, 94(4), 601-616. Falkner, G., Hartlapp, M., and Treib, O. (2007): Worlds of compliance: why leading approaches to european union implementation are only ‘sometimes‐true theories’. European Journal of Political Research, 46(3), 395-416. Falkner, G., and Treib, O. (2008): Three worlds of compliance or four? The EU‐15 compared to new member states. Journal of Common Market Studies, 46(2), 293-313. Goggin, M. L., Bowman, A. O.'M., Lester, J. P., and O'Toole, L. J. (1990): Implementation Theory and Practice: Toward a New Generation. Glenview: Scott. Grüne-Yanoff, T. and Hertwig, R. (2016): Nudge Versus Boost: How Coherent are Policy and Theory? Minds and Machines 26(1), 149–183. Hjern, B., and Porter, D. O. (1981): Implementation Structures: A New Unit of Administrative Analysis. Organization Studies, 2(3), 211–227. Hood, C. (1986): The tools of government, Chatham: Chatham House Publishers Hood, C., and Margetts, HZ. (2007): The tools of government in the digital age, Basingstoke: Palgrave Macmillan. Howlett, M. (2004): Beyond good and evil in policy implementation: Instrument mixes, implementation styles, and second generation theories of policy instrument choice. Policy and Society, 23(2), 1-17. 19 Howlett, M. (2005): What is a policy instrument? Tools, mixes, and implementation styles. Designing government: From instruments to governance, 31-50. Howlett, M. (2009): Governance modes, policy regimes and operational plans: A multi-level nested model of policy instrument choice and policy design. Policy Sciences, 42(1), 73-89. Howlett, M., and Ramesh, M. (1993): Patterns of Policy Instrument Choice: Policy Styles, Policy Learning and the Privatization Experience. Policy Studies Review, 12(1/2), 3–24. Howlett, M., and Mukherjee, I. (2017): Policy design: From tools to patches. Canadian Public Administration, 60(1), 140-145. Howlett, M., and Rayner, J. (2007): Design principles for policy mixes: cohesion and coherence in ‘new governance arrangements’. Policy and Society, 26(4), 1-18. Howlett, M. (2011): Revisiting policy design: The rise and fall (and rebirth?) of policy design studies. In General Conference of the European Consortium for Political Research (ECPR), Section (Vol. 63). Howlett, M., Mukherjee, I., and Woo, J. J. (2015): From tools to toolkits in policy design studies: the new design orientation towards policy formulation research. Policy & Politics, 43(2), 291-311. Hupe, P. L., and Hill, M. J. (2016): ‘And the rest is implementation.’ Comparing approaches to what happens in policy processes beyond Great Expectations. Public Policy and Administration, 31(2), 103-121. Hupe, P., and Sætren, H. (2015): Comparative Implementation Research: Directions and Dualities. Journal of Comparative Policy Analysis: Research and Practice, 17(2), 93-102. John, P., Smith, G., and Stoker, G. (2009): Nudge nudge, think think: two strategies for changing civic behaviour. The Political Quarterly, 80(3), 361-370. Jordan, A., and Matt, E. (2014): Designing policies that intentionally stick: policy feedback in a changing climate. Policy Sciences, 47(3), 227-247. Kahneman, D., Slovic, P., and Tversky, A. (Eds.). (1982): Judgment under uncertainty: Heuristics and biases. New York, NY: Cambridge University Press. Levi-Faur, D. (2010): Regulatory Architectures for a Global Democracy: On Democratic Varieties of Regulatory Capitalism. in: T. Porter and K. Ronit, (Eds) The Challenges of 20 Global Business Authority: Democratic Renewal, Stalemate, or Decay? Suny University Press, 2010, pp. 205-226. Lipsky, M. (1980): Street level bureaucrats. Nova York: Russel Sage. Matland, R. E. (1995): Synthesizing the implementation literature: The ambiguity-conflict model of policy implementation. Journal of public administration research and theory, 5(2), 145-174. Mettler, S., and SoRelle, M. (2017): Policy feedback theory. in: C. Weible and P. Sabatier (Eds.), Theories of the policy process, Boulder, CO: Westview Press, 151-181. Mortensen, P. B., and Green-Pedersen, C. (2015): Institutional Effects of Changes in Political Attention: Explaining Organizational Changes in the Top Bureaucracy. Journal of Public Administration Research & Theory, 25(1), 165–189 Nilsen, P., Ståhl, C., Roback, K., and Cairney, P. (2013): Never the twain shall meet?-a comparison of implementation science and policy implementation research. Implementation Science, 8(63). Ostrom, E. (2010): Beyond markets and states: polycentric governance of complex economic systems. Transnational Corporations Review, 2(2), 1-12. Peters, B. G. (2015): Pursuing horizontal management: The politics of public sector coordination. University Press of Kansas. Pierce, J. J., Siddiki, S., Jones, M. D., Schumacher, K., Pattison, A., and Peterson, H. (2014): Social construction and policy design: A review of past applications. Policy Studies Journal, 42(1), 1-29. Pierson, P. (1993): When effect becomes cause: Policy feedback and political change. World politics, 45(4), 595-628. Pressman, J. L., and Wildavsky, A. B. (1973): Implementation: how great expectations in Washington are dashed in Oakland: or, why it's amazing that federal programs work at all, this being a saga of the Economic Development Administration as told by two sympathetic observers who seek to build morals on a foundation of ruined hopes. Univ of California Press. 21 Pülzl, H., and Treib, O. (2007): Implementing public policy. Handbook of public policy analysis: Theory, politics, and methods, 125, 89-107. Rowe, M. (2007): Rendering visible the invisible: police discretion, professionalism and decision-making. Policing & Society, 17(3), 279-294. Sabatier, P. A., and Mazmanian, D. A. (1979): The Conditions of Effective Implementation: A Guide to Accomplishing Policy Objectives. Policy Analysis, 5(4), 481–504. Sager F., and Rielle Y. (2013): Sorting through the garbage can: under what conditions do governments adopt policy programs? Policy Sciences 46(1): 1–21. Sager, F., Thomann, E., Zollinger, C., van der Heiden, N., and Mavrot, C. (2014): Street-level Bureaucrats and New Modes of Governance: How conflicting roles affect the implementation of the Swiss Ordinance on Veterinary Medicinal Products. Public Management Review, 16(4), 481-502. Saetren, H. (2014): Implementing the third generation research paradigm in policy implementation research: An empirical assessment. Public Policy and Administration, 29(2), 84-105. Schneider, A. L., and Ingram, H. M. (1997): Policy design for democracy. University Press of Kansas. Soss, J., and Schram, S. F. (2007): A public transformed? Welfare reform as policy feedback. American Political Science Review, 101(1), 111-127. Strassheim, Holger and Silke Beck (Eds.) (2019): The Research Handbook of Behavioural Change and Public Policy. Cheltenham: Edward Elgar. Thaler, R. H., and Sunstein, C. R. (2008): Nudge: Improving Decisions About Health, Wealth, and Happiness. New Haven, CT: Yale University Press. Thomann, E., Lieberherr, E., and Ingold, K. (2016): Torn between state and market: Private policy implementation and conflicting institutional logics. Policy and society, 35(1), 5769. Tosun, J. (2013): Environmental policy change in emerging market democracies: Eastern Europe and Latin America compared (Vol. 41). University of Toronto Press. 22 Tosun, J. (2017): The behaviour of suppliers and consumers in mandated markets: the introduction of the ethanol–petrol blend E10 in Germany. Journal of Environmental Policy & Planning, 1-15. Tosun, Jale und Samuel Workman (2017): Struggle and Triumph in Fusing Policy Process and Comparative Research. In C. Weible und P. Sabatier (Hrsg.), Theories of the Policy Process. Boulder, CO: Westview Press, 329–362 Tosun, J., Unt, M., and Wadensjö, E. (2017): Youth‐oriented Active Labour Market Policies: Explaining Policy Effort in the Nordic and the Baltic States. Social Policy & Administration, 51(4), 598-616. Tsebelis, G. (2002): Veto players: How political institutions work. Princeton University Press. Vancoppenolle, D., Sætren, H., and Hupe, P. (2015): The Politics of Policy Design and Implementation: A Comparative Study of Two Belgian Service Voucher Programs. Journal of Comparative Policy Analysis: Research and Practice, 17(2), 157-173. Van Meter, D. S., and Van Horn, C. E. (1975): The policy implementation process: A conceptual framework. Administration & Society, 6(4), 445-488. Van der Knaap, P. (1995): Policy evaluation and learning: feedback, enlightenment or argumentation?. Evaluation, 1(2), 189-216. Wälti, S. (2004): How multilevel structures affect environmental policy in industrialized countries. European journal of political research, 43(4), 599-634. Weaver, R. K. (2015): Getting People to Behave: Research Lessons for Policy Makers. Public Administration Review, 75(6), 806-816. Wellstead, A. M., Stedman, R. C., and Howlett, M. (2011): Policy analytical capacity in changing governance contexts: A structural equation model (SEM) study of contemporary Canadian policy work. Public Policy and Administration, 26(3), 353-373. Winter, S., (1990): Integrating Implementation Research, in: D. J. Palumbo and D. J. Calista (Eds), Implementation and the Policy Process: Opening up the Black Box (New York: Greenwood Press), pp. 19-38. Winter, S., (2012): Implementation. in: B. G. Peters and J. Pierre (Eds) Handbook of Public Administration, 2nd ed. (London: Sage), pp. 255–263. 23