Academia.eduAcademia.edu

Ethics of Coding: A Report on the Algorithmic Condition

How can we think adequately about the relation between knowledge and ethics in societies that are governed by algorithmic digital systems and objects endowed with agency?

ETHICS OF CODING A REPORT ON THE ALGORITHMIC CONDITION Suggested Citation: Colman, F., Bühlmann, V., O’Donnell, A. and van der Tuin, I. (2018). Ethics of Coding: A Report on the Algorithmic Condition [EoC]. H2020-EU.2.1.1. – INDUSTRIAL LEADERSHIP – Leadership in enabling and industrial technologies – Information and Communication Technologies. Brussels: European Commission. 732407,https://cordis.europa.eu/project/rcn/207025_en.html. pp.1–54. In-text: (Colman et al., 2018) Cover Illustration: © Sam Skinner ETHICS OF CODING A REPORT ON THE ALGORITHMIC CONDITION Authors Professor Felicity Colman, Kingston University, London, United Kingdom (KU) Professor Vera Bühlmann, Technische Universität Wien, Austria (TU) Professor Aislinn O’Donnell, National University of Ireland Maynooth, Ireland (NUIM) Professor Iris van der Tuin, Utrecht University, Netherlands (UU) Project Number: 732407 Project Acronym: EoC Project Title: Ethics of Coding Report Period: 01/01/2017 – 01/12/2017 http://cordis.europa.eu/project/rcn/207025_en.html H2020-EU.2.1.1. – INDUSTRIAL LEADERSHIP – Leadership in enabling and industrial technologies – Information and Communication Technologies ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION TABLE OF CONTENTS CONTRIBUTORS ONE: INTRODUCTION: THE ETHICS OF CODING 1.1 EoC Aims and Scope 1.2 Knowledge: Coding Ethics 1.3 Ethics TWO: ALGORITHMS 2.1 The Algorithmic Condition 2.2 Algorithmic Modalities, and the Behaviour of Mathematical Models THREE: DATA 3.1 Data 3.2 Data Sovereignty FOUR: CODE 4.1 Coding definitions 4.2 Social Coding 4.3 Ethical Coding 4.4 Educational Coding FIVE: ETHICS AND MANNERS OF CONDUCT 5.1 Ethics 5.2 Doing Ethics 5.3 EoC Ethics Based on Codes of Conduct for digital citizenship SIX: QUANTUM LITERACY 6.1 Quantum 6.2 Quantum Literacy (QL) SEVEN: CONCLUSION AND RECOMMENDATIONS. THE ALGORITHMIC CONDITION 7.1 Algorithmic Conditions – Summary Observations 7.2 Ethical Coding 7.3 Recommendations 4 8 12 14 16 17 20 21 26 28 30 31 34 36 37 40 41 44 45 50 EoC GLOSSOMATICS 47 REFERENCES 53 APPENDIX A: DEDA 62 3 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION AUTHORS Professor Felicity Colman, Kingston University, London, United Kingdom (KU) Professor Iris van der Tuin, Utrecht University, Netherlands (UU) Professor Aislinn O’Donnell, National University of Ireland Maynooth, Ireland (NUIM) Professor Vera Bühlmann, Technische Universität Wien, Austria (TU) RESEARCH ASSOCIATES Marien Baerveldt, MSc, Utrecht University and Art of Hosting, Netherlands Gerda Palmetshofer, Technische Universität Wien, Austria Rumen Rachev, Institute for Cultural Inquiry, Utrecht University, Netherlands Sam Skinner, Kingston University, London, UK Whitney Stark, Institute for Cultural Inquiry, Utrecht University, Netherlands STAKEHOLDERS AND PARTICIPANTS Diana Alvarez-Marin, ETH Zurich, Chair for CAAD, Switzerland Elie Ayache, ITO 33, Paris, France Jesse Balster, Utrecht University Computer Science, Netherlands Marguerite Barry, University College Dublin, Ireland Prof. Maaike Bleeker, Theatre and Performance Studies, Utrecht University, Netherlands Prof. Peg Birmingham, Philosophy, DePaul University, Chicago, MI, USA Prof. Rosi Braidotti, University Professor Utrecht University, Netherlands Christl de Kloe MA, Utrecht Data School, Utrecht University, Netherlands Pierre Cutellic, ETH Zurich, Chair for CAAD, Switzerland Dr. Sylvie Delacroix, Reader in Legal Theory and Ethics, UCL Laws and Computer Sciences, UK Adam Harkens, Queens University Belfast, Ireland Cliona Harmey, National College of Art and Design, Ireland Prof. Joanna Hodge, Professor of Philosophy, Manchester Metropolitan University, UK Dr. Yuk Hui, ICAM Leuphana University Lüneburg, Germany Dr. Thomas King, Digital Ethics Lab, Oxford Internet Institute, University of Oxford, UK Dr. Kristian Faschingeder, Department for Architecture Theory and Philosophy of Technics ATTP, TU Wien, Austria Georg Fassl, Department for Architecture Theory and Philosophy of Technics ATTP, TU Wien, ,Austria Dr Noel Fitzpatrick, DIT, Ireland Dr. Jessica Foley, CONNECT, Dublin, Ireland Ine Gevers, Curator Niet Normaal Foundation, Netherlands Aline Franzke, MA Utrecht Data School, Utrecht University, Netherlands 4 CONTRIBUTORS David Gauthier, MA, Netherlands Institute for Cultural Analysis, University of Amsterdam, Netherlands Helena Grande, MA Institute for Cultural Inquiry, Utrecht University, Netherlands Prof. Ludger Hovestadt, ETH Zürich, Information Science and Architecture, Chair for CAAD, Switzerland Noelia Iranzo Ribera, MSc, History & Philosophy of Science Utrecht University, Netherlands Dr. Laura Karreman, Theatre and Performance Studies, Utrecht University, Netherlands Dr. Aphra Kerr, Maynooth University, Ireland Prof. Rob Kitchin, Maynooth University, Ireland Dr. Nikolas Marincic, ETH Zurich, Chair for CAAD Prof. Jan Masschelein, KU Leuven Martin McCabe, Dublin Institute of Technology Dr. Chris Meyns, History of Philosophy, Utrecht University, Netherlands Eleni Mina, MA, Bio Informatics, Leiden University, Netherlands Dr. Vahid Moosavi, ETH Zurich, Chair for CAAD, Switzerland Philippe Morel ENSA Paris, France Prof. Marcel Alexander Niggli, Chair for Criminology and Philosophy of Law, University Fribourg, Switzerland Dr. Jorge Orozco, ETH Zurich, Chair for CAAD, Switzerland Joe Oyler Maynooth University, Ireland Poltak Pandjaitan, ETH Zurich, Chair for CAAD, Switzerland Prof. Sandra Ponzanesi Professor of Gender, Utrecht University, Netherlands Wessel Reijers, ADAPT, Trinity College Dublin, Ireland Angela Rickard, Maynooth University, Ireland Dr. Egle Rindzeviciute, Kingston University, London, UK Miro Roman, ETH Zurich, Chair for CAAD, Switzerland Dr. Mirko Tobias Schafer, New Media and Digital Culture & Director of Utrecht Data School, Utrecht University, NL Prof. Anne-Françoise Schmidt, Prof. of Epistemology, MINES ParisTech, Paris, France Dr. Oliver Schürer, Department for Architecture Theory and Philosophy of Technics ATTP, TU Wien, Austria Sharon Todd, Maynooth University, Ireland Whitney Stark, MA, Institute for Cultural Inquiry, Utrecht University & BAK—a base for art, science, and politics, NL Dr. Nancy Vansieleghem, Luca School of Art Michael Veale, UCL Department of Science, Technology, Engineering and Public Policy, UK Prof. Scott Wilson, Kingston University, London, UK Prof. Michael Wheeler, University of Stirling, UK Dr. Ben Williamson, University of Stirling, UK Bram van den Boomen, Utrecht University Computer Science, Netherlands Ruben van Doorn, Artificial Intelligence and Liberal Arts & Sciences, Utrecht University, Netherlands 5 ONE INTRODUCTION: THE ETHICS OF CODING [EOC] ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 1.1 EoC – AIMS AND SCOPE This report on the Ethics of Coding [EoC] presents a snapshot view through an investigation on the current state of what we call “the algorithmic condition”. By speaking of the algorithmic condition, we pick up today, in critical manner, Hannah Arendt’s question of the condition of possibility for leading an “active life” as the conditions of possibility for politics. For Arendt, this question emerged out of an altered status of knowledge that resulted in her time from being related to nature of the earth as investigated from a viewpoint in the universe, rather than one situated firmly on earthly grounds (Arendt, 1958). In doing so this report brings together discourses and objects, of the sciences and the humanities, and seeks to present a spectrum of the diversity of issues generated by this altered, novel condition, and survey of the wide-ranging considerations and potential applications of this topic. Further, this report on the ethics of coding and the algorithmic condition asks: How can we think adequately about the relation between knowledge and ethics in societies that are governed by algorithmic digital systems and objects endowed with agency? In order to attend to these latter questions, we looked at Jean-Francois Lyotard’s report from 1979 on the altered status of knowledge in “computerized societies”. Raising Arendt’s question of critique (and transcendentality of conditions of possibility) with regard to how we think about “human nature”, as well as by relating coding and programming to Lyotard’s particular notion of language games and paralogisms (Lyotard, 1979), we propose to take into account, from both viewpoints, an emerging novel “literacy” which we propose to call a “quantum literacy”. With this, we want to direct attention to the principle inadequacy of thinking about numbers and letters, mathematics and language, as two separate domains of which the former is concerned with the necessary whereas the latter deals with the contingent and interpretable. Code confronts us with an “impure” reason that cannot rid itself of amphiboly, but that is nevertheless “computable”. An algorithm is a finite set of instructive steps that can be followed mechanically, without comprehension, and that is used to organise, calculate, control, shape, and sometimes predict outcomes, applied across various fields. Algorithms are as old as mathematics itself (Ritter, 1989a, 1989b) and are used in multiple domains of human life; from making food using recipes, solving math problems in engineering, to controlling complex transport and distributions systems. Developed for calculating, and reasoning problems in a computational manner, algorithms are classically used in math. With the rise of computers and computation especially since the 1980s, more and more problems from more and more diverse fields are being treated as “computable” problems; which means they are being tackled in algorithmic manner. The outlook of Quantum Computing (cf. Chang et al., 1998; Shor, 1994), or even Universal Quantum Computing (propagated e.g. by IBM) prognoses to be able to treat all problems – be they as “wicked”, i.e. socially complex and principally unsolvable (Rittel et al., 1973; Churchman et al. 1967) as they may – as “computable” problems” that compute in term of Qubits rather than Bits. 8 ONE: INTRODUCTION: THE ETHICS OF CODING [EOC] The use of algorithms (for the detection of patterns in information; the organisation and analysis of data; the implementation of data codes and sequences; and other computational systems) in societies not only effects a practice-based shift in knowledge production and acquisition, but also produces a logic which is symbolic but also which manifests as a reality. This logic - which is logistically constituted – is what we refer to in this report as the algorithmic condition – that alters the cultural and social reality it organises, through its procedural dynamics – which we address as the ethics of coding. Algorithms engage the data as they are written, but their codes can and do evidence degrees of bias (Amoore, 2009; Bishop, 2018; Cheney-Lippold, 2011; Hayles, 2017); and the degree of code bias thus requires consideration of the ethics of coding. The domain of ethics is a customarily contingent, yet also dynamic field. In consideration of technology, culture, philosophy, and all fields of Information and Communications Technology [ICT], that enable an algorithmic condition, we focus here on the notion of “ethics” and make the distinction between the discourses and practices of ethics and morals; where “morals” refer to belief systems that impose value judgments, we think about ethics as referring to the encoding and decoding of types of practices; we think of this as an ethics because all encoding and decoding process involve encryption and deciphering that can embed different degrees of contingency and ambiguous determination. Hence, we find it useful to speak of an ‘ethos’ of coding, maybe even different ‘ethos of different eras’ (Hodge, 2017). An ‘ethos’ includes: 1. codes of practices (generally engaged by groups and individuals for regulation purposes); 2. rule-based systems that advocate specific groups and individuals; and or 3. existing orders of knowledge linked to those rules, governing what is deemed appropriate and or inappropriate behaviour (for example, when applied in relation to contemporary conditions that designate national identity and its ethos; as practice, power structures, and implementation “contracts” (Braidotti, 2006; Hodge, 2002, 2017). When one considers ethical guidelines for research innovations, as they are in use today, one usually encounters a perfunctory (or “tick-box”) approach to the consideration of “the ethics of research” involving degrees of human sampling, confidential information, and or security issues for data acquisition, storage, retrieval and processing has been the dominant practice when considering the “ethics” of a thing, design, practice, process, or situation. However, when we take into account some of the essential changes that innovations in ICT have brought to European communities across the sciences, business, healthcare, education, and social models, and at the level of the possibilities of “life” itself – such as in the field of bioinformatics as the science of collecting and analysing complex biological data such as genetic code (cf. Kennedy, 2016; Schäfer and Van Es, 2017), or in the area of technological developments of robotics, automation, and artificial intelligence (AI) (cf. Burrell, 2016; Greenwald, 2017; Powell, 2017), or in the arena of changing modes of work (Moore, 2018), or in the domains of energy (Hovestadt, et al., 2017), then it becomes clear that this radical change in the conditions for life and biosocial 9 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION relating requires a more specific form of ethical guidance; an ethics that is able to meet the demands of an algorithmic environment, and the forms of information and “knowledge” generated by this environment. Where ethics currently exist in ICT research and innovation their focus usually lies on the consideration of privacy, the security of data, or copyright law information (see Floridi, 2013). While these considerations are unquestionably necessary, for the use and implementation of new ICT models there are other aspects that have changed since the computerisation (and thus datafication) of societies since the late 1970s (Nora & Minc, 1980; Hayles, 2012; Lyotard, 1979); namely the ethics of the production and implementation of forms of knowledge in a digitised economy (as, for example, the “data economies” that are part of the Digital Single Market [DSM]). The question of the veracity of news information (highlighted in stories ranging from “affective visualisation” to “fake-news” in 2017) are a case in point of the significance of the question of ethical knowledge formation. For example, a focus in news is frequently of the ‘personal’ use of data, rather than the account of the effects of the algorithm/s in use (Edwards & Veale, 2017); and information on how the algorithm works remains opaque (De Laat, 2017). Points of comparison for the scale of change and significance of the algorithmic conditions of society, governance, and lifestyles include the Industrial Revolution (1760+), and the Datafication of Society (1978+). This report argues for an involvement with ethics that extends beyond existing tick-box models; instead advocating that an ethics of coding be responsive to the specificities of future ICT design, infrastructural forms of services, production types, and the types of applications required in the Digital Single Market (DSM) (including all of the initiatives in establishing the codes of conduct related to the DSM, such as the electronic communications code), the future Internet, and for the proposed development of multi-scaled networking technologies for the European Federated Experimental Infrastructure for Future Internet Research & Experimentation (FIRE). This report is meant to be of use for any infrastructure, system, network, organisation, institution, or individual engaging or participating with/within an algorithmic condition. Algorithms are not only a component of research – they are simultaneously part of the “toolkit” in the world of the Internet of Things (IoT). In view of this, their ethical implementation should be considered in terms of their contribution to any decentralised or cybernetic system, in terms of its use, transparency, data sovereignty, and any algorithmic procedures. The report is written in a generic way that enables it to be applied, not only to specifically European models of ICT, but also to non-European technologies, suggesting how they might be utilised within the European Union [EU]. 10 ONE: INTRODUCTION: THE ETHICS OF CODING [EOC] With this ethics of coding comes the requirement for a new literacy that is able to be expressive and articulate within the algorithmic condition; not simply a case of media literacy, but rather, as the report describes, the need for a new quantum literacy. Information, ICT and Media literacies are part and parcel of contemporary curricula (or they are simply assumed to have been acquired by students) at all levels of education in European countries, and globally, contingent upon access to ICT (see also Lambert, 2001). Informed by the transition to datafied societies, such literacies assume an interface culture of spatial navigation through texts, images and the built environment with mobile phones, tablet computers and laptops. We argue that the algorithmic condition extends beyond, and fundamentally changes, such spatial modes of relating by foregrounding the temporal logics upon which both interfaces and navigational practices rely. Navigating through nonlinear time changes the critical, albeit rather predictable, “jumping of scale” (Smith, 1992) and introduces unpredictable critical and creative activities in, and by, the data-technology-human apparatus (cf. Barad, 2007; Bühlmann et al., 2017). This report highlights the significance of understanding what it means to take an ethical stance in relation to the processual dynamics of algorithms, through an awareness of the types of codings produced in algorithmic societies. Firstly, the EoC report adopts the philosophical approach of Jean-François Lyotard (1979) to knowledge production, and resituates it in the three core coding spheres in which we find occurrences of the algorithmic governance of society; the social, ethical and educational realms, which collectively give rise to what we want to call “knowledge sovereignty”. Secondly, the report is informed by the ethical position of Hannah Arendt (1958, 1961, 1963, 1978, 1989, 1994, 2003), whose work reconsiders what the notion of ethics could possibly be after the types of human behaviour that occurred in Europe after the atrocities of the second world war. Arendt’s work asks how questions of the inequities of human rights might be resolved, and how the creation of a shared and common responsibility can be enabled through the collective actions of a community (see Birmingham, 2006). Further, the report notes that the algorithmic condition is altering what comes to be recognised as ‘human rights,’ if the condition lends itself to a state not only of change of the legal status for entities of all kinds, for example, through a change in militarisation practices (Schuppli, 2014); but also a change in the very notion of ‘community,’ through new forms of digital citizenship generated through algorithmic gatherings, and cloud-based commons (cf. Amoore 2013; Bühlmann et al., 2017; Colman 2015; Manovich, 2013; Terranova, 2004; Thacker, 2015; Tselentis & Galis, 2010; World Bank, 2016). 11 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 1.2 KNOWLEDGE: CODING ETHICS When we consider the relation between ethics and knowledge with regard to an “algorithmic condition” of possibility for “leading an active life” (Arendt, 1957), we need to acknowledge the immediate precursor of this condition; that of the relatively rapid uptake of computerisation across European societies. Lyotard’s report on what was termed “the postmodern condition” (Lyotard, 1979, 1984) describes this transition period – between the early paradigm of operations research, the cybernetics systems era (described in Wiener, 1961; and its consequences in Galison, 1994; Guattari, 1995; Kline, 2015; Rindzeviciute, 2016), and the era of personal computating after the 1980s (Lyotard, 1984). Examining the period from the 1950s to the end of the 1970s, Lyotard describes the “transition” in the information economy brought about by ICT (Lyotard, 1979, 1984). Lyotard’s philosophical approach to the alteration of knowledge argues for the need for new kinds of narratives, both for – and of – computerised societies. At its core, Lyotard’s report offers a critique of the “metanarratives” that had structured accounts of knowledge prior to the reconstruction of Europe. Lyotard highlights how the end of the so-called “great narratives” of knowledge marked a shift from their previously central and hierarchical role to a devolved understanding of knowledge production in societies. With this came a change in understanding the ways in which models of knowledge are generated and implemented. In terms of understanding this shift, Lyotard focusses on methods of expressing new forms of knowledge that are generated by computerisation, examining the forms of arguments, the concept of deductive reason, axiomatics, claiming statements, probable (modal) reasoning, rule-sets (computational grammars, transformational syntaxes, dealing with semantics in terms of corpora of data), and so on. Through the post- industrial digital devolution of hierarchical knowledge, Lyotard proposes a generalisation of Wittgensteinian “language games” instead of arguments and deductive reasoning as a means by which to engage with the new models of knowledge that were emerging in the computerised society of the late 1970s. Using a framework of play, Lyotard picks up the computational role of algebra. Thinking about algebra, we understand that play (the Wittgensteinian game) has a set of definite rules, but the rules can be combined into various systems and their terms can, accordingly, be factorized (or raised to their exponentials) in a principally indefinite amount of manners. In terms of their algebraicness, Lyotard’s notion of play deals with a “logistic logic”, with a symbolical system of substitute positions that establish a vicarious order and amphibolic order. Lyotard seeked to addressed this with his notion of “paralogisms”. The structure of a symbolic system differs from a system in analytical terms (fixed axioms and elements (basic notions), rather than a set of fixed but cominable and articulatable rules). In a game, the position of these basic notions are “empty”, “unoccupied” 12 ONE: INTRODUCTION: THE ETHICS OF CODING [EOC] (vicarious) and hence, within a certain limits, open (variously endowable/entitle-able with capacities and kinds of values, and hence adaptable in a great variety of manners), whereas, for example, in an axiomatic system, the rules are fixed as well as the definition and (predicate) valuation of the basic terms. Where predicate logic is fundamental for analytical (axiomatic) systems, in the play paradigm of Lyotard’s language game there is a logistic disposition on which the symbolic system depends. Following this logic of non-fixed meanings; speculative functions; and other novel modalities and behaviours (which have been observed in relation to “algorithm” and “data” [described at 2; 3.1]), the consideration of the ethics of coding in the plurality of algorithmically established and induced (generated and generative) economies that need to play together in one single market (Digital Single Market) ask for a manner of governing that accounts for all that challenges the EU’s project of a DSM, like digital (non-national) import/ export trading between digital economies and communities via crypto-currencies, and block chain organisation. Social coding occurs at this fundamental level of economics – which is a core part of the knowledge “turn” that Lyotard suggests – where game and game theory work on probability, prediction, and also behaviourism. This report’s interest in trading, currencies, probability theory, fluctuations, networked infrastructures, and education models (including the proposition for a Quantum Literacy [6.2]) thus highlights for any consideration of the algorithmic condition, the active and dynamic nature of epistemologies as well as ontologies, and the kinds of ethical coding engaged by that condition. The EoC report describes how new, emergent forms of knowledge generation in the period 1980–2020 require novel registers to address what we propose to call the algorithmic human condition. Ours is a period during which algorithms have enabled the use of data in both prescriptive terms (using algorithmic calculations to confirm previous knowledge categories) and in generative ways (where algorithmic calculations are exploratory, and open ended). This use of algorithms has been productive of different forms of knowledge, and “data” is now understood in terms of code that can be made to produce different things, and which has neither a positive, nor a negative nor anf definitive (intrinsic) ‘meaning.’ Code is code (and not number, character, letter or amino-acid) insofar as its relation to what it encrypts is entirely arbitrary. Following this, the report asks, if knowledge models are formed of and contribute to a reality that involves actions reasoned in terms of by algorithmic calculations, then what is the status of ethics in coding algorithmic conditions? This question targets that what we describe as the algorithmic condition is being both generated by, and generative of technic models (by data-technology-human apparatuses) rather than humanistic ideologies. 13 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 1.3 KNOWLEDGE: ETHICS CODING In addition to Lyotard’s attention to the technical, the modal production of knowledge, and the requirement for education to be responsive to such significant societal change, the EoC research considers the algorithmic condition through the lens of Arendt’s notion of the vita activa in her book The Human Condition (1958). In this text, Arendt set out what the framework for a twentieth-century ethics could be, set in the post European World War II context midtwentieth-century, booming cybernetics society and the development of information infrastructures (variously described in the work of Bell, 1973; Galison, 1994; McLuhan, 1964; Kline, 2015; Rid, 2016; Rindzeviciute, 2016; Siskin, 2016; Wiener, 1961). Arendt’s proposal for “The Human Condition” (1958) begins with a consideration of the novel technology of this era, most notably the Sputnik 1 satellite, which was the world’s first artificial satellite, launched by the Soviet Union into an elliptical low-earth orbit on 4 October 1957. Arendt uses this perspective to rethink the human condition as being unbound, face to face to nothing anymore, now that we can watch the world from a point of view in outer space. Arendt’s Human Condition narrates the shift in technologically-given perspective towards pluralist societies, especially that in terms of a two-fold process of individual expropriation on the one hand and the accumulation of social wealth on the other. Arendt’s reference and her subsequent discussion asks for a reimagining of the political realm that were adequate to these developments – in this pluralist technological world perspective – as actioned by people. Arendt’s vita activa further describes how actions are temporally bound and contingent in each case and, for each collective, also bound to their contemporaneous technological forms. Arendt’s framework questions how we might redesign the topology of contemporary experience, one which is given perspective through its technological communications platforms – what we describe today as ICT, the IoT, and also as Universal Quantum Computing models (see 6.1). Across the member states of the European Union, this technological topology is mapped via policy statements on the IoT, algorithms, data (in terms of security, health, governance issues), and in the infrastructural histories behind the use of algorithmic process organisation throughout Europe (seen, for example, in the development of the common market for coal and steel in Europe in the 1950s). The topology of code today comes in “algorithmic formality”: a play between “temporalized” and contingent symbolic systems that afford data to be recognised according to certain formal patterns. 14 TWO ALGORITHMS 15 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 2.1 THE ALGORITHMIC CONDITION Algorithms are used in all sorts of calculations and predictions. They are used in code encryption or breaking, data sorting, data management, information sorting, organisation, and analysis. With the huge data sets generated through the computing power of the twentieth century, nearly all infrastructural and organizational structures are being reset on a novel basis, that of algorithmic control. This triggers a kind of frenzy: the more kinds of problems that can be treated in this way, the more novel applications pop up to extend the scope of algorithmic control, planning, and generation of scenarios. The algorithmic control of coding information is employed across a broad area of humanoriented organisational activities, including transport systems, food production, processing, and distribution, health, education, finance, the media industries, and security. In addition to organisational procedural systems, algorithms are employed to process huge demographic data sets, which can be used, bought and sold as data sets in the market place. As identified by diverse scholars – from the analysis of the operation of Hindu-Arabic numerals by the Baghdad-based ninth-century scholar Abdullah Muhammad bin Musa al-Khwarizmi (Crossley & Henry, 1990: 104), to the twentieth-century culture and media scholar Wendy Hui Kyong Chun (2011) – the manipulation of codes and symbols of information by an algorithm engage in “code logos”. This refers to the way in which an algorithm operates according to how it is conditioned: “code as source, code as true representation of action, indeed, code as conflated with, and substituting for, action” (Chun, 2011: 19). In referring to the algorithmic condition, we infer the ways in which algorithms of all kinds play together and condition our environments and communities (cf. Amoore, 2009; Amoore & Piotukh, 2015), and the material intra-actions within those environments (we refer to “intraactions” not “inter-actions” here, signalling the algorithmic agency of the code at work, following Barad, 2007: 141). The actions of social; vernacular life are themselves information, which collated,constitute and allow to detect (encrypt and decipher) patterns in recorded data. 16 TWO: ALGORITHMS 2.2 ALGORITHMIC MODALITIES, AND THE BEHAVIOUR OF MATHEMATICAL MODELS In terms of ICT, the term algorithm refers to the encoding of computerised procedures, which can be subsequently used to extract data sets from anything. Amoore and Raley note that: “Algorithms increasingly have the capacity to analyse across different forms of data (images, text, video, audio) and across cloud-based spatial locations of data” (2017: 6). Since the mass computerisation of societies from the 1980s, the shift to a social environment – where the digitisation of communications enabled new modes of algorithmic conditions – has radically altered communities, produced new community forms, both actualised and virtual, and new behaviours (cf. Dobnikar et al., 2011; Finn, 2017; Floridi, 2015; Parisi, 2013; Simon 2015). What does it mean when the algos (a market traders term used in mainstream media for automating trading to generate profits at a frequency impossible to a human trader, thereby ruling out the human (emotional) impact on trading activities; it became popular in the media when the comments by the French President Hollande with regard to the concept of hard Brexit caused the British sterling currency to plunge in value 6% in 2 minutes in October 2016) acts upon the matter of information that has been put into action by the political field, affecting both the field and the information in ways unforeseen by the society that invented the algorithm? What does this kind of algorithmic behaviour do to the data sets that it rearranges? Algorithms put data into a coded state of performance, relative to mathematical models. For example, any regulation of the digital market occurs at the level of code and algorithm, with regard to sets rules based on pricing, quantity, timing, and other mathematical models. This is what has produced, and continues to produce, the material conditions made by market transactions that determine the forms and types of social, political, and cultural systems that direct, and are reflective of life today. Algorithms provide the set of instructions used to plan, control and predict outcomes in unstable environments, and with regard to problems that would classically count as too complex and unsolvable (“wicked”, Churchman, 1967). Algorithms are described theoretically in terms of how they can be understood as a recipe for achieving x; with the mathematical models put to work by algorithms, this x’s own and ongoing active behaviour can be taken into account of the performed calculation, as for example with the algorithms that organise data in social media, or the algorithms that organise data for transport safety. Algorithmic performance, and the mathematical models behind it, forever “approximate” their referent. But this approximation is not one in epistemological terms only, concerning a relation of representation. And neither is it one in ontological terms only, where an object would acquire indefinite “depth” or “withdrawal” (as object-oriented philosophers like to call it). It is performative approximation that produces what it will have referred to (when studied retrospectively). Prospectively, however, this indefiniteness is where the algorithmic condition and the human condition encounter one another: even when we have the set of instructions, there are multiple “unstable” factors that contribute to the possible dynamic outcomes that form and characterise x. These unstable factors and processual dynamics represent what the 17 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION report identifies as the algorithmic modalities that affect a model, its behaviour and its outcomes, leading to a range of forms of data use and manipulation (see Poser, 2013 on modalities; Rouvroy & Berns, 2013 on data governmentality). 18 THREE DATA ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 3.1 DATA Algorithmic infrastructures are creating a different kind of condition for using data, as Daniel Bell (1973) suggested. Data protection principles, formalised in the 1995 EU Data Protection Directive, and the 2016 EU General Data Protection Regulation (GDPR) seek to address some of the issues raised by the use of algorithms, and generated by different forms of algorithmic behaviours, but discussion of these conditions are so far limited when it comes to the facilitation of an ethical climate. In particular, as even mundane data might be mined, stored, and then transformed into collaborative filtering algorithms (and others, as computing capacity expands and novel algorithms can be written), into new data sets about individuals that may be unexpected or have potential to harm them, the data gathered about individuals becomes more distant from the data applied to them (e.g. through decision-making), placing existing legal frameworks under strain. One of the most important aspects of the algorithmic turn and the focus on data-driven decisions is the proliferation of pre-emption (rather than regulation), trying to predict patterns and behaviours based on standardised algorithmic models. This is what the Belgian juridical scholar Antoinette Rouvroy calls “data behaviourism”, which provides a “new way of producing knowledge about future preferences attitudes, behaviours or events without considering the subject’s psychological motivations, speeches or narratives, but rather relying on data” (Rouvroy, 2012: 1). What is new here is that this kind of data behaviourism does not simply produce subjects, since data does not consider subjects as “flesh and blood persons” – as concrete agents, driven by deep and complex wishes and feelings – but rather this type of behaviourism bypasses consciousness and reflexivity, and operates on the mode of alerts and reflexes. In other words: 20 THREE: DATA the algorithmic condition affects potentialities rather than actual persons and behaviours (as described by Edwards and Veale, 2017), yet at the same time, the condition influences human cultures. Hence, in considering the algorithmic condition, the inference is that we have moved from the Spinozist ethical question of “What a body can do” to a predictive code question of “What does the algorithm do to bodies?”. As the algorithm exerts a power over bodies that it directs, this constitutes a politics; entering the domain of ethics, and the question of Artificial Intelligence [AI]. Machine learning is tethered not only to the available databanks, but also to the contingencies written into the algorithm, experienced by the machine (in its glitches and power surges and infections, etc.), and ethical biases are cached within the patterns of data upon which algorithms are trained. At its core, data behaviourism is producing the same power structures from the past (for example, see Noble, 2018), and degrees of bias, but this time providing the legitimacy to solidify power relations that may have earlier been questioned and rigorously scrutinised. Describing this data state, Rouvroy suggests: We are no longer dealing with things, since there are no longer any things, there are no longer resilient objects: there are only networks of data evolving in real time and that aggregate from time to time as profile, patterns and so on. But raw data seem to be speaking by themselves. We no longer distinguish what used to come under the sign or the signal and the thing. What is lost with this entanglement is the possibility of critique. (Rouvroy in Rouvroy & Stiegler, 2016: 7) 21 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION Bias is a difficult word to translate across disciplines, as it is used with many meanings. Statisticians see bias as something that allows data to exhibit patterns, whereas social scientists of many flavours consider bias in the sense of emergent patterns which are undesirable, and can spring from a multitude of areas. Additionally, a range of computer science researchers have been involved in creating approaches to ‘de-bias’ algorithmic systems.1 These define fairness, often in terms of protected characteristics in the law and ways in which to integrate or mitigate them, for example during the training or refining of machine learning systems. Yet these systems assume that organisations have access to the sensitive data they need to mitigate these issues, such as race, sexuality or religion — which they often do not — as well as a definite classification of these in their grey zones, which could always be otherwise (Binns et al., 2017). Furthermore, issues of fairness stretch beyond protected characteristics into the justification of actions being taken, and it is thus possible to make a system that is perceived as unfair, even if equality law is strictly obeyed.Bias can be introduced in a number of ways, for example: 1 22 See www.fatml.org. THREE: DATA • • • • Data might represent historical patterns – such as entrenched racism or sexism – which we do not wish to replicate. As a result, it might hold a mirror up to the world, when a mirror is not what we want a decisionsystem to be based upon. (see also discussion in Rouvroy & Berns, 2013) Zipf’s law of minimal effort, whereby only a minority of users online generate content. Peter Lunenfeld (2011) has characterised this as a “secret war between downloading and uploading”, consumption and creation. This is essentially captured by the idea that one is not “sampling at random” from an entire “population” of data, and that even where the measurement of a phenomenon is agreed upon, data concerning it might not be measured or captured evenly. Algorithms may engender their own data bias, through the “presentation bias” of results to generate “clicks”, ratings or comments. Placing items in search results higher up the page is shown in studies to be “the most important factor used in determining the quality of the result, not the actual content displayed in the top-10 snippets” (Bar-Ilan et al., 2009). “Filter bubbles” are based upon a paradox whereby no algorithm can recommend something new that you may like, if it does not use data from other people (Pariser 2011), which at scale can produce a herding effect. Analogously, a recent study in the Journal of the Royal Society Interface showed that sheepdogs think algorithmically, and that algorithms make good sheepdogs (Strömbom et al., 2014). The study described an algorithm derived from sheepdog/herd behaviours in which a single individual, through the introduction of predatory pressure inducing flocking, which in turn make herds more easily “driven”, is capable of influencing the group behaviour of a massive and unwilling crowd. 23 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 3.2 DATA SOVEREIGNTY “Data sovereignty” is a term that tries to resolve such conflicts with regard to geopolitical issues – which law applies to data depending on where it is stored. It is a question of where and how data is hosted. Servers and providers used to be geopolitically localisable, however, this principle no longer applies in relation to cloud hosting. Technologies like blockchain offer new concepts of data sovereignty. Amoore and Raley (2017: 3) draw attention to how “the embodied actions of algorithms […] extend cognition, agency and responsibility beyond the conventional sites of the human, the state and sovereignty.” In the German language, the term data sovereignty is also linked to a discussion called “informationelle Selbstbestimmung”, which has been an (ongoing) discussion since the 1970s dealing with issues around privacy, and which in the context of modern data processing, the protection of the individual against unlimited collection, storage, use and disclosure of his/her personal data is encompassed by the general personal rights of the German constitution. In terms of international law, data sovereignty is often related to who owns the data – that is, in terms of nation states – due to geopolitical location (Vaile et al. 2013). However, this principle is in conflict with cloud computing (distributed storage). Homi Bhabha’s (2017) treatment of spectral, vernacular sovereignty in relation to memory (collective identity) is relevant here, as a means by which to uncover the philosophical problem implied by this notion of data sovereignty. 24 FOUR CODING ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 4.1 CODE DEFINITIONS The term “code” is being interpreted in several ways. The etymological roots of the term “code” are Latin – codex – meaning a tree trunk or block of wood from which sheaves/sheets of papyrus, or paper books were made. Code also comes from an explicit and systematical writing down of laws, with reference to Hamurabi and Justinian law. By the late sixteenth century codexes referred to collections of statues / laws, and specific codexes are still referred to; e.g. The Codex Alimentarius (Food Code). With regard to algorithms, code is often referred to in terms of systems of symbols used for information control and programming. “Code” can also be used as a term to describe conversions, and the writing of messages and instructions. “Code” in this sense refers to particular media of communication (as in Morse code), but also to protocols and rules for communication. In computer programming, to “code” (in certain languages) refers to the formulation of instructions, the running of protocols, and the setting up of algorithms to perform work. “Code” also refers to pre-specific language or economy of symbols that, like an alphabet (lexicality, grammar, syntax) rather than like systems of symbols (axioms and elements) that can represent other things indexically, and thus hide things, or enable access to things, as in cryptography through encryption keys and codes. It is in these senses that we investigated the term in relation to the consideration of the algorithmic today; code that preserves the “materiality” of the thing to which it refers (as in the materiality of the book), yet also is symbolic; and inclusive of the implications of quantum technologies and mechanics, where specific coding enables, governs, or directs societal and social operations. From an analytical point of view, it is significant that code produces “systems of symbols” (rather than a symbolical system) – precisely because of this, a structural approach and a systematic approach are not mutually exclusive: they are the two sides on one coin, and their interplay is generative on either side. This difference between “system” and “structure” is a problem that lies at the very core of all approaches that endeavour to formulate a “general linguistics” – at least since Ferdinand de Saussure (1857–1913). Taking into account its “materiality”, code can deal with this problem because it does not try to “solve” it, but rather provides for a comparative and comparatistic practice of how it can be “resolved” – in the sense of how we speak of a mixed substance being “resolved” in chemistry. In ICT code is taken to refer, not only to the formulation of a set of instructions (codes in general) for data in a computer programme or genetic information, but also as inferring the material and discursive data (social codes) that are used across all communication and media forms (Stocker & Schöpf, 2003: 15), all of which contribute to generic knowledge codes (educational canons, rules of law) that subsequently contribute to the forms and governance of communities. The work of the pillars named in the Juncker Commission on the Digital Single Market provide the contextualising framework for the codes produced in the European ICT environment in, and through, which the EU’s nations must work together, in terms of the law 26 FOUR:CODING and its ethical applications.2 The conception of “Europe” itself shifts under this condition, as much as it does under the ethical shifts identified by Arendt (cf. Balibar, 2015; Gasché, 2008; Hoskyns, 1996; Tselentis et.al., 2010). In its consideration of the different forms of coding, as a working hypothesis the EoC assumes that, if we are communicating not only numerically (counting) but also in units (measurements) of data through specific sets of algorithmic instructions that organise procedures, then that algorithmic condition is altering knowledge forms, their ontologies, and epistemic and classificatory fields via the formality of informatics generated, and the fields of possibilities enabled. Such enabling is selective, it is empowerment as well as disempowerment. The implications of machine learning can be generative of forms of “moralism” that are different to the ethical considerations of technology, and are not helpful for engaging an economical ethics of resources, work, and governance. However, machine learning can also contribute to the generation of “local” morals or simply idiosyncratic “manners of conduct”. Such local morals may embody newly acquired experiences that do not in themselves lay any claim to any formal status, or to be of general validity. “Moralisms” only arise when local customs claim generalised validity. Thus, an “ethics of coding” does not seek to contradict the accreditation of local morals; rather, it is an ethics precisely in that it can embrace non-generalisable situatedness by making their particular kind of encoding translatable, and hence compatible, with the encodings of other local morals. The coding of concepts happens twofold, and in parallel, at both an ideological and informatic level. In terms of the symbolic and historical levels, these are maintained and driven by two things: the recorded expression of codes (by specific laws, cultures and communities that are generationally and politically contingent), and the valuation of its actions. Within algorithmically governed areas of society, concepts are not coded at these socio-political levels (albeit that through data behaviourism coding for the future is entangled with power structures from the past), but by informatic, modal operations. It is this latter focus that deserves attention in its own right, and that is of specific interest for us when looking at the three fields that we consider subordinate to what we call Ethics of Coding in the Algorithmic Condition: Social Coding, Ethical Coding, Educational Coding. 2 https://ec.europa.eu/commission/priorities/digital-single-market_en 27 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 4.2 SOCIAL CODING In consideration of social coding, we ask: What are the areas of social governance in which such a Code Logos may be evidenced – in terms of the design of infrastructures for communications; the direction of funding for communications (and for whom)? How do contemporary social coding forms contribute to forms of knowledge about society? What are the models and conditions that enable the forms that societies take, given their particular algorithmic conditions, as well as their larger, environmental algorithmic condition? Finally, we ask: what are the shared understandings of ethics generated by these conditions? Forms of social coding capture the informatics of the algorithms at work. For example, Chun (2011) refers to the notion of a “code logos”, which urges us to think about what the frameworks of normative coding are that are enacted today by digital technologies that enable coding to occur. Social coding is used across, and is implicit in, the design of a range of ICT platforms, including network architectures that regulate the control of security identity data, digital social platforms, and collective awareness platforms, defining the biopolitical terms of heredity, genetic codes, gendered codes (see Van der Tuin, 2015), class coding, ethnicity coding, ideological coding, theological coding, and economic coding – all of which define rights of access as well as support states of sovereignty. In the context of the algorithmic condition, the notion of “data sovereignty” within the legal framework of the “European Union” is representative of the shift in codification of social knowledge forms. In The Coming of Post-Industrial Society: A Venture in Social Forecasting (1973) Daniel Bell argues that the most important shift for society is the “codification of theoretical knowledge”. Linking Bell’s conception of social coding to Susanne Langer’s concern with “The New Key” in philosophy (as just one of many such examples), we can see that even as early as the 1940s, the change in social epistemic structure brought by knowledge shifts has many ramifications for societies: Here, suddenly, it becomes apparent that the age of science has begotten a new philosophical issue, inestimably more profound than its original empiricism: for in all quietness, along purely rational lines, mathematics has developed just as brilliantly and vitally as any experimental technique, and, step by step, has kept abreast of discovery and observation; and all at once, the edifice of human knowledge stands before us, not as a vast collection of sense reports, but as a structure of facts that are symbols and laws that are their meanings. A new philosophical theme has been set forth to a coming age: an epistemological theme, the comprehension of science. The power of symbolism is its cue, as the finality of sense-data was the cue of a former epoch. (Langer, [1942] 2009: 16 original emphasis). 28 FOUR:CODING The formulation of the algorithmic condition as a state contributes to changes in cultural epistemologies concerning issues of data (such as how we understand what forms of computation, time, memory, archiving, and storage might be), communications, and the notion of “history” itself. As commentators including Bernard Stiegler, Donna Haraway, and Félix Guattari note, a “lack of consistency” (the “contingency” of any “foundational” point) is one of the traits of the technological systems, machines, and subjects produced in, and by, post-industrial capitalism (cf. Guattari, 1995: 48; Haraway, 1997: 121; Stiegler, 1998: 177). With this modal mode of operation come new expressions, and new ways of articulating the conditions of the present. In the chapter “Locked In: The Algorithmic Basis of Sociality” in her book The Culture of Connectivity: A Critical History of Social Media Dutch media scholar José van Dijck notes: Connectivity quickly evolved into a valuable resource as engineers found ways to code information into algorithms that helped brand a particular form of online sociality and make it profitable in online markets – serving a global market of social networking and user-generated content. (Van Dijck, 2013: 4) Van Dijck’s focus and language details the use of social media platforms and how they engage algorithms to service the market place. By attempting to critique in traditionally modern and postmodern keys, we can only project power relations from the past, the present and into the future onto the ICT platforms, whereas we know that both they and their workings are determined yet at the same time not exhausted by such power relations. Summary Points on Social Coding To speak of an ethics of “something” indicates a possessiveness; the idea that there is a realm already defined (such as the uses of big data by commercial companies such as Google, which may be said to purchase the data, or use the data made available to them in the marketplace). As we engage with some of the nuances of (context contingent) social coding, those prefigured areas are readily identifiable. As Lyotard demonstrated, the concept of knowledge has changed. However, we might look to critical models of “social physics” or sociology to consider further the collapse of knowledge models. For example, as French philosopher Auguste Comte’s three-stage positivism (see Scharff, 2012) identifies in the third stage (industrialism) how we no longer look for causes, origins, etc., rather there is nothing to be learnt anymore (learning is characteristic for the second stage; the spiritual/metaphysical one). In the third stage we only concern ourselves with the laws of what is known, and this idea is in conflict with both coding and programming. 29 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 4.3 ETHICAL CODING We have looked at “ethical coding” through the lense of what it means to be “citizens” in the algorithmic condition (Bühlmann, 2018). To ask (and think) what it means to be citizens in a digital world infers a capacity to act responsibly. Yet with regard to whom, and within which domain? In considering what coding ethically – and an “ethics of coding” – entails, we endeavour to maintain an open mind, by attending instead to modes of exchange, translation, and transference. Hence the consideration (for ICT, the IoT, technology design), of a new role of “economy” to address judiciary laws through coding (Ayache, 2010, 2017). When we begin to explore algorithmic conditions that give rise to ethical coding, we note that the language used to express this condition shifts between a consideration of attention to the differences between the object and/or concept produced, and the material, technological platform generating it. Here we have to address the conflation of the product(s) of algorithmic infrastructures – the technical objects (data sets; digital infrastructures) – with their conceptual etymology and “promise”. “It is injustice, not justice, which brings us into normative politics” (Bhabha, 2017 citing Avishai Margalit 142). Homi Bhabha takes Margalit’s observation as his point of departure for what he refers to as spectral sovereignty – where the concept of a “nation state” continues, even after its form has been altered, changed, or perhaps even destroyed (for example, the “nation states” of the indigenous peoples of Palestine or Kashmir). Following Bhabha, Bruce Robbins and Paulo Lemos Horta argue: The nation state persists in spectral or compromised form, [as an] absolutely contemporary, part of any properly cosmopolitan aspiration in the digital era … [introducing into] identity a primordial indefiniteness – one might say, a refusal to be pinned down by the question, ‘Where are you from?’ For Bhabha, this indefiniteness parallels the role of dignity in the discourse of human rights: it is the proper basis for a cosmopolitan ethics. (Robbins & Horta, 2017: 13) Judith Butler (2004) also refers to the concept of “spectral sovereignty” in relation to the discussion of precarious life, questions of power, ethics, violence and mourning – which Mary Lou Rasmussen and Valerie Harwood (2009) apply in relation to issues of contemporary governmentality and the nation state’s approach to “identity”. If we can be civic citizens of this digital world, how then might we be lawful within this digital world? What is its relation to jurisprudence, and to law? Must we be ethical in a novel way because laws are being mechanised (Niggli, 2014)? What does an ethics for coding consist of? How to lead (as Arendt puts it in The Human Condition) an active and a free life? (see Birmingham, 2006) Within algorithmic conditions we ask, what is digital citizenship? 30 FOUR:CODING 4.4 EDUCATIONAL CODING “Educational coding” in the context of this report is concerned with schooling (conventionally understood) and with educational institutions. It also seeks to examine the kinds of education, schooling or literacy that are required to engage with the algorithmic condition, across a wide range of institutions. Commitments to teaching digital literacy, coding, and even algorithmic literacy have, thus far, failed to engage sufficiently with the algorithmic condition. The literacy required to address this condition extends beyond simple programmatic interventions or skillsbased approaches that seek to develop competence in becoming more adept users of technology or even a better understanding of how big data works. Coding literacy does not involve simply introducing Science, Technology, Engineering and Mathematics [STEM] students, designers, engineers and computer scientists to arts, humanities and social sciences curricula, or doing the converse for students in across the science, technology, engineering, and either art and mathematics, or applied mathematics [STEAM]. An extensive literature documents and examine the risks posed by algorithms – not only in terms of coding (for bias and discrimination) – but also in terms of their application and interpretation, whilst noting how algorithms can alert us to our own biases (Kitchin & Dodge, 2011). The opaqueness of pervasive machine learning algorithms requires additional ethical considerations and safeguards. However, as Alison Powell articulated in her evidence to the UK’s Science and Technology Committee consultation on automated decision making, transparency alone is not enough to make algorithms accountable (Powell, 2017). For example, traditional notions of transparency do apply to neural networks that distribute intelligent decision-making across linked nodes, in which the decision-making process cannot be unpicked (cf. Burrell, 2016; Kroll, 2017). Furthermore Paul B. de Laat (2010, 2017) observes that transparency is complicated by the fact that underlying data sets cannot be made freely available due to privacy constraints machine learning models may promote “gaming of the system”, thus a solution or salience would appear to lie in the use of intermediate parties involved in their application. As algorithms interface with datasets, the literature highlights the dangers of creating feedback loops that further target and exclude marginalised populations. As a consequence, understanding big data means understanding the construction of “evidence” and the responses to such “evidence”, in a manner that remains vigilant and critical, and that does not see such outputs as definitively authoritative. Yet, we can also make algorithms and big data our objects of study through a variety of lenses and, in the pedagogical context, a “science of relations”. However, the critical leverage derived from exposing the limitations of contemporary algorithmic culture and its practices can – important as this is – promote fear and impede opportunities to see the benefits that can arise from these practices. In the context of education, 31 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION a good deal of contemporary discourse surrounding technology involves creating preventative and awareness-raising interventions, targeting vulnerability, safeguarding, cyber-bullying, and privacy. In this respect, programmes and responses follow the model of public health interventions rather than developing educational responses to the contemporary questions arising from algorithmic culture and questions of algorithmic governance. Making something an object of study is part of this process of schooling – the “suspension” that allows for a non-instrumental engagement with whatever is on the table. In education, something is put on the table to allow us to gather around it (cf. O’Donnell, 2012, 2014). Masschelein and Simons (2009) take up this Arendtian metaphor in order to show the importance of mediators – the object, the matter at hand, the world – in the educational endeavour, and the process of suspension required in order to study it and detach it from its use value. This offers a promising way of thinking about education more broadly, in the sense that it invites a range of bodies, such as research centres, industries, and other organisations, including public sector bodies, to introduce “educational moments” that are uncoupled from the demands of production or consensus. As part of an ethics of coding, this would cultivate and promote practices of listening, enquiring, singularising, and studying. Bernard Stiegler’s (2012) re-appropriation of the concept of the “pharmakon” invites both vigilance and generosity when responding to practices and this ethic in the educational context, and his attentiveness to the co-imbrication of the human/artefact coalition and the co-genesis of humanity and technology allows for other, more complex stories to be told. For Stiegler, we become human through our processes of mediation – techne is part of becoming human (see also Vlieghe, 2013). It is not a matter of being for or against technology per se, but rather, of thinking about education and the “grammar” and the “objective” of schooling itself. 32 FIVE ETHICS AND MANNERS OF CONDUCT ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 5.1 ETHICS The evaluation of the ethics of data use, and or the ethical ecology of decision support systems that engage with data, happens across a number of situations within any given algorithmic environment, or condition. Edwards and Veale (2017) describes forms of evaluation of algorithms, in relation to public sector practitioners of machine learning, whilst Guy Abel (2010) notes a similar process in relation to international migration and population behaviour and change. Additionally, as national debates in, for instance, the Netherlands (Kool, Timmer & Van Est, 2015) and France (CNIL, 2016) reveal, there are multiple positions on what actually constitutes an “ethics” of behaviour. The very notion of marking “change” per se, as a registration point, first of all makes a presumption of something “new”, and second, presumes the possibility of marking its correspondences, predicting its intentionality – significant and insignificant – and what Poser describes as “physical and causal necessity” (2013). The action of singling out a specific instance (for example, asking which ICT or IoT project has answered its brief in the most appropriate way) involves quantitative measurements that assess meaning through its “critical mass number of normalized instances” (Bühlmann & Hovestadt, 2013: 14), such as we see in the statistical measurement of the sales of objects through informatics algorithms. When dealing with social change, the measurement of the mass and the marginal engages qualitative judgments (Schwab, 2016), and a marking out of differences that can lead to moralistic pronouncements, which is fundamentally different to an ethics that seeks to insert and implement a communityactioned, political ethic (Arendt, 1958).3 Ethics can be defined, using a term from Thomas Gieryn (1983), as “boundary work”. Ethics comes from the Greek term ethos, meaning “habitual character and disposition” in plural, “manners”, and refers to a person’s, community’s, or institution’s activities and conduct. In contrast, “moral” comes from the Latin term moralis, referring to the proper behaviour of a person in a particular community or culture. Given the comparative and integrative vector of ethics (that seeks to abstract from particular morals and make them compatible) we speak of an Ethics of Coding in the Algorithmic Condition in order to address the level of society (not communities), for which we regard as constitutive pluralist and also arbitrary ways of institutionalizing morals (see Benveniste, 1969; Lambert, 2001). There are many aspects to consider with the use of these two terms. One that seems important to us is that, with regard to coding, ethics has less to do with something “being well understood” and “finding a common ground”, and the comfort that goes along with this desire, 3 34 E.g. MIT’s work on a “moral machine” for IOT/ automated vehicles makes value judgements with broad criteria. http://moralmachine.mit.edu/ FIVE: ETHICS AND MANNERS OF CONDUCT than with affirming the necessity of not being entirely comfortable, as individuals, in (public) situations: such forms of uncomfortability ask for what we call “forms of conduct” – those forms can be encoded and deciphered in various manners. Hence there is room for “play” with regard to forms of conduct. In that sense, codes of conduct provide “lenses” through which to look at formalised concepts that are taken as norms and standards. When codes of conducts are instituted, they are typically rendered in terms that claim a general validity as standards, norms, etc. Such processes of generalisation are, of course, different in the case of standards from those of norms. The distinctions with which we work involve the following: A standard can constitute a system (as in IT standards), whereas a norm is derived from an established system, or from a projective, ideal system. These distinctions are important, because in ICT we have networks that are inadequately addressed when regarded as systems. Systems is a term in logics (elements and axioms), whilst networks is a term in logistics (particles, nodes, and distributions). This distinction has important implications for how we reason, assess, and evaluate systems and networks. “Codes of conduct”, as we understand them, open up interrogative, critical spaces with regard to behaviours – either those that are prescribed normatively, or those perceived as self-evident or “natural”, as habits, customs and traditions. Codes of conduct thus involve bearing with the discomfort that stems from “coding” such behaviour to constitute and fit for public domains where one’s own custom is just one of many possible forms of behaving. We call “coding a habit” the moment of opening up a gap of arbitration, of which we need to be conscious and which must be given a form in order to “stay with the trouble” (Haraway, 2008). This is the condition of possibility for translating between different behaviours without imposing an inflexible hierarchy. The paradigmatic shifts across societies due to the algorithmic conditions that now organise life thus require more than just velocity and form predictive tools with which to measure and critique normative modes of “quantitative standardization” (Bühlmann & Hovestadt, 2013: 14-15), in order, not only to account for shifts in organisation theory, but also to have ethical procedures incorporated into all decision support systems that engage with data. 4 Thus, a key question for an ethics in – and of – the DSM environment (for e-health, population, migration, security, communications, education, and research) is: How can we develop, use, and evaluate decision support systems that engage with data? The question that arises is: How does one become a digital citizen with a clear commitment to ethics? 4 For example, all of the current initiatives in establishing the codes of conduct related to the DSM, such as the electronic communications code https://ec.europa.eu/digital-single-market/en/news/proposed-directiveestablishing-european-electronic-communications-code 35 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 5.2 DOING ETHICS Mirko Tobias Schäfer (2017) notes that the Data Ethics Decision Aid (DEDA, developed by Aline Franzke and Schäfer 2017) enables new ICT and new technology project groups to think through the stages of their creation, design, or plan, in terms of the whole assemblage and use of data sets generated by that assemblage; allowing the project designer to ask who manages, and who is responsible if things go wrong. Who will produce documentation of ethical issues, and make decision-making processes transparent and public, and how and where will this take place? How does the project / object develop accountability so that a governing body is able to explain what they do with the data, and who has access to it? Schafer notes that such questions add a new layer of deliberation within democratic societies. Schafer develops a data ethics decision aid to assist with such questions. It follows a flowchart of steps to guide the designer of new data products (hardware and software) (see Appendix A). The steps involve consideration of the following elements: > 1. Algorithms > 2. Source > 3. Anonymity > 4. Visualisation > 5. Access > 6. Open Access and Re-use of Data > 7. Responsibility > 8. Transparency and accountability > 9. Privacy > 10. Bias > 11. Informed consent These topics provide the categories (and not the classification) for quality standards and the process of certification. The distinction between categories and classes here is crucial: categories establish the structure of orders of belonging, it is in that sense that they are relative to criteria. Classes, on the other hand, are relative to criteria in how they administrate such orders of belonging (and thus they presuppose the structure of such orders as given unproblematically). Certification in this context is understood in a technical manner; that is as an exchange of trust, and not as a document of “proof” (such as in the passing all check list points with a certain percentage). In the algorithmic condition we note that trust is an ethical arrangement that is different from the concept of proof. There are different lineages for ethics-ready certifications, where a user pays in order to attain a certain degree of “competency”. However, use of the DEDA is not, we suggest, to be taken as a certification mark. Rather, it should be used in conjunction with a code of conduct in order to generate an explicit ethics with regard to a particular code of conduct, specific to the community of the user. 36 FIVE: ETHICS AND MANNERS OF CONDUCT 5.3 EOC –ETHICS BASED ON CODES OF CONDUCT FOR DIGITAL CITIZENSHIP Following the Data Ethics Decision Aid (DEDA) model of an ethics process for thinking about the design of new technologies, we propose an ethics based on codes of conduct (rather than an “ethics tool”) as a framework to be considered alongside the DEDA. This ethics interrogates critically particular codes of conducts by taking as its guiding points the following for consideration of users of the DEDA. Such treatment of Codes of Conduct asks for consideration of: 1. Knowledge sovereignty of the thing/condition/concept (knowledge as a shared and public domain that is not “owned” by anyone), 2. Local processes of certification (productive of an ethics in a particular domain, eg; architecture, or automotive industries); 3. Legal basis for certification of the thing/condition/concept (an ethics that registers a “maturity” with legal capacity, enabling certification/ technology readiness). Each category requires different levels, generating an explicit form for “codes of conduct” for digital citizenship within specific communities of practice – for example, in engineering, architecture, health, education, or security. Such a process causes us to reflect on a new notion of the public, one relative to standards and practices in applied cryptography (e.g. the generalized usage of block chain technology and crypto-currencies beyond the field of banking properly, to re-organize institutions and corporations at large). This is especially important because in our view it provides a novel domain for policy writing, that concerns the self-governing of the “communities of witnesses” that replace former administrative hierarchies when organizations and institutions are being re-thought in terms of block chain technology. The claimed “absolute transparency” of these communities of witness raises novel challenges for policy writing to respond to, and it posits the possibility for policy writing to find a way out of its own modernist bias (protocols of political correctness and tick-box forms as demonstration of proof do not, per se, instigate a more ethical behaviour – it may even trigger the opposite). It should be noted that there are two main philosophical traditions of how to think of “law” (in the European, but also in other contexts). The United Kingdom follows the common-law tradition (as does the United States), whilst most European countries follow the civil law tradition. This creates confusion, not only at the level of international law, but also for the law of nations (Völkerrecht), as it applies to UNESCO, public international law, and European law. These are all different “levels of abstraction” with regard to the “force of law” that come into effect with regard to diverse situations. Inevitably, there are many conflicts between them, some of them lurking in the religious cultural roots (Catholicism and Reformation). The understanding of “sovereignty” (in the context of data sovereignty, and as we suggest: knowledge sovereignty) crucially depends upon these different traditions in the philosophy and history of law: the sovereign in common law is not the same as that in civil law. Homi K. Bhabha (2017) addresses 37 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION this (at least implicitly) when he speaks of “spectral sovereignty (ghosts of national sovereigns that work on theological and affective energy), “vernacular cosmopolitans” and “cosmopolitan memories”. In other words, an ethics of coding consists of the identification of the formal terms for codes of conduct for digital citizenship within specific communities of users – to be set by that community. In consideration of codes of conduct and the use of the Data Ethics Decision Aid to generate a code of conduct based ethics for a specific system or object, Intellectual Property [IP] cannot play a fundamental role anymore. 38 SIX QUANTUM LITERACY ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 6.1 QUANTUM What are the implications for generating an ethics of coding for algorithms, applicable across different social, educational, and philosophical applications (see Bühlmann et al., 2015), within the field of information as we understand it today – namely as data (Hui, 2017: 235) – a field in which the possibilities afforded by quantum physics instigated a huge categorical shift in physics itself, as well as in metaphysical articulations? (Deutsch, 1985) The term “quantum” is derived from the Latin and in the general sense refers to the relation between quantity (counting, “one”) and magnitudes (measurements, “units”) (see work by German mathematician Emmy Noether on the role of magnitudes [cf. Weyl, 1981]). After the work of scientists in the twentieth century, including Niels Bohr, Albert Einstein, Max Plank, Werner Heisenberg, Erwin Schrödinger, John Bell, Richard Feynman, David Deutch, and others (Lévy-Leblond, 1976; Whitaker, 2012), a shift from classical towards quantum thinking inserted an understanding of the fields of relationality into empirical and philosophical analysis, thus reconfiguring the conception of epistemologies as well as, more recently, that of ontologies (see the trends towards Data Ontologies and Semantic Web), in all disciplines. Quantum-generated epistemologies are articulated by thinkers such as Karen Barad (2007) and Arkady Plotnitsky (1994, 2006, 2009); and evidenced in popular culture, where disturbing questions of the question of algorithmic reality, AI, human, and non-human agency are explored, such as in episodes of the television series Black Mirror (Charlie Brooker, 2011-2018). In the algorithmic condition, reference to quantum refers to the rendering of appearances that cause all quanta phenomena visible and material (photons move in particles and waves: light, literally, is “material”(particle view) as well as “immaterial” (wave view) (see Feynman, 2014; Colman, 2017) – in both technical, and spectral senses. Within this condition, how the quantum processing of data will effect the categories of time, space, light (as energy and mass), and all of the framing discourses of life, approximating such issues as nature, culture, aesthetics, technology, politics, science, philosophy, gender, ethnicity, theology, health, food production, communication, and information remains to be observed. It is important to note that in this physical term of “quantum”, all electronics based computers (as opposed to analog computers, which work on the level of electric control only by operating with tubes directly, without an intermediary of code-based arbitrage) are already to be regarded as “quantum computers”, in that they involve the same mathematics (complex analysis rather than real analysis) as applied in quantum physics (see Brillouin, 1969). The rather recent hype with regards to a kind of “Quantum Computers” that are said to introduce an entirely new generation of computing involves a different paradigm of “logical computability”: it proposes to go from the Universal Turing Computing paradigm to one called “Universal Quantum Computing” (e.g. IBM’s marketing language). The promise this makes concerns the dealings with the so-called Halting or Decision Problem in the Turing Computing Paradigm, and is not primarily a question of physics, but one of logics and metaphysics. (This is something we will investigate in detail in our continuing work after this report). 40 SIX: QUANTUM LITERACY 6.2 QUANTUM LITERACY [QL] As quantum conditions become knowledge generators in what we call the Algorithmic Condition, new syntaxes, semantics, and grammars with regard to digital informatics are required with which to express and articulate this era’s socio-material reality, and the formalization of the codes of conduct it is enabling/instituting. We argue that we are currently living in a novel era of sophistics, for which a quantum literacy is required in order to articulate the algorithmic condition (see Bühlmann, Colman & Van der Tuin, 2017). Our emphasis on a literacy (rather than a logics) means to express that a literacy empowers one to make sound arguments as well as to lie with arguments, to compose poetry as well as prose and reflective accounts. The reason at work in sophistics is not “pure” but “cunning” and “witty”. The political and philosophical conditions of accountability and responsibility in the novel Algorithmic Condition are yet to be engendered (if they are not to institute an orthodox form of governance). As Lévy-Leblond, J.-M. argues: It is quite clear that in the actual practice of physics, no one can be content with the use of the sheer mathematical formalism, even though this formalism is a necessary and fundamental constituent of the considered theoretical domain. A metalanguage is necessary as well, so that the names given to the mathematical objects and formal concepts of the theory enable its statements to fit in the general discourse. The choice of the terminology thus is a very delicate affair, with deep epistemological implications. If adequate, it may greatly help the understandability of the crucial points as it may hinder it in the contrary case. (Lévy-Leblond, 1976: 171) 41 SEVEN THE ALGORITHMIC CONDITION: CONCLUSIONS AND RECOMMENDATIONS ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 7.1 ALGORITHMIC CONDITIONS – SUMMARY OBSERVATIONS This report draws attention to the question of ethics in relation to ICT issues in Europe; the role of education in addressing the change in knowledge generation, its conditions of production and exchange in this society; and the need to develop new forms of literacy that are responsive to this condition. The algorithmic condition can be characterised in a number of ways: • Algorithmic conditions arise from economic infrastructures that direct knowledge; • Algorithmic conditions arise from calculations can be used for analysis that can be exploratory, or for compliance (as in corroboration of data sets); • Algorithmic conditions give rise to social coding, and the language used to express this condition shifts between a consideration of code in terms of control and governance, as well as a consideration of code as generative and open; • Algorithmic conditions can be understood as informatic codes (not as pre-determined symbols) that are generative of novel forms, things, and concepts, and taken as material-semiotic indicators. They are used to govern the domains of information flows, including security, privacy, communication, transport, distribution, big data, machine learning, and algorithmic reason; • Algorithmic conditions are generative of a particular ethics concerning codes and formula; cryptocurrencies, digital citizenship, the acquisition of privacy, architecture, and the speed of thought; • Algorithmic conditions are used as ethical matter – expressed as positional, economic, social, political measurement and governance “tools”. However, the rationality of these tools cannot decide the ethicality of the matter at stake – it can only resolve it, in a quasi-chemical manner that produces novel aggregations and constellations in the symbolical fabric of “ethical matter”; • Further, the question of “ethics” itself is inevitably a question of “force” as it is implemented by the ethos of its technological platform (Hodge, 2017). Additionally, the tag of “ethics” can be used as a carrier for the potentially unethical (Powell & Nemorin, 2017); used merely for compliance – ‘tick-box’- requirements. 44 SEVEN: THE ALGORITHMIC CONDITION – CONCLUSIONS AND RECOMMENDATIONS 7.2 ETHICAL CODING When working with an algorithmic environment (e.g. the DSM) in the generation of information economies by platforms such as Google and Twitter, the issues of agency and governmentality in relation to the condition of the community are raised. However, the ethics of the algorithmic condition is determined by the practices and habits of digital (world) citizens within local communities. The actions of the algorithm are frequently hidden by its systems embedded within infrastructures of communication and distribution of datamining and warehousing, code, and in the political, financial, and economic applications of that data as code in a range of applications, including security, secrecy, legal, educational, and commercial uses. However, education of the ways in which the condition operates and is generated – and with what potential outcomes – can enable an ethical environment that is not concerned with the commerce of weaponising, or controlling, but with more sustainable engagements with life. As Arendt (1958) argues, the actions of people are what determine and create the political fabric of a collective community. Membership of this community is contingent upon the terms of responsibility created through actions of living. Change, as Arendt describes it, is propelled in some instances by technological change. This kind of change can be articulated as relational change – that is, it occurs as a result of political and or ecological crises (such as we see in the after-effects on communities post-war, post-famine, or post-state of emergency). However, when examining the variations of ethics of coding that occur, we see that change is more readily understood as a resulting from a change in the ethos of a community and also from the contingencies of locally situated algorithmic conditions. With the implementations of a computerised society, a new community of users is enabled, and new forms and different uses of ICT emerge, shaping the nebulous algorithmic conditions of the era of 1980–2020. Ethical coding in ICT platforms engages multi-levels of standardisation, that are adjusted according to fluctuations in social coding (as caused by events; natural or man-made disasters; accidents; cultural and political changes that affect societies; innovations in technology; changes in knowledge produced through ICT that affect research across all areas of life; healthcare; security; governance; business; and citizenship). We conclude that in order to address these issues ethically, we need to approach them from a transversal/operational point of view rather than by simply gathering situational knowledge of how the concrete consequences of such multi-level standardisation can be documented today. To address “the algorithmic condition” with regard to ICT platforms only would be a reductive action when our interest is in a corresponding ethics; we need to take into account also these platform’s derivative “non-spaces”, such as the darknet, etc. Hence, the issue of “standardisation” must be addressed before and relative to their different backgrounds: epistemology and ontology, economy, political theory, and law. 45 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION 7.3 RECOMMENDATIONS 1. 2. 46 When developing, using, and evaluating decision support systems that engage with data, it is recommended that the Data Ethics Decision Aid (DEDA) is used in order to generate an ethics based upon Codes of Conduct. This ethics should be detailed as part of a support system’s specifications, and can be noted at the technology-ready point of development of new systems. The need for the development of a quantum literacy that is able to be expressive and to articulate the various domains that pertain to the algorithmic condition. EoC GLOSSOMATICS ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION ALGORITHM: A set of instructions that organises a particular procedure (a formalised task) into steps that can be mechanically performed. Algorithms are used to control and stabilise actions in unstable environments. ANALOG COMPUTING: Early computers that worked on the level of electric control with vacuum tubes, without an intermediary mediation by code (electronics) and the arbitrage that goes along with electronics. We oppose analog computing to quantum computing (see below). CIPHER: The spelling-out of nothing (form Arabic sifr for “empty, nothing, naught”) into a set, or an alphabet of “elements” such that any of the formulations in the code-terms of a cipher can cancel each other out. (See Rottmann, 1987; Barad, 2012; Bühlmann, 2018b). ICT ALGORITHM: The encoding of computerised procedures into a set of steps; used to find data in the control of process or the provision of relevant information. INDEXES diverse “manners” (“rationalisations” [pl., hence “manners”]) of how computerised procedures can be discretised and dissolved into combinatorial clouds of possible sequentiality. They are “probabilisers”; tools used to prepare data such that it can be identified according to different relevancy measures, and sorted and treated variously. They are applied widely, from generic data compression algorithms such as .jpg and gif, to specific contexts for the mining of big data, as in image recognition software or, more generally, pattern detecting applications. CODE: Systems of symbols used for information control and programming. It is significant that code is a “language”, or “an economy” of symbols, not a symbolical system. Because of this, a structural approach (symbols, algebra, groups) and a systematic approach (forms and elements, axiomatic) are not mutually exclusive with regard to code. This exclusiveness between “system” and “structure” is a problem that lies at the very core in all approaches that try to formulate a “general linguistics”, at least since Saussure. Taking into account the “materiality” of code can deal with this problem because it does not try to “solve” it, but rather provides a comparative method for how it can be “resolved” – in the same sense that we speak of a mixed substance being “resolved” in chemistry. Technically, a particular code always has a double make-up: it is always relative to a logarithmic base (decimal, binary, hexagonal, a particular alphabet etc.) and it is governed by a particular set of rules (grammars, syntaxes). The logarithmic base steps in the place where in axiomatics where in axiomatics is the place of common notions (elements); it indexes their place logistically, according to a vicarious order of substitute positioning. DATA: refers to measurements of quantified information. All data is relative to how it has been recorded. In popular language, data often refers to the information used by particular markets; it often tends to be naturalized uncritically, by conflating it with code as the ‘nature’ of the algorithm. 48 EoC GLOSSOMATICS DATA SOVEREIGNTY: Digital data subject to the laws of the country in which it is stored in server farms. As a legal paradigm, it comes into crisis with cloud computing and its geographically distributed manner of storage. DEDA: Data ethics decision aid, https://dataschool.nl/deda/?lang=en DIGITAL CITIZENSHIP: Citizenship that relates to belonging to communities of users that have attained a degree of digital competence (see https://ec.europa.eu/jrc/en/ digcomp/digital-competence-framework), but with the added steps of striving for a societal ethics based on a great number of different codes of conduct. ELECTRICS/ELECTRONICS: When the field of electronics was invented in the 1880ies, electrical devices had already been around for at least 100 years (e.g. Volta’s electric batteries, or Samuel Morse’s electric telegraph). The difference lies in how devices manipulate electricity to do their work. Electrical devices take the energy of electric current and transform it in simple ways into some other form of energy – most likely light, heat, or motion. Electronic devices do much more. Instead of just converting electrical energy into heat, light, or motion, electronic devices are designed to manipulate the electrical current itself to coax it into doing interesting and useful things. The very first electronic device invented in 1883 by Thomas Edison manipulated the electric current passing through a light bulb in a way that let Edison create a device that could monitor the voltage being provided to an electrical circuit and automatically increase or decrease the voltage if it became too low or too high. ETHICS: We understand ethics as “boundary work” that manifests in contracts that organise protocols for actions. Ethics comes from the Greek term ethos, meaning “habitual character and disposition” in plural, “manners”, and refers to a person’s, community’s, or institution’s activities and conduct. When habits are culturally sanctioned into customs, they become constitutive for morals. Problematizing such processes of sanctioning as processes of institutionalization, there is a comparative and integrative vector of ethics (that seeks to abstract from particular morals and make them compatible) that is capable to address the level of society (not communities). INFORMATICS: from information and the suffix -ics, for all that pertains to it. The science of knowing how to “read” and “write” amidst the circuit of information, as manners of processing data for storage and retrieval. 49 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION IGNORANCE, AND ITS COST: To plead ignorance in actions is not an acceptable argument (Arendt, 1963, 1978, 2003). To be ignorant is not innocent. “I did not mean to do that” does not solve the problem that something is done – cf. Les rencontres philosophique de lÚNESCO: Qu est-ce qu on ne sait pas? (Serres, Ricoeur, Blanchot, etc.) MACHINE: The universal Turing machine (UTM). MATHEMATICAL MODEL: A schematisation tool that informs the set-up of algorithms. A model always reduces the complexity of a situation in a definite and particular way. Different models of the same situation do not contradict each other from a quantum literacy perspective, but provide for different “resolutions” in the material-code sense elaborated above (> CODE). As such, models lend themselves as a tool to come up with manners of schematisation and formatting. They can be characterised in their different capacities and capabilities, and they can be compared and profiled with regard to each other. While different models may all be valid, they are not equally well suited for being applied to a particular problem. (Models play as central a role in quantum literacy, as arguments do in linguistic literacy). MODAL: A measure of a proposition, law, predication, or knowledge model. A modal is a qualifier for the quantitative measurement of contingency and necessity stated by systems of propositions, laws, predications, or a particular knowledge model. MORAL: “Moral” comes from the Latin term moralis, referring to the proper behaviour of a person in a particular community or culture. They embody contingent customs and traditions. NORM: A norm is derived from an established system, or from a projective, ideal system. It claims validity in a generally committing, “necessary” sense. 50 EoC GLOSSOMATICS QUANTUM: A unit (quantity) of mass (energy and information) relational to its interactions with other units of mass (energy and information). A quantum unit (quantity) measures “undecided” energy as “pre-specific mass” (materiality neither reducible to classical forces, nor to any one homogenous generic force in particular, like heat). It is a unit that measures probability amplitudes (waves of possibilities that propagate across resolutions), and hence a unit of measurement that is not absolute but conditioned, yet variously so (in this variation it differs from the Kantian notion of transcendentality). Quantum measurement provides for a semiotic-material practice of measurement that is performative in a social sense (interpretative, enactive, continuous) and that can be theorised individually (hence biased), yet in an objective and anonymous sense (formal, distinct, “polite”). In this latter respect, the performativity of quantum measurement is unlike other forms of social performativity such as rituals – forms that can be theorised only on a level of collective and culturally specific, shared customs, rather than individually, anonymously, or formally. QUANTUM COMPUTING: We maintain that digital computing is quantum computing, in that it makes the shift from electrics to electronics. The digital code is code on a binary logarithmic base, this alone is the main difference to a decimal coding system for example (which takes ten as its logarithmic base) or any other coding system. The binary base is more abstract then the others in that it can express the terms of any other coding system in its base. This abstraction is similar to that between different number domains: while all natural numbers can be expressed as rational numbers, and while all rational and irrational can be expressed as real numbers, the reverse does not apply. Electronics that works with binary code involves not only real analysis (as analog computers do) but complex analysis. This makes such computers quantum computers. The same reasoning explains also for the multiplication of digital communication channels that all “surf” one and the same analog communication channel (a certain frequency). The frequency serves as the physical “substrate” or “carrier” for an indefinitely great number of multiple digital channels. 51 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION QUANTUM LITERACY: Literacy for expressing heterogeneous quanta conditions. Literacy in expressing, articulating, estimating, cultivating and hence understanding quanta conditions in their semiotic-material heterogeneity. Quantum literacy is literacy in formality and their symbolic conditions, rather than in one particular form or format and its set of characters (as when “literate” is classically referred to mastering an alphabet, or a theory of numbers, or a theory of forms). STANDARD: A standard can constitute a system (as in IT standards), on arbitrary grounds. In this, it is unlike a norm which is derived from an established system or convention. THE PRICE OF INFORMATION: To turn data into information means to strip this data of much of the possibilities/contingencies it embodies (as data). Hence, information comes at a cost – it destroys virtual conditions of possibility for the coexistence of differences (Brillouin, 2004: 303). TURING MACHINE: A Turing Machine is a mathematical model of computation which manipulates symbols mechanically on a strip of tape according to a table of rules. (Turing, 1936, 1938). UNIVERSAL TURING MACHINE: The universal Turing machine (UTM) can simulate an arbitrary Turing machine on arbitrary input. The challenge is in how the mechanical demonstration mode at work in a Turing Machine can be accommodated in one homogenous (gravitated) universal dynamics. This is not directly a question of physics, it touches upon the role of abstract mathematics with regard to empirical physics, and hence touches upon the problems of how to think about the Laws of Nature. The challenge that quantum physics poses to classical (Newtonian) physics is that it relativizes the latter’s central and integrative role of classical dynamics over pre-classical mechanics; quantum physics introduces, before all else, a novel mechanics that is itself analytical/algebraic. The philosophical challenges this novel paradigm for physics poses is how to relate quantum mechanics to the algebraic conception of natural laws as Laws of Conservation (Noether, 1918; Kosmann-Schwarzbach, 2011). 52 REFERENCES 53 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION Abel, G. J. (2010). Estimation of international migration flow tables in Europe. Journal of the Royal Statistical Society: Series A (Statistics in Society), 173(4), 797–825. Amoore, L. (2009). Algorithmic war: Everyday geographies of the war on terror. Antipode, 41(1), 49–69. Amoore, L., & Piotukh, V. (Eds.). (2015). Algorithmic life: Calculative devices in the age of big data. London: Routledge. Amoore, L., & Raley, R. (2017). Securing with algorithms: Knowledge, decision, sovereignty. Security Dialogue, 48(1), 3–10. Arendt, H. (1958). The human condition. Chicago: University of Chicago Press. Arendt, H. (1961). Between past and future: Eight exercises in political thought. London: Penguin Books. Arendt, H. (1963). Eichmann in Jerusalem. London: Penguin Books. Arendt, H. (1978). The life of the mind. New York: Harcourt Brace Jovanovich. Arendt, H. (1989). Lectures on Kant’s political philosophy. R. Beiner (Ed.). Chicago: University of Chicago Press. Arendt, H. (1994). Essays in understanding 1930–1954: Formation, exile and totalitarianism. Jerome Kohn (Ed.). New York: Schocken Books. Arendt, H. (2003). Responsibility and judgment. Jerome Kohn (Ed.). New York: Schocken Books. Ayache, E. (2010). The blank swan: The end of probability. Hoboken: Wiley. Ayache, E. (2017). Price Writing, “Towards a Quantum Literacy: Spectral Sovereignty, Citizenship, and Personhood in a Digital World,” Ethics of coding: A report on the algorithmic condition [Seminar], Technische Universität Wien. Balibar, E. (2015). Borderland Europe and the challenge of migration. Open Democracy. Retrieved from https://www.opendemocracy.net/can-europe-make-it/etienne-balibar/borderlandeurope-and-challenge-of-migration [Accessed March 12, 2016] Barad, K. (2007). Meeting the universe halfway: Quantum physics and the entanglement of matter and meaning. Durham, NC: Duke University Press. Barad, K. (2012). What is the Measure of Nothingness: Infinity, Virtuality, Justice. Berlin: Hatje Cantz. Bar–Ilan, J., Keenoy, K., Levene, M., & Yaari, E. (2009). Presentation bias is significant in determining user preference for search results – A user study. Journal of the Association for Information Science and Technology, 60(1), 135–149. Bell, D. (1973). The coming of post-industrial society: A venture in social forecasting. New York: Basic Harper. Benveniste, É. (1969). Le vocabulaire des institutions indo-européennes. Collection ‘Le Sens Commun’. Paris: Les Éditions de Minuit. Bhabha, H. (2017). Spectral sovereignty, vernacular cosmopolitans, and cosmopolitan memories, Robbins, B. and Horta, P. L. (Eds.), Cosmopolitanisms (pp. 141–152). New York: New York University Press. 54 REFERENCES Binns, R., Veale, M., Van Kleek, M., & Shadbolt, N. (2017, September). Like trainer, like bot? Inheritance of bias in algorithmic content moderation. In: Ciampaglia, G.L. and Mashhadi, A. and Yasseri, T., (Eds.) Lecture Notes in Computer Science, Vol. 10540. (405–415). Springer: Cham, Switzerland. Bishop, S. (2018). Anxiety, panic and self-optimization: Inequalities and the YouTube algorithm. Convergence, 24(1), 69–84. Birmingham, P. (2006). Hannah Arendt and human rights: The predicament of common responsibility. Bloomington: Indiana University Press. Braidotti, R. (2006). Affirmation versus vulnerability: On contemporary ethical debates. Symposium: Canadian Journal of Continental Philosophy, 10(1), 235–254. Brillouin, L. (2004). Science and information theory (2nd ed.). New York: Dover Phoenix. (Original work published 1962) Bühlmann, V. (2018a). The digital, a continent? Within the object-space of cunning reason. Tsagdis, G. and Lindberg, S. (Eds.), Intersections: At the Technophysics of Space [Special issue]. Azimuth: Philosophical Coordinates in Modern and Contemporary Age. Bühlmann, V. (2018b). Generic Mediality: On the Role of Ciphers and Vicarious Symbols in an Extended Sense of Code-based ‘Alphabeticity’, Braidotti, R. and Dolphijn, R. (Eds.), Philosophy After Nature. (31–54), London: Rowman & Littlefield. Bühlmann, V., & Hovestadt, L. (2013). Introduction: Printed physics. In: Bühlmann, V. and Hovestadt, L. (Eds.), Printed Physics (71–112). Metalithikum 1. Vienna: Ambra. Bühlmann, V., Hovestadt, L., Alvarez-Marin, D., Roman, M., & Michael, S. (2015). Quantum City. Vienna: Birkhäuser Vienna. Bühlmann, V,. Colman, F., Van der Tuin, I. (2017). Introduction to New Materialist Genealogies: New Materialisms, Novel Mentalities, Quantum Literacy, Minnesota Review, (88), 47–58. Burrell, J. (2016). How the Machine “thinks”: Understanding Opacity in Machine Learning Algorithms, Big Data & Society 3, no. 1 (January- June): 1–12. Butler, J. (2004). Precarious life: The powers of mourning and violence. New York: Verso. Chuang, I. L., Vandersypen, L. M., Zhou, X., Leung, D. W., & Lloyd, S. (1998). Experimental realization of a quantum algorithm. Nature, 393 (6681), 143. Cheney-Lippold, J. (2011). A new algorithmic identity: Soft biopolitics and the modulation of control. Theory, Culture & Society, 28(6), 164–181. Chun, W. (2011). Programmed visions: Software and memory. Cambridge, MA: MIT Press. Colman, F. (2015). Digital biopolitics: The image of life. Wilmer, S. E. and Zukauskaite, A. (Eds.), Resisting biopolitics: Philosophical, political, and performative strategies (189–201), London: Routledge. Colman, F. J. (2017). The Earth as material film: Benning’s light glance making the materialimage, Lubecker, N. and Rugo, D. (Eds.), James Benning’s Environments: Politics, Ecology, Duration, (114–126), Edinburgh & Oxford University Presses. Clastres, P. (2010). Archaeology of violence. (Herman, J. , Trans.). Cambridge, MA: MIT Press. (Original work published 1980) 55 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION Crossley, J. N., & Henry, A. S. (1990). Thus spake al-Khwarizmi: A translation of the text of Cambridge University Library Ms. Ii. vi. 5. Historia Mathematica, 17(2), 103–131. Churchman, C. West (1967). Wicked Problems. Management Science. 14 (4). De Laat, P. B. (2010). How can contributors to open-source communities be trusted? On the assumption, inference, and substitution of trust. Ethics and information technology, 12(4), 327–341. De Laat, P. B. (2017). Big Data and Algorithmic Decision-Making: Can Transparency Restore Accountability?, ACM SIGCAS Computers and Society (47:3): 39–53. Deutsch, D. (1985, July). Quantum theory, the Church–Turing principle and the universal quantum computer. Proceedings of the Royal Society London (Vol. 400, No. 1818, 97–117). Dolphijn, R., & Van Der Tuin, I. (2012). New materialism: Interviews and cartographies. Ann Arbor, MI: Open Humanities Press. Retrieved from http://quod.lib.umich.edu/o/ ohp/11515701.0001.001 Dobnikar, A., Lotric, U., & Ster, B. (2011). Proceedings (Part I) from ICANNGA 2011: Adaptive and natural computing algorithms 10th international conference. Ljubljana, Slovenia, April 14–16, 2011. Berlin: Springer. Edwards, L., & Veale, M. (2017). Slave to the algorithm? Why a “right to an explanation” is probably not the remedy you are looking for. Duke Law & Technology Review. 16(1), pp. 18–84. Available at SSRN: https://ssrn.com/abstract=2972855 “Ethics” [European Commission definition] Horizon2020: https://ec.europa.eu/programmes/ horizon2020/en/h2020-section/ethics Feynman, R. (2014 [1986]). QED The Strange Theory of Light and Matter. Princeton: Princeton University Press. Finn, E. (2017). What algorithms want: Imagination in the age of computing. Cambridge, MA: MIT Press. Floridi, L. (2013). The ethics of information. Oxford: Oxford University Press. Floridi, L. (Ed.). (2015). The onlife manifesto: Being human in a hyperconnected era. Cham, Switzerland: Springer. Franzke, A. and Schäfer, M. (2017) Data ethics decision aid [DEDA]. https://dataschool.nl/ deda/?lang=en Galison, P. (1994). The ontology of the enemy: Norbert Wiener and the cybernetic vision. Critical inquiry, 21(1), 228-266. Galison, P. (1996). Computer Simulations and the Trading Zone, Galison, P. and Stump, D. J. (Eds.), The Disunity of Science: Boundaries, Contexts, and Power. Stanford: Stanford University Press. pp. 118–157. Gasché, R. (2008). Europe, or the infinite task: A study of a philosophical concept. Stanford: Stanford University Press. Gieryn, T. F. (1983). Boundary-work and the demarcation of science from non-science: Strains and interests in professional ideologies of scientists. American sociological review, 781–795. 56 REFERENCES Greenwald, A. G. (2017). An AI stereotype catcher: An artificial intelligence method identifies implicit human biases such as gender stereotypes. Science, 356(6334), 133–134. doi: 10.1126/science.aan0649 Goriunova, O. (Ed.). (2014). Fun and software: Exploring pleasure, paradox and pain in computing. London: Bloomsbury. Guattari, F. (1995). Chaosmosis: an ethico-aesthetic paradigm (Bains, P. and Pefanis, J., Trans.) Bloomington: Indiana University Press. Halpern, O. (2014). Beautiful data: A history of vision and reason since 1945. Durham, NC: Duke University Press. Hansen, M. B. (2012). Bodies in code: Interfaces with digital media. London: Routledge. Haraway, D. J. (1988, Autumn). Situated knowledges: The science question in feminism and the privilege of partial perspective. Feminist Studies, 14(3), 575–599. Haraway, D. J. (1991). Simians, cyborgs, and women: The reinvention of nature. London: Free Association Books. (Original work published 1984) Haraway D. J. (1997) Modest_Witness@Second_Millennium. FemaleMan©_Meets_ Oncomouse. New York: Routledge Hayles, N. K. (2012). How we think: Digital media and contemporary technogenesis. Chicago: University of Chicago Press. Hayles, N. K. (2017). Unthought: The power of the cognitive nonconscious. University of Chicago Press. Hodge, J. (1995). Heidegger and Ethics. London: Routledge. Hodge, J. (2002). Ethics and Time: Levinas Between Kant and Husserl. Diacritics., 32(3), 107–134. Hodge, J. (2017). Technology, technics, ecotechnics: Heidegger, Stiegler & Nancy. Ethics of coding: A report on the algorithmic condition [Seminar]. ICA, London. Hoskyns, C. (1996). Integrating gender: Women, law and politics in the European Union. London: Verso. Hovestadt, L., Bühlmann, V., Alvarez-Marin, D., Roman, M., & Michael, S. (2015). A Quantum City. Mastering the Generic. Vienna: Birkhäuser Vienna. Hovestadt, L., Bühlmann, V., & Michael, S. (2017). A genius planet: A radical pathway from energy scarcity to energy abundance. Vienna: Birkhäuser. Hui, Y. (2016). On the existence of digital objects. Minneapolis: University of Minnesota Press. Hui, Y. (2017). The question concerning technology in China. Falmouth: Urbanomic. Hutcheson, G. D. (2005). Moore’s law: the history and economics of an observation that changed the world. Electrochemical Society Interface, 14(1), 17–21. Kennedy, H. (2016). Post, Mine, Repeat: Social Media Data Mining Becomes Ordinary. Basingstoke: Palgrave Macmillan. Kirby, V. (2011). Quantum anthropologies: Life at large. Durham, NC: Duke University Press. Kitchin, R., & Dodge, M. (2011). Code/space: Softwear and everyday life. Cambridge, MA: MIT Press. Kline, R. R. (2015). The cybernetics moment: Or why we call our age the information age. Baltimore, MD: Johns Hopkins University Press. 57 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION Kool, L., Timmer, J., & Van Est, R. (2015). De datagedreven samenleving: Achtergrondstudie. Den Haag: Rathenau Instituut. Kosmann-Schwarzbach, Y. (2011) The Noether Theorems. Invariance and Conservation Laws in the Twentieth Century. New York: Springer. Kroll , J. A., Huey, J., Barocas, S., Felten, E. W., Reidenberg, J.R., Robinson, D. G., Yu, H. (2017). Accountable Algorithms, University of Pennsylvania Law Review 165, 633–705. Lambert, G. (2001). Report to the academy: Re the New conflict of the faculties. Aurora, CO: The Davies Group. Retrieved from https://www.academia.edu/213546/Report_to_the_ Academy_re_the_New_Conflict_of_Faculties_ Langer, S. K. (2009). Philosophy in a new key: A study in the symbolism of reason, rite, and art. Cambridge, MA: Harvard University Press. (Original work published 1942) Lévy-Leblond, J.-M. (1976). Towards a proper quantum theory (hints for a recasting). Retrieved from http://ucispace.lib.uci.edu/bitstream/handle/10575/1146/levy-leblond%20--%20 towards%20a%20proper%20quantum%20theory%20(1976).pdf?sequence=1 Lévy-Leblond, J.-M. (1999). Quantum words for a quantum world. In D. Greenberger, Reiter, W. L. and Zeilinger A. (Eds.), Epistemological and experimental perspectives on quantum physics (75–87). New York: Springer. Lovink, G. (2002). Dark fiber: Tracking critical internet culture. Cambridge, MA: MIT Press. Lunenfeld, P. (2011). The secret war between downloading and uploading: tales of the computer as culture machine. Cambridge, MA: MIT Press. Lyotard, J.-F. (1984). The postmodern condition: A report on knowledge. (Bennington, G. and Massumi, B., Trans.) Manchester: Manchester University Press. (Original work published 1979). Lyotard, J.-F. (1979). Les problèmes du savoir dans les sociétés industrielles les plus développées. Quebec: Conseil des Universités. Original report retrieved from http://www.cse.gouv. qc.ca/FR/Publications_CUniv/ Masschelein, J., & Simons, M. (2009). From active citizenship to world citizenship: A proposal for a world university. European Educational Research Journal, 8(2), 236–248. Manovich, L. (2013). The algorithms of our lives. The Chronicle of Higher Education. Retrieved from http://chronicle.com/article/The-Algorithms-of-Our-Lives-/143557/ [Accessed March 15, 2016] McLuhan, M. (1964). Understanding Media: Extensions of man. New York: McGraw Hill. Moore, P. V. (2018). The Quantified Self in Precarity: Work, technology and what counts. London & New York: Routledge. Niggli, A. (2014). Mechanical justice. In V. Bühlmann & L. Hovestadt (Eds.), Domesticating symbols. Metalithicum II, (pp. 20–45). Vienna: Ambra. Noble, S. U. (2018). Algorithms of oppression: How search engines reinforce racism. New York: New York University Press. Noether, E. (1918). Invariante Variationsprobleme, Nachr. D. König. Gesellsch. d. Wiss. zu Göttingen, Math.-phys. Klasse. 235–257. 58 REFERENCES Nora, S., & Minc, A. (1980). The computerization of society: A report to the president of France. Cambridge, MA: MIT Press. (Original work published 1978) O’Donnell, A. (2012). Transformative education. In: Kieran, P. and Grenham, T. (Eds.), New educational horizons in contemporary Ireland (277–338), Oxford: Peter Lang. O’Donnell, A. (2014). Experimental philosophy and pedagogy. In: Hickey-Moody, A. and Page, T. (Eds.), Material cultures of resistance (21–40), Kentucky: Rowman & Littlefield. Parisi, L. (2013). Contagious architecture: Computation, aesthetics and space. Cambridge, MA: MIT Press. Pariser, E. (2011). The filter bubble: What the Internet is hiding from you. Penguin UK. Plotnitsky, A. (1994). Complementarity: anti-epistemology after Bohr and Derrida. Durham, NC: Duke University Press. Plotnitsky, A. (2006). Reading Bohr: Physics and philosophy. Dortrecht: Springer. Plotnitsky, A. (2009). Epistemology and probability: Bohr, Heisenberg, Schrödinger, and the nature of quantum-theoretical thinking. New York: Springer. Poser, H. (2013). Technology and Modality. In Bühlmann,V. and Hovestadt, L. (Eds.), Printed Physics. – Metalithikum I, (pp. 71–112), Vienna: Ambra. Powell, A. (2017). What to Do about Biased AI? Going beyond Transparency of Automated Systems, (6 September), [blog] http://www.alisonpowell.ca/?p=793. Powell, A., & Nemorin, S. (2017). A virtues-based approach to making sense of ethics in IoT. Presentation at Ethics of coding: A report on the algorithmic condition [Seminar]. ICA, London. Rasmussen, M. L., & Harwood, V. (2009). Young people, education and unlawful non–citizenship: Spectral sovereignty and governmentality in Australia. Globalisation, societies and education, 7(1), 5–22. Rid, T. (2016). Rise of the machines: A cybernetic history. New York: WW Norton & Company. Rindzeviciute, E. (2016). The power of systems: How policy sciences opened up the Cold War World. Ithaca, NY: Cornel University Press. Rittel, Horst W. J.; Webber, M. (1973). Dilemmas in a General Theory of Planning. Policy Sciences. 4: 155–169. Ritter, J. (1989a), “Babylon–1800” in Serres, M. (Ed.) A History of Scientific Thought. Paris: Blackwell, 17–43. Ritter, J. (1989b), “Measure for Measure: Mathematics in Egypt and Mesopotamia”. In: Serres, M. (Ed.) A History of Scientific Thought. Paris: Blackwell, 44–72. Robbins, B., & Horta, P. L. (2017). Introduction. In: Robbins, B. and Horta, P. L. (Eds.), Cosmopolitanisms (1–20). New York: New York University Press. Rottmann, B. (1987), Signifying Nothing: The Semiotics of Zero. Standford: Stanford University Press. Rouvroy, A. (2012). The end(s) of critique: data-behaviourism vs. due-process, Privacy, Due Process and the Computational Turn. Ed. Mireille Hildebrandt, Ekatarina De Vries, Routledge: 143–167. 59 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION Rouvroy, A. & Berns, T. (2013). Gouvernementalité algorithmique et perspectives d’émancipation. Réseaux, (1), 163–196. Rouvroy, A. & Stiegler, B. (2016). The digital regime of truth: From the algorithmic governmentality to a new rule of law. La Deleuziana, 3, 1–24. Schäfer, M. T. (2017) Investigating Datafication by Doing Entrepreneurial Research EoC seminar: The Social Ethics of Coding, at Utrecht University, 30 March. Schaefer, M.T . & Van Es, K. F. (2017). The Datafied Society – Studying Culture through Data. Amsterdam: Amsterdam University Press. Scharff, R. C. (2012). Empirical Technoscience Studies in a Comtean World: Too Much Concreteness?. Philosophy & Technology, 25(2), 153–177. Shor, P. W. (1994). Algorithms for quantum computation: Discrete logarithms and factoring. In: Foundations of Computer Science, 1994 Proceedings., 35th Annual Symposium (124–134). Schuppli, S. (2014). Deadly algorithms: Can legal codes hold software accountable for code that kills? Radical Philosophy, (187), 2–8. Seeing Data (2017). http://seeingdata.org/ Simon, J. (2015). Distributed epistemic responsibility in a hyperconnected era. In: Floridi, L. (Ed.) The Onlife Manifesto. (145–159). Cham, Switzerland: Springer. Siskin, C. (2016). System: The shaping of modern knowledge. Cambridge, MA: MIT Press. Smith, N. (1992). Contours of a spatialized politics: Homeless vehicles and the production of geographical scale. Social Text, 33, 54–81. Smith, B. H. (2006). Scandalous knowledge: Science, truth and the human. Edinburgh: Edinburgh University Press. Stiegler, B. (1998, 2008, 2010). Technics and Time (Vols. 1–3). Stanford: Stanford University Press. Stiegler, B. (2012). Relational ecology and the digital pharmakon. Culture Machine, 13, 1–19. Stocker, G., & Schöpf, Ch. (Eds.). (2003). Code: The language of our time. Exh. Cat. Linz, Austria. Strömbom, D., Mann, R. P., Wilson, A. M., Hailes, S., Morton, A. J., Sumpter, D. J., & King, A. J. (2014). Solving the shepherding problem: heuristics for herding autonomous, interacting agents. Journal of the royal society interface, 11(100). Terranova, T. (2004). Network culture: Politics for the information age. London: Pluto Press. Thacker, E. (2015). Biophilosophy for the 21st century, Wilmer, S. E. and Zukauskaite, A. (Eds.), Resisting biopolitics: Philosophical, political, and performative strategies (123–155). London: Routledge. Tselentis, G., & Galis, A. (Eds.) (2010). Towards the future Internet: Emerging trends from European research. Amsterdam: IOS press. Turing, A. M. (1936). “On Computable Numbers, with an Application to the Entscheidungsproblem”. Proceedings of the London Mathematical Society. 2 (published 1937). 42: 230–265. 60 REFERENCES Turing, A.M. (1938). «On Computable Numbers, with an Application to the Entscheidungsproblem: A correction». Proceedings of the London Mathematical Society. 2 (published 1937). 43 (6): 544–6. Vaile, D., Kalinich, K. P., Fair, P. V., & Lawrence, A. (2013). Data sovereignty and the cloud: A board and executive officer’s guide. Cyberspace Law and Policy Centre, UNSW Faculty of Law, Version 1.0, 2 July http://www.bakercyberlawcentre.org/data_sovereignty/index.html Van Der Tuin, I. (2015). Generational feminism: New materialist introduction to a generative approach. New York: Lexington Books. Van Dijck, J. (2013). The culture of connectivity: A critical history of social media. Oxford University Press. Van Es, K. F. & Schaefer, M. T. (2017). Introduction: Brave New World. In: Van Es, K. and Schäfer, M. T. (Eds.), The Datafied Society: Studying Culture through Data (pp. 13–22). Amsterdam University Press. Veale, M., & Edwards, L. (2017, November 15). Clarity, surprises, and further questions in the Article 29 Working Party Draft Guidance on Automated Decision-Making and Profiling. SSRN. Available from https://ssrn.com/abstract=3071679 Vlieghe, J. (2013). Education in an age of digital technologies. Philosophy & Technology, 27, 519-537. doi:10.1007/s13347-013-0131-x Whitaker, A. (2012). The new quantum age: from Bell’s theorem to quantum computation and teleportation. Oxford: Oxford University Press. Wiener, N. (1961). Cybernetics or control and communication in the animal and the machine (2nd ed.). Vol. 25. Cambridge, MA: MIT Press. World Bank. (2016). World Development Report 2016: Digital Dividends. Washington, DC: World Bank. Weyl, H. (1981). Emmy Noether. In: Emmy Noether 1882–1935 (112–152). Birkhäuser Boston. 61 ETHICS OF CODING: A REPORT ON THE ALGORITHMIC CONDITION APPENDIX A DEDA Source: https://www.uu.nl/en/news/utrecht-data-school-develops-data-ethics-decision-aid Source: 62 https://www.uu.nl/en/news/utrecht-data-school-develops-data-ethics- EoC Ethics of Coding: A Report on the Algorithmic Condition