Machine Habitus. Massimo Airoldi
Читать онлайн книгу.such as Google, Amazon and, beginning in the mid 2000s, YouTube, Facebook and Twitter, lie at the core of this historical transition toward the datafication and algorithmic ordering of economy and society (Mayer-Schoenberger and Cukier 2013; van Dijck 2013; Zuboff 2019).
Digital platforms are at once the main context of application of these autonomous machines and the ultimate source of their intelligence. Platformization has been identified as one of the causes of the current ‘eternal spring’ of AI research, since it has finally provided the enormous amount of data and real-time feedback needed to train machine learning models, such as users’ profile pictures, online transactions or social media posts (Helmond 2015). Together with the development of faster and higher performing computers, this access to ‘big’ and relatively inexpensive data made possible the breakthrough of ‘deep learning’ in the 2010s (Kelleher 2019). As Mühlhoff notes (2020: 1869), most industrial AI implementations ‘come with extensive media infrastructure for capturing humans in distributed, human-machine computing networks, which as a whole perform the intelligence capacity that is commonly attributed to the computer system’. Hence, it is not by chance that the top players in the Internet industry, in the US as in China, have taken the lead of the AI race. In 2016, Joaquin Candela, director of the Facebook Applied Machine Learning Group, declared: ‘we’re trying to build more than 1.5 billion AI agents – one for every person who uses Facebook or any of its products’ (Higginbotham 2016, cited in A. Mackenzie 2019: 1995).
Furthermore, while in the Digital Era algorithms were commercially used mainly for analytical purposes, in the Platform Era they also became ‘operational’ devices (A. Mackenzie 2018). Logistic regressions such as those run in SPSS by statisticians in the 1980s could now be operationally embedded in a platform infrastructure and fed with thousands of data ‘features’ in order to autonomously filter the content presented to single users based on adaptable, high-dimensional models (Rieder 2020). The computational implications of this shift have been described by Adrian Mackenzie as follows:
if conventional statistical regression models typically worked with 10 different variables […] and perhaps sample sizes of thousands, data mining and predictive analytics today typically work with hundreds and in some cases tens of thousands of variables and sample sizes of millions or billions. The difference between classical statistics, which often sought to explain associations between variables, and machine learning, which seeks to explore high-dimensional patterns, arises because vector spaces juxtapose almost any number of features. (Mackenzie 2015: 434)
Advanced AI models built using artificial neural networks are now used in chatbots, self-driving cars and recommendation systems, and have enabled the recent expansion of fields such as pattern recognition, machine translation or image generation. In 2015, an AI system developed by the Google-owned company DeepMind was the first to win against a professional player at the complex game of Go. On the one hand, this landmark was a matter of increased computing power.3 On the other hand, it was possible thanks to the aforementioned qualitative shift from a top-down artificial reasoning based on ‘symbolic deduction’ to a bottom-up ‘statistical induction’ (Pasquinelli 2017). AlphaGo – the machine’s name – learned how to play the ancient board game largely on its own, by ‘attempting to match the moves of expert players from recorded games’ (Chen 2016: 6). Far from mechanically executing tasks, current AI technologies can learn from (datafied) experience, a bit like human babies. And as with human babies, once thrown into the world, these machine learning systems are no less than social agents, who shape society and are shaped by it in turn.
Figure 1 Algorithms: a conceptual map, from Euclid to AlphaGo
Critical algorithm studies
When algorithms started to be applied to the digital engineering of the social world, only a few sociologists took notice (Orton-Johnson and Prior 2013). In the early 2000s, the sociological hype about the (then new) social networking sites, streaming platforms and dating services was largely about the possible emancipatory outcomes of an enlarged digital connectivity, the disrupting research potential of big data, and the narrowing divide between ‘real’ and ‘virtual’ lives (Beer 2009). However, at the periphery of academic sociology, social scientists working in fields like software studies, anthropology, philosophy, cultural studies, geography, Internet studies and media research were beginning to theorize and investigate the emergence of a new ‘algorithmic life’ (Amoore and Piotukh 2016). In the past decade, this research strand has grown substantially, disrupting disciplinary borders and setting the agenda of important academic outlets.4 Known as ‘critical algorithm studies’ (Gillespie and Seaver 2016), it proposes multiple sociologies of algorithms which tackle various aspects of the techno-social data assemblages behind AI technologies.
A major part of this critical literature has scrutinized the production of the input of automated calculations, that is, the data. Critical research on the mining of data through digital forms of surveillance (Brayne 2017; van Dijck 2013) and labour (Casilli 2019; Gandini 2020) has illuminated the extractive and ‘panopticist’ character of platforms, Internet services and connected devices such as wearables and smartphones (see Lupton 2020; Ruckenstein and Granroth 2020; Arvidsson 2004). Cheney-Lippold (2011, 2017) developed the notion of ‘algorithmic identity’ in order to study the biopolitical implications of web analytics firms’ data harnessing, aimed at computationally predicting who digital consumers are. Similar studies have also been conducted in the field of critical marketing (Cluley and Brown 2015; Darmody and Zwick 2020; Zwick and Denegri-Knott 2009). Furthermore, a number of works have questioned the epistemological grounds of ‘big data’ approaches, highlighting how the automated and decontextualized analysis of large datasets may ultimately lead to inaccurate or biased results (boyd and Crawford 2012; O’Neil 2016; Broussard 2018). The proliferation of metrics and the ubiquity of ‘datafication’ – that is, the transformation of social action into online quantified data (Mayer-Schoenberger and Cukier 2013) – have been indicated as key features of today’s capitalism, which is seen as increasingly dependent on the harvesting and engineering of consumers’ lives and culture (Zuboff 2019; van Dijck, Poell and de Waal 2018).
As STS research did decades earlier with missiles and electric bulbs (MacKenzie and Wajcman 1999), critical algorithm studies have also explored how algorithmic models and their data infrastructures are developed, manufactured and narrated, eventually with the aim of making these opaque ‘black boxes’ accountable (Pasquale 2015). The ‘anatomy’ of AI systems is the subject of the original work of Crawford and Joler (2018), at the crossroads of art and research. Taking Amazon Echo – the consumer voice-enabled AI device featuring the popular interface Alexa – as an example, the authors show how even the most banal human–device interaction ‘requires a vast planetary network, fueled by the extraction of non-renewable materials, labor, and data’ (Crawford and Joler 2018: 2). Behind the capacity of Amazon Echo to hear, interpret and efficiently respond to users’ commands, there is not only a machine learning model in a constant process of optimization, but also a wide array of accumulated scientific knowledge, natural resources such as the lithium and cobalt used in batteries, and labour exploited in the mining of both rare metals and data. Several studies have looked more closely into the genesis of specific platforms and algorithmic systems, tracing their historical evolution and practical implementation while simultaneously unveiling the cultural and political assumptions inscribed in their technicalities (Rieder 2017; D. MacKenzie 2018; Helmond, Nieborg and van der Vlist 2019; Neyland 2019; Seaver 2019; Eubanks 2018; Hallinan and Striphas 2016; McKelvey 2018; Gillespie 2018). Furthermore, since algorithms are also cultural and discursive objects (Beer 2017; Seaver 2017; Bucher 2017; Campolo and Crawford 2020), researchers have investigated how they are marketed and – as often happens – mythicized (Natale and Ballatore 2020; Neyland 2019). This literature shows how the fictitious representation of calculative devices as necessarily neutral, objective and accurate in their predictions is ideologically rooted in the techno-chauvinistic belief that ‘tech is always the solution’ (Broussard 2018: 7).