International Workshop: “AI & The Future of Protest Politics”
The Boston-Bergen Forum on Digital Futures invites to a two-day hybrid event, bringing together prominent philosophers to discuss the impact of AI-based platforms and their underlying technological and economic principles on political discourse and protest.
Main content
We are happy to invite to our international workshop “AI and the Future of Protest Politics: Politics and Emotions in the Age of Digital Transformation and Surveillance Capitalism”, a two-day, hybrid event taking place at the Philosophy Department in Bergen, on 30th November (Thursday) and 1st December (Friday) 2023, from 9:00—17:00 (CET).
Our goal is to bring together prominent scholars from different disciplines to discuss the impact of AI-based platforms, and their underlying technological and economic principles, on political discourse and protest.
The event will draw on the Boston-Bergen Forum on Digital Futures—an international research network among the ‘Culture, Society & Politics’ and the ‘Practical Philosophy’ research groups at UiB’s Philosophy Department, the MediaFutures Research Centre Bergen, the Applied Ethics Center at UMass Boston, and the MIT Program Human Rights and Technology.
This project is Co-lead by Professor Franz Knappik (Bergen) and Dr. Christopher Senf (Bergen) as well Professor Christoph Trattner (MediaFutures) and Professor Nir Eisikovist (UMass).
Program:
1st Workshop Day, Thursday, November 30th
09.15 am.: Welcome Note
Section I: Making AI Intelligible
09.30 am.: Keynote by Herman Cappelen (Hong Kong, online):
“AI and The Commodification of Meaning”
10.15 am.: Q&A (Moderation, Jesse Tomalty)
11.00 am.: Rosalie Waelen (Bonn):
“The Struggle for Recognition and AI’s Impact on Self-development”
11.30 am.: Q&A (Moderation, Chris Senf)
***
12.30 am.: Lunch break at Café Christie
***
Section II: Contesting the Attention Economy
14.00 pm.: Sebastian Watzl (Oslo):
“What is Wrong with How Attention is Commodified?”
14.30 pm.: James Williams:
“Silly Questions”
15.00 pm.: Q&A (Moderation, Franz Knappik)
Section III: Algorithms of (In)justice
16.00 pm.: Kjetil Rommetveit (Bergen):
“(How) Can you code rights and morality into digital infrastructures and AIs?”
16:30 pm.: Q&A (Moderation, Carlota Salvador Megias)
17.15 pm.: End
2nd Workshop Day, Friday, December 1st
10.00 am.: Welcome Note
Section I: Political Technologies
10.15 am.: Eugenia Stamboliev (Vienna):
“Protesting the Classification of Emotions (affects) and its Technological means”
10.45 am.: James Hughes (UMass, Boston):
“Communication Technologies: Hegemonic, Radicalizing and Democratic”
11.15 am.: Q&A (Moderation, Franz Knappik)
***
12.15 am.: Lunch break at Café Christie
***
Section II: Future Struggles
13:30 pm.: Paul Raekstad (Amsterdam):
“Domination Without Dominators: The Impersonal Causes of Oppression”
14:00 pm.: Alec Stubbs (UMass, Boston, online):
“Becoming Achievement Subjects in the Attention Economy”
14.30 pm.: Q&A (Moderation, Chris Senf)
***
15.15 pm.: Coffee break
***
Section III Activism and Philosophy in the Age of AI
15.30 pm.: Maria Brincker (UMass Boston, online):
“What kind of space is a ‘platform’ with its own goals?”
16.00 pm.: Kade Crockford (ACLU Massachusetts & MIT Media Lab, online):
“All Politics is Local: Fighting Face Surveillance from the Ground Up in Massachusetts”
16.30 pm.: Q&A (Moderation, Chris Senf)
17:30 pm.: End
Abstracts
1) Herman Wright Cappelen (Hong Kong)
“AI and the Commodification of Meaning”
AI systems, owned by private corporations, will soon have the ability to control the meaning of the sentences we speak and interpret. This can be seen as a form of commodification of speech act content, a more serious form of commodification than e.g., artistic commodification. The determination of meaning by AI raises concerns about corporate control over language, reminiscent of Orwellian scenarios. Often, the goals behind these communicative exchanges will be foreign to individuals, who may not endorse or even be aware of them. The result is a form of meaning alienation.
2) Rosalie Waelen (Bonn)
“The struggle for recognition and AI’s impact on self-development”
Critical theories, with their focus on power dynamics and emancipation, offer a valuable basis for the analysis of AI’s social and political impact. Axel Honneth’s theory of recognition is one such critical theory. Honneth’s theory of recognition adds to the present AI ethics debate, because it shines light on the different ways in which AI reinforces or exacerbates struggles for recognition. Moreover, through the lens of Honneth’s theory of recognition, one learns how AI can harm people’s self-development. This presentation highlights some of those contemporary struggles for recognition and their (potential) impact on people´s self-development.
3) Sebastian Watzl (UiO Oslo)
“What is Wrong with How Attention is Commodified?”
Our attention is commodified: it is bought and sold in market transactions when individuals lend out the ability to control their attentional capacities in exchange (for example) for technological services. What is wrong with that? Attention markets, we argue, resemble labor markets. By drawing on the ethics of commodification and core features of attention, we show that attention markets, while not always morally wrong, carry special moral risks: because of how attention shapes beliefs and desires, subjective experience and action, they are prone to be disrespectful, alienating, and provide fertile grounds for domination. Our analysis calls for regulatory interventions.
4) James Williams
“Silly Questions”
What kinds of questions about the attention economy are worth asking at present? The absence of any meaningful structural reform, along with the continued march of automation, make this analytical focus all the more important. Here I will discuss a series of questions which are crucial yet have been largely overlooked. Some of them are also quite silly. This is a feature, not a bug; as Whitehead noted, 'the silly question is the first intimation of some totally new development.’
5) Kjetil Rommetveit (UiB Bergen)
“(How) Can you code rights and morality into digital infrastructures and AIs?”
In 1980 philosopher of technology Langdon Winner famously asked 'Do Artifacts Have Politics?' This question was followed up by Latour's (1994) and Verbeek's (2008) analyses of technological mediation of morality. Whereas these questions were once provocative, in recent AI regulations they have become part of official governance mechanisms. In this talk I present some novel approaches to governance in the EU through, specifically the risk-based approach and the design-based approach. Situating these within a wider techno-regulatory imaginary, I provide examples of how these instruments play out in practice. I end on some critical questions: what kind of politics do in-built morality have? And what implications can be discerned for critical publics?
6) Eugenia Stamboliev (Vienna)
“Protesting the classification of emotions (affects) and its technological means”
To critique affect technology, we need to politicize emotionality and affectivity newly. Today, we are witnessing the emergence of intrusive algorithmic technologies, such as AI, in our daily lives. These technologies, designed to measure and control lives, information, and data, are intended to nudge and influence our political moods and public sentiments as much as they are to measure expressions and emotions. In this talk, I will discuss the history of two types of "affect technologies" (AT) and offer some criticism on their goals and applications. First, ATs intended to measure and classify emotions and affection emerged from the cognitive turn in computer studies. While popular, these ATs are normatively problematic and flawed, but still influence the design and economic models underlying many recognition systems. Second, ATs expected to drive, manage, and influence political beliefs and public moods are underlining architectures that do more than manage emotions via technological means, but they are part of the devaluation of emotions through political campaigning. Protesting the shortcomings of ATs, means calling into question both the normative and political agendas underlying affect technologies, as well as offering new and positive approaches on affectivity that are beyond the scope of measurement and control, but remain politically crucial for democratic protest while avoiding commercial and technical exploitation.
7) James J. Hughes (UMass Boston & IEET)
“Communication Technologies: Hegemonic, Radicalizing and Democratic”
Books, radio and television all transformed political mobilization, by both elites and radicals. How different is the Internet, social media and algorithmically driven communication? Are we more likely to form radical sub-communities, each with its own reality (e.g. MAGA)? Can we envision democratic countervailing institutions emerging from the "commodification, outrageification, and gamification of protest" by platform companies? Will the algorithmic rules and required moderation included in the EU AI Act, DMA and DSA reduce ideological hegemony, improve collaboration and decrease toxicity in these environments?
8) Paul Raekstad (Amsterdam)
“Domination Without Dominators: The Impersonal Causes of Oppression”
Social movements of the last centuries have been naming and analyzing the complex forms of personal and impersonal domination that they fight to overcome. Yet current theories of domination have largely been unable to make sense of the latter. Theories of domination as being subject to the will, or arbitrary power, of another rule them out, while extant theories of impersonal domination are often unsystematic or narrowly focused. My paper tries to remedy this by developing a systematic theory of impersonal domination, distinguish some important types thereof, and show why it matters for universal human emancipation.
9) Alec Stubbs (UMass Boston)
“Becoming Achievement Subjects in the Attention Economy”
This talk argues that digital platforms are increasingly dependent on the involvement of users, who engage in 'prosumption' – the active production and consumption of digital content. This talk will explore how this engagement leads to a form of self-exploitation drawing from the work of Byung Chul-Han. Additionally, this talk builds from the Italian Autonomists' concept of 'the social factory,' suggesting that our online interactions blur the lines between the social world and traditional forms of production. Finally, I will end by considering the rise of generative AI and its potential impact on content creation, shaping the future landscape of digital interaction.
10) Maria Brincker (UMass Boston)
“What kind of space is a ‘platform’ with its own goals?”
How are we to understand our political actions on surveillance and algorithm-driven for-profit platforms? Current social media platforms present users with possibilities of building vast networks and achieving massive, fast reach to highly dispersed groups. Hence, they present incredible opportunities for expanded agency, organizing, and information sharing. However, these platform ecosystems also present users with highly unusual affordance spaces, which might pose challenges to our agency. Proprietary algorithms, vast data harvesting and camouflaged behavior modification tools are used to drive platform company interests – often conflicting with those of users. We engage in political movements to shape the future, but how do our actions on these platforms in fact shape our future and our extra-situational spaces?
11) Kade Crockford (ACLU & MIT)
“All Politics is Local: Fighting Face Surveillance from the Ground Up in Massachusetts”
In 2019, the ACLU of Massachusetts launched a campaign to bring democratic control over government use of facial surveillance technology. Over the following two years, we passed eight bans on government use of face surveillance in cities and towns across the state, including in Massachusetts’ four largest cities: Boston, Cambridge, Springfield, and Worcester. We also passed a state law creating some regulations on police use of the technology statewide. During this talk, campaign leader Kade Crockford will discuss how the ACLU’s campaigners dreamed big, built a coalition, and fought from the ground up to defeat the narrative of technological determinism, and how you can do it, too.
Join our workshop, whether in person or remotely!