Digitization and the rise of artificial intelligence forecast radical change on all aspects of human practice, especially given the ever-improving abilities of algorithms in tasks like pattern recognition and their practical application. Powerful technology arises from AI research, opening the gate for various forms of cultural and societal engineering, i.e., a reshaping of culture and society by dint of algorithmic models and "intelligent" applications. To date, however, even highly-trained algorithms are only outperforming humans in very specific tasks with limited scope (e.g., chess), as opposed to banal – yet cognitively highly complex – everyday actions like assessing the immediate consequences of a lie.
Thus, although the development of artificial intelligence is still in its beginnings, it has already triggered an enormous rush of utopian and dystopian thinking. While some dream of immortality and the vanquishing of poverty, disease, and warfare, others foresee a grim future for those parts of humanity that will find themselves outpaced by technology. Potential consequences of the changes imposed by technological advancement on human practice reach from the level of the individual, through cultural techniques, to the organization of society as a whole, raising fundamental questions, such as:
- How does artificial intelligence impact our understanding of the human mind, especially in relation to the role of its computational equivalents that reach more and more aspects of everyday life (e.g., chatbots, driverless mobility, risk assessment software in the banking and insurance sector)?
- What are the consequences of digitization and machine learning algorithms for education and our understanding of learning and creativity (e.g., in schooling through adaptive tutors, but also against the background of our current notion of creativity as a unique human ability)?
- How will the increasing use of computational methodology change the ways we relate to the past and envision the future (e.g., by reading), both in academia and in society? How can the enrichment of algorithmic models with methods and results from the humanities shape and improve computational assessment of human practice (e.g., data mining of big text corpora, automated translation, racial bias in neural networks)?
- How does the use of artificial intelligence in all domains of human practice influence how we deal with complexity (e.g., of society) and human control thereof? Can computational methods help to reduce, organize, and analyze cultural complexity, or do they pose a threat to human control over different aspects of the lifeworld (e.g., security and network technology, automation of industrial production, autonomous weaponry)?
Against the background of such questions, the conference aims to foster an open and critical reflection on the consequences of cultural and societal engineering. Namely, algorithmic models influence society and culture far beyond the limited scope of their practical application. They reshape communicative processes as well as the ways we interact with the world; and thus, subject culture and society to new forms of ‘engineering’. By including scholars and scientists from the humanities and social sciences (in the broad sense of the word) as well as from AI research, this conference focuses on the implementation of advanced technology in different domains of everyday life, i.e., in relation to concrete areas of application (e.g., pattern recognition) as well as associated human-machine interfaces (e.g., devices and applications). In doing so, the conference investigates not only opportunities and shortcomings of AI research, but also implications and potential structural effects of technological innovation for the organization of societal practice (e.g., work) and techniques of cultural self-reflection (e.g., history). It will not only ask what technologies can do (or will be able to do in the future), but also how these capabilities can be compared and related to their human equivalents, e.g., perception, cognition, and communication.
Accordingly, the conference will establish a reflection on the methodological differences between the different areas of scholarship involved. Hosted by a faculty that unites humanities and social sciences, the conference aims at exploring how empirical sciences based in statistical evidence and scholarship grounded in interpretation can learn from one another, how they can profit from computational methodology, and how they are potentially altered by information technology. A central perspective in this context relates to the question of how humanities and social sciences develop and implement models – and to what extent they reflect upon the preconditions and implications of their ways of modeling their objects of interest. It is assumed that while the humanities can learn a lot from the rigorous procedures of building, testing, and evaluating models in (data-driven) scientific research, the sciences can surely benefit from the critical approach the humanities take when dealing with concepts and models for the (interpretation-driven) analysis and assessment of human practice, e.g., in assessing unintended consequences of models, or their role in society as a whole.
The conference will focus on the following areas of interest, each of which will assemble representatives of different disciplines:
Section I – Mind and Consciousness
- Consciousness & computing: How do models of the mind and of consciousness relate to models of data processing as used in AI development?
- Experience & effectiveness: To what extent are AI algorithms capable of self-experience, or does their effectiveness rely precisely on them being not capable of experience?
- Action & assessment: How does our notion of social activity and of assessing others change with the emergence of AI as a new type of intelligent, but unconscious social actor?
Section II – Learning and Inventing
- Neurons & networks: How can models of “self-learning” neural networks be compared to models of learning developed by psychology, the cognitive sciences, and pedagogy?
- Classification & creativity: How do our notions of genuine human competencies, such as complex pattern recognition, strategic thinking, and exploration, change with the emergence of AI?
- Exploration & education: How does AI change the ways in which we try to foster innovation and reshape educational systems?
Section III – Reading and Data Modeling
- Production & processing: How does AI change the ways in which humans produce and process language, texts, and meaning?
- Meaning & machines: To what extent can human procedures of establishing meaning and sense be established in ‘intelligent’ algorithms?
- Cognition & corpora: How do the emergence of AI and new ways of corpus analysis change the ways in which we interpret the past, envision the future, and process material created by humans in general?
Section IV – Complexity and Control
- Complexity & control: How can AI foster the ways in which we control societal and environmental complexity – and how might it at the same time bring about a loss of human control?
- Analysis & adaptation: How does AI enhance our capability of analysing complex systems and thereby make processes of decision-making more adaptive?
- Transparency & transformation: How does AI affect the transparency of societal and political processes, and how does it transform our engagement with complex ecologies?
Confirmed Keynote Speakers
- Armin Grunwald (Karlsruher Institut für Technologie)
- Jürgen Fohrmann (Rheinische Friedrich Wilhelms-Universität Bonn)
- Lyse Langlois (Université Laval, Quebec)
- Giuseppe Longo (Ècole Normale Supérieure, Paris)
- Joanna Zylinska (Goldsmiths, University of London)
Detailed Descriptions of the Thematic Sections
Section I – Mind and Consciousness: The comparison between models of consciousness and the mind as developed in psychology, cognitive sciences, and philosophy, on the one hand, and the models used by computer scientists to build AI, on the other hand, promises a clearer perspective on a topic that is currently subject to many mystifications. From a scholarly point of view, one of the big issues related to the development of AI is the difference between computing and consciousness. For example, while human actors are self-aware and capable of switching between a first- and a third-person-perspective, those notions hardly apply to current AI applications, i.e., an algorithm has no experiences of itself and of "what it is like" to be AI. Do we have to understand this as a limitation or rather as a necessary precondition for the effectiveness of an artificially-trained system? How does the lack of consciousness shape our ways of interaction with AI applications and our expectations in relation to such interactions (e.g., ethically, aesthetically, economically)? More concretely, this section will investigate how the emergence of AI as a new type of intelligent, but unconscious social actor changes both the conceptualization of society and the role of human individuality, i.e., how is the relevance of personal experience re-evaluated in a technologically-driven societal practice, and how is human individuality subjected to algorithmic assessment?
Section II – Learning and Inventing: The development of AI, particularly of "self-learning" neural networks, depends on models of learning that this section will juxtapose to models of learning as developed, on the one hand, by neurology and cognitive sciences, and, on the other hand, by pedagogy. This section will contrast the ways in which these respective disciplines model the acquisition of competences in pattern recognition, strategic thinking, and human interaction. But it will also explore how they describe creativity, i.e., the emergence of new, unpredictable insights, experiences, and competences. Thereby, it will also discuss how our notion of inventing and exploring might change due to technological progress, taking into account that not only "inspiration", but also more concrete competencies such as the ability to decide what to learn and explore have often been considered uniquely human. As a result of these conversations, this section will try to understand the consequences of AI for the organization of education and evaluate the claims made by proponents of the application of AI to learning processes.
Section III – Reading and data modeling: This section will focus on the operational basis of the humanities, but also of cultural (re)production in general. It will ask how human cognition and communication create texts, language, and cultural artifacts; how, in this process, significance and meaning are constituted; to what extent these processes rely on irreducibly contingent (i.e., unforeseeable) operations; and to what extent these operations can be modeled by algorithms. More concretely, this section will examine how human cognition and communication can be compared to automated processing of information, e.g., in machine translation tasks, machine-machine-communication, or the processing of historical corpora. Further, how is reading a book different from a computational "reading" of big data, and how can the humanities make use of algorithmic operations to challenge and improve hermeneutic methods? This section will foster a dialogue between scholars from the humanities (digital and non-digital), from cognitive sciences, and computer science. It will try to evaluate how an algorithmic assessment of the human past might affect societal self-description or even society’s relation to time.
Section IV – Complexity and Control: Complexity is a key concept for the self-description of modern societies and therefore a central object of investigation both in the sciences and humanities. The concept itself presents as epistemologically challenging, and its descriptions have roots in interdisciplinary approaches such as systems theory, ecology, and cybernetics. Yet, AI adds another layer to such theoretical attempts to reduce and describe complexity and thereby – maybe – control it. Not only is it possible to analyze complex ecological systems (e.g., climate, public opinion, economy, health) with the help of AI. These increasingly complex information technologies have themselves become integral parts of these ecologies. The emergence of networks that interlink humans and AI, e.g., in communication technology or finance, brings about both new utopias of transparency and fears of total surveillance. Artificial control systems challenge established forms of decision making and cherished concepts of (Western) societies which are inherently linked to the question of human control, e.g., democratic governance, privacy, or personal responsibility for actions. Thus, if we rely more and more on new ecologies that we do not fully understand, nor oversee, does the increase of technologically-driven processing (and production) of complexity inevitably lead to a loss of human control? Or can AI be beneficial to overcome pressing ecological and societal problems? This section addresses sociologists and political scientists as well as computer scientists. It will try to assess the popular claim that AI is a solution for problems of control in highly complex ecologies and thereby foster a discussion on how the emergence of AI changes societal discourse.