Prof. Dr. Steffen Augsberg
Justus-Liebig-Universität Giessen
Chair of public law
- Phone number: +4964199-21090
- Email: steffen.augsberg@recht.uni-giessen.de
Ulrich von Ulmenstein
Justus-Liebig-Universität Giessen
Chair of public law
- Phone number: +491744753018
- Email: u_ulmenstein@outlook.de
Christina Lauppert von Peharnik, Ass. lur.
Justus-Liebig-Universität Giessen
Chair of public law
- Phone number: +4964199-21092
- Email: christina.lauppert-von-peharnik@recht.uni-giessen.de
Description: Legal Subproject
Ethical Subproject
Dr. Matthias Braun
Friedrich-Alexander-Universität Erlangen-Nürnberg
Chair of Systematic Theology II (Ethics)
- Phone number: +49 9131 85-26078
- Email: matthias.braun@fau.de
Max Tretter
Friedrich-Alexander-Universität Erlangen-Nürnberg
Chair of Systematic Theology II (Ethics)
- Phone number: +49 9131 85-26889
- Email: max.tretter@fau.de
Description: Ethical Subproject
With regard to solidarity, there is a deep-rooted understanding that solidarity requires at least a degree of uncertainty. The ethical subproject ties in with this idea by questioning existing normative discourses on the relationship between solidarity and its respective entanglement with degrees of certainty.
- The first task focuses on the epistemological question of the extent to which acts of giving in solidarity are challenged by the promises of new AI-generated degrees of certainty.
- A second task is to elaborate in detail to what extent the clinical use of AI questions trust in institutions – as a prerequisite for solidarity.
- The third task is to examine the changing forms of individual and collective controllability in times of clinical use of AI. We already have a more or less sharp (culturally handed down) idea of individual possibilities of control – such as the right not to know, claims to transparency or responsibility and liability – which are necessary prerequisites not only to be able to make free decisions, but also to decide under which conditions giving in solidarity is an expression of individual freedom. An important question will be how these forms of controllability are challenged when they not only have to cope with degrees of uncertainty, but are also confronted with the idea of a growing corpus of (postulated) certainty. While this in itself is a complex issue, things become even more complicated when we think about modes of collective controllability and their embedding in more or less sharp concepts of spatiality and temporality.</small>
Prof. Dr. Nora Szech
Chair of Political Economy
- Phone number: +49721608-43809
- Email: nora.szech@kit.edu
David Ehrlich, M.sc.
Karlsruher Institut of Technology
Institute of Economics
- Phone number: +49721608-43808
- Email: david.ehrlich@kit.edu
Description: Economic Subproject
CwiC investigates the normative and behaviour-scientific challenges in dealing with the new possibilities for prediction in and with AI at the interface of science, society and technology. The ability to predict future developments with unprecedented accuracy affects many – if not all – areas of social life.
In the health sector in particular, the new type of prediction allows for much more precise planning, but on the other hand also calls into question central points of reference: individual self-concepts, our general distinction between illness and health, as well as traditional norms, such as the concept of basic solidarity, which is also fundamental with regard to concepts of social security.
In the CwiC subproject Economics, the handling of AI will therefore be examined more closely from the perspective of behavioral science. How do individuals deal with the new possibilities? Are better forecasting possibilities welcome at all, or rather not? Are concepts of insurance questioned or changed?
The demand for and acceptance of more precise information through the use of AI is probably strongly normative. Therefore, we are also interested in the question how social information – especially normative information – influences the use of AI. To investigate behavioral effects, we conduct economic studies at the KIT KD2 Lab. There, the participants make incentive-based decisions – i.e. decisions that have real consequences, for example of a monetary nature – to ensure a high validity of the observed results.