Do artificially intelligent systems reflect a certain gender, race or class? What and whose politics is currently being consolidated in algorithmic culture?
As AI finds its way into the mundanity of everyday life, constantly scanning and categorising us to provide the highest level of comfort and efficiency, the systems that surround us are increasingly mediating our bodies, actions and behaviours. Search engine queries, autocomplete functions, auto-image tagging, personal aides, smart-speakers, wearables – all nicely displayed and packaged to ease our regular routines. Currently, however, we can witness a great deal of concealed racial and gender bias in the design of these systems and objects – be it through ‘personal assistants’ with female names and voices or soap dispensers that only works on white hands.
In this block-seminar we will take an applied and interdisciplinary approach to exploring forms of bias in the design of AI. We will locate real-world examples on the topic, engage personally with the systems, and inspired by approaches from queer and feminist theory and technology – we will prototype forms of 'hacking back' (no prior experience with design or technology necessary).
Prof. Dr. des. Michelle Christensen and Prof. Dr. des. Florian Conradi
Technische Universität Berlin Fakultät I Geistes- und Bildungswissenschaften
Institut für Philosophie, Literatur-, Wissenschafts- und Technikgeschichte
23.-27.03.2020 / 10:00 - 18:00
30.-31.03.2020 / 10:00 - 18:00
Einstein Center Digital Future
Robert Koch Forum
Max. 15 participants