16–20 Feb 2026
Campus Akademie, Bäckerstraße 13, 1010 Wien
Europe/Zurich timezone

Lecturers and Courses

Monday: "Introduction to ML/AI"

Michal Hradiš (Brno), What is AI? Overview lecture

Claudius Krause (MLA²S, ÖAW), Machine Learning at the ÖAW

 

Tuesday: Vibe Coding & „Promptotyping”

Christopher Pollin (DH Craft Graz)

This workshop introduces current approaches to AI-assisted software development. Starting with Vibe Coding and Context Engineering, participants will learn the Promptotyping methodology, which enables researchers without programming skills to develop their own data-driven tools and interfaces. Through hands-on exercises, participants will learn to construct with (Frontier-)LLMs and iteratively build functioning research artifacts (tools, models, workflows).

Dr. Christopher Pollin is the founder of Digital Humanities Craft OG and an independent researcher specializing in applied generative AI. He received his PhD from the University of Graz on digital modeling of historical information and is developing the Promptotyping methodology for the systematic creation of research tools using LLMs. He teaches Digital Humanities, Generative AI and Prompt Engineering at several Austrian universities.

 

Wednesday: "Ethics and AI"

Bernd Wachmann (ÖAW), "The ÖAW AI Strategy"

Patrick Rarivoson, "Open EuroLLM"

Laura Gandlgruber (University of Vienna), "Beyond Tools: Embedding Responsible AI in Teaching and Learning"

This presentation outlines the University of Vienna's AI policy for teaching and learning, demonstrating how the use of artificial intelligence can be institutionally governed and didactically integrated. It covers key areas such as the systematic development of AI competencies for students and teaching staff, the creation of qualification and information resources, and the integration of AI into teaching, learning, and assessment formats. Using concrete measures and implementation examples, the presentation shows how universities can provide clear orientation, safeguard academic integrity, and embed AI as a sustainable component of academic education.

Mag. Dr Laura Gandlgruber, BA, MA, is Programme Manager for Artificial Intelligence in Studies and Teaching at the University of Vienna, as well as being a certified AI manager. She lectures in artificial intelligence and scientific work at the University of Vienna, the FH der WKW, and the University of Vienna's Postgraduate Centre. As an enthusiastic advocate of AI, she keeps up to date with the latest developments in artificial intelligence and its application in teaching, studies and university administration. She designs and leads workshops for teachers and administrative staff, and is also a highly sought-after expert on AI-related topics for events and interviews.

Charlotte Spencer-Smith (CMC), Governing Machines: From Moderation to Alignment

Philip Winter (VRVis), Bias in ML

This talk provides an accessible introduction to bias in machine learning. Beginning with a concise ML crash course that covers key concepts like model training and learning paradigms, we investigate various types of social and technical biases that may occur in modern deep learning applications. Real-world consequences are examined, highlighting societal harms alongside technical pitfalls like reduced generalization and inaccurate predictions. The talk concludes with practical mitigation strategies.

Dr. Philip Winter is a researcher in the fields of Machine Learning and Astrophysics. He studied Astrophysics at University of Vienna (BSc & MSc) and University of Tübingen (Dr), as well as Machine Learning at JKU Linz. He contributes to a variety of ML/DL-related scientific projects for continual learning, semantic segmentation, computer vision, generative modeling, medical applications, and certification in cooperation with companies including AGFA Healthcare, TÜV Austria, and Audi. Moreover, he is experienced in simulating astrodynamical processes such as gravitational n-body systems and plastic-elastic collision processes for planet formation. Since 2022, Philip is working as a Machine Learning researcher in the Biomedical Image Informatics group at VRVis.

Thursday: "From idea to final project"

Stavrina Dimosthenous (University of Manchester), Data/Code Management, reproducibility

Felix Schmitt (ÖAW/CLIP), Introduction to CLIP 

 

Friday: "Explainability"

Pia Sommerauer (Vrije Universiteit Amsterdam), "Evaluation and Interpretability of NLP systems" 

(Large) language models achieve impressive results on various NLP tasks. At the same time, it is difficult to tell to what degree language models can reflect a human-like understanding of semantics and how specific information is reflected. In this session, we will take a critical look at the strengths and weaknesses of (L)LLMs and consider how LLM-based systems can be evaluated. In a practical, hands-on session, we will build a challenge dataset using a behavioral testing approach.  

Pia Sommerauer is an assistant professor at VU Amsterdam. Her research focuses on understanding how people use language to form (social) categories and on how language models represent social and semantic categories. She approaches these questions from interdisciplinary perspectives and collaborates with experts in fields such as communication science and philosophy. She teaches courses in programming and Natural Language Processing to diverse and interdisciplinary groups of students.

Gunnar König & Moritz Grosse-Wentrup, "Explainability of Neural Networks in STEM fields"