16–20 Feb 2026
Campus Akademie, Bäckerstraße 13, 1010 Wien
Europe/Zurich timezone

Contribution List

21 out of 21 displayed
Export to PDF
  1. Dr Claudius Krause (MBI Vienna (ÖAW)), Claus Trost (Erich Schmid Institute of Materials Science of theAustrian Academy of Sciences), Jan Odstrčilík (Institut für Mittelalterforschung, ÖAW), Kati Heinrich (IGF | ÖAW), Nicki Holighaus (Acoustics Research Institute, Austrian Academy of Sciences)
    16/02/2026, 12:00
  2. Michal Hradiš (Brno University of Technology)
    16/02/2026, 12:30

    1) This lecture introduces what “AI” means today and how it relates to machine learning, with an emphasis on large language models (LLMs) and AI agents. We’ll build intuition for core ML ideas—supervised vs. self-supervised learning, loss functions, optimization, and generalization.

    regression, classification, generation, language models, etc.

    Go to contribution page
  3. Dr Claudius Krause (MBI Vienna (ÖAW))
    16/02/2026, 14:30
  4. Michal Hradiš (Brno University of Technology)
    16/02/2026, 16:00

    2) This lecture explains how today’s language models work at a high level: key building blocks, typical training stages, and the resulting capabilities and limitations. We’ll finish with a brief Python demo using the OpenAI API and a local model via Ollama.

    LLMs and LLM architecture, transformers, supervised and unsupervised training in more detail

    Go to contribution page
  5. Christopher Pollin (DH Craft Graz)
    17/02/2026, 09:00
  6. Christopher Pollin (DH Craft Graz)
    17/02/2026, 11:00
  7. Christopher Pollin (DH Craft Graz)
    17/02/2026, 14:00
  8. Christopher Pollin (DH Craft Graz)
    17/02/2026, 16:00
  9. Philip Winter (VRVis)
    18/02/2026, 09:30
  10. Patrick Rarivoson
    18/02/2026, 11:30
  11. Bernd Wachmann (ÖAW)
    18/02/2026, 14:00
  12. Charlotte Spencer-Smith (CMC)
    18/02/2026, 14:45
  13. Laura Gandlgruber (Vienna University)
    18/02/2026, 16:15
  14. Stavrina Dimosthenous (University of Manchester)
    19/02/2026, 09:00

    Delving into the world of data-driven research presents exciting possibilities. However, there are a few topics that need to be addressed before one can jump right in.
    Research data management (RDM) planning, software version control, automation, and data cleaning might appear dry at first glance, and they might indeed be to many. Despite these topics and practices seeming like ‘checkbox’...

    Go to contribution page
  15. Stavrina Dimosthenous (University of Manchester)
    19/02/2026, 11:30

    Delving into the world of data-driven research presents exciting possibilities. However, there are a few topics that need to be addressed before one can jump right in.
    Research data management (RDM) planning, software version control, automation, and data cleaning might appear dry at first glance, and they might indeed be to many. Despite these topics and practices seeming like ‘checkbox’...

    Go to contribution page
  16. Stavrina Dimosthenous (University of Manchester)
    19/02/2026, 14:00

    Delving into the world of data-driven research presents exciting possibilities. However, there are a few topics that need to be addressed before one can jump right in.
    Research data management (RDM) planning, software version control, automation, and data cleaning might appear dry at first glance, and they might indeed be to many. Despite these topics and practices seeming like ‘checkbox’...

    Go to contribution page
  17. Erich Birngruber (CLIP)
    19/02/2026, 16:00
  18. Gunnar König
    20/02/2026, 09:00

    Due to their predictive accuracy, machine learning models are
    increasingly employed in high-stakes decision making and in scientific
    contexts. These predictive capabilities often come at the price of
    interpretability: The models are too complex to be inherently
    intelligible by their human users.
    In an attempt to restore this interpretability, a broad range of methods
    have been developed,...

    Go to contribution page
  19. Moritz Grosse-Wentrup
    20/02/2026, 11:30

    Machine learning (ML) models are increasingly deployed in high-stakes
    environments, e.g., in the health domain, where ethical and legal
    considerations require models to be interpretable. Despite substantial
    progress in interpretable ML (IML), several key challenges remain. These
    include distinguishing between interpreting the model and using the
    model to interpret the data-generating...

    Go to contribution page
  20. Pia Sommerauer (Vrije Universiteit Amsterdam)
    20/02/2026, 13:30