LIT AI Lab/ELLIS Unit Linz Seminar.

The LIT AI Lab kicks off the “LIT AI Lab/ ELLIS Unit Linz Seminar” in summer 2021, which shall bring together researchers and students interested in the field of Artificial Intelligence.

You can follow the talks online via zoom and/or in person (3G rules apply) as mentioned below.

Wednesday, 1 September 2021 at 10am
Room S3 055 or via Zoom (link further below)

Thomas Koller
Hochschule Luzern Informatik

Challenging Human Supremacy: Evaluating Monte Carlo Tree Search and Deep Learning for the Trick Taking Card Game Jass

Abstract: Despite the recent successful application of Artificial Intelligence (AI) to games, the performance of cooperative agents in imperfect information games is still far from surpassing humans. Cooperating with teammates whose play-styles are not previously known poses additional challenges to current state-of-the-art algorithms. In the Swiss card game Jass, coordination within the two opposing teams is crucial for winning. Since verbal communication is forbidden, the only way to transmit information within the team is through a player’s play-style. This makes the game a particularly suitable candidate subject to continue the research on AI in cooperation games with hidden information. In this work, we analyse the effectiveness and shortcomings of several state-of-the-art algorithms (Monte Carlo Tree Search (MCTS) variants and Deep Neural Networks (DNNs)) at playing the Jass game.

 

Zoom link: https://jku.zoom.us/j/98246734423?pwd=QXJ4ekJBUkNHb3M0bXFKVEVuQ1h1dz09

Thursday, 19 August 2021 at 2:30pm
Zoom (link below)

Alexander Kolesnikov
Google Switzerland

New Vision Architectures Beyond CNNs

Abstract: Convolutional Neural Networks (CNNs) have been solely dominating the field of computer vision for nearly a decade. In this talk I will present two recent papers that propose new and highly competitive architecture classes for computer vision. In the first part I will present the Vision Transformer model (ViT), which is almost identical to the standard transformer model used in natural language processing, but happens to work surprisingly well for vision applications. In the second part of the talk, I will present the MLP-mixer model: an all-MLP architecture for vision. It can be seen as a simplified ViT model without the self-attention layer. Nevertheless, it also demonstrates strong results across a wide range of vision applications.

 

Zoom link: https://jku.zoom.us/j/98246734423?pwd=QXJ4ekJBUkNHb3M0bXFKVEVuQ1h1dz09 (Meeting-ID: 982 4673 4423, Passwort: 550113)

Wednesday, 11 August 2021 at 10am
Lecture hall 19 or via Zoom (link further below)

Richard Küng
Institute for Integrated Circuits, Johannes Kepler University Linz

Provably efficient machine learning for quantum physics

Abstract: Classical machine learning (ML) provides a potentially powerful approach to solving challenging problems in quantum physics and chemistry. However, the advantages of ML over more traditional methods have not been firmly established. We prove that classical ML algorithms can efficiently predict ground state properties of a physical system, after learning from data obtained by measuring related systems. We also prove that classical ML algorithms can efficiently classify a wide range of quantum phases of matter. Our arguments are based on the concept of a classical shadow, a succinct classical description of a quantum state that can be constructed in feasible quantum experiments and be used to predict many properties of the state.
 
This is joint work with Robert Huang (Caltech), Giacomo Torlai (AWS), Victor Albert (University of Maryland) and John Preskill (Caltech+AWS).

 

Zoom link: https://jku.zoom.us/j/98246734423?pwd=QXJ4ekJBUkNHb3M0bXFKVEVuQ1h1dz09 (Meeting-ID: 982 4673 4423, Passwort: 550113)