Distributed Computing
ETH Zurich

Seminar in Deep Neural Networks (FS 2026)

Organization

When & Where: Tuesdays 10:15 - 12:00, ETZ E 9.
First seminar: 17.02.2026
Last seminar: 26.05.2026
Easter holidays: No seminar on 07.04.2026
Coordinators: Frédéric Berdoz, Roger Wattenhofer.

Background

This is a seminar, we will focus on recent research and skip most of the basics. We assume that all participants are familiar with the fundamentals on deep neural networks. If you feel like you cannot follow the discussions, please check out this playlist, this lecture, the book by Francois Chollet on Deep Learning with Python, or any other lectures or books on deep neural networks. As a seminar participant, you are asked to attend all the talks and make a presentation.

Seminar Timeline

Preparation Timeline

Your Presentation

Grade

The most important part of your grade will be the quality of your presentation, both content and style. In addition, we grade how well you direct the discussions with the audience, during and after the presentation. Beside your final presentation, we also grade how actively you participate in the discussions throughout the semester and we value the quality of your mentor-only test presentation. Attendance is mandatory.

Schedule

Date Presenter Title Mentor Slides
February 17 Frédéric Berdoz Introduction to Scientific Presentations - [pdf]
February 24 Jonathan Maillefaud Generative Graph Pattern Machine Saku Peltonen -
- TBA - -
March 03 Tom Haidinger Neural Message-Passing on Attention Graphs for Hallucination Detection Florian Grötschla -
Yucheng Wang MiMo-Audio: Audio Language Models are Few-Shot Learners Luca Lanzendörfer -
March 10 Frederico Aberle Does Reinforcement Learning Really Incentivize Reasoning Capacity in LLMs Beyond the Base Model? Frédéric Berdoz -
Daven Lim Do Language Models Use Their Depth Efficiently? Joël Mathys -
March 17 Mara Godeanu SAM Audio: Segment Anything in Audio Florian Grötschla -
Klejdi Sevdari VOICECRAFT: Zero-Shot Speech Editing and Text-to-Speech in the Wild Luca Lanzendörfer -
March 24 Rui Wang WHALE: Towards Generalizable and Scalable World Models for Embodied Decision-making David Jenny -
Maximilian Schlegel Extending the Context of Pretrained LLMs by Dropping Their Positional Embeddings Frédéric Berdoz -
March 31 Ioana Gabor The quest for the GRAph Level autoEncoder (GRALE) Joël Mathys -
Peter Jolles Graph Positional Encoding via Random Feature Propagation Saku Peltonen -
April 07 - Easter break - -
April 14 Prathamesh Tagore Sign and Basis Invariant Networks for Spectral Graph Representation Learning Saku Peltonen -
Sichao Ma Continuous Audio Language Models Florian Grötschla -
April 21 Benedict Armstrong Energy-Based Transformers are Scalable Learners and Thinkers Joël Mathys -
Felix Quernheim Generalizable Insights for Graph Transformers in Theory and Practice Joël Mathys -
April 28 - No Seminar - -
May 05 Felix Schatzl DISCOVER: Automated Curricula for Sparse-Reward Reinforcement Learning David Jenny -
- TBA - -
May 12 Federico Lin Differentiable Euler Characteristic Transforms for Shape Classification Saku Peltonen -
Georgios Sideris Collaborative Learning in the Jungle Antonio Di Maio -
May 19 Baraq Ben-Or Lipshitz LeJEPA: Provable and Scalable Self-Supervised Learning Without the Heuristics Andreas Plesner -
Samvit Mavinkurve You Cannot Feed Two Birds with One Score: the Accuracy-Naturalness Tradeoff in Translation Till Aczel -
May 26 Lin Che Overtrained Language Models Are Harder to Fine-Tune Andreas Plesner -
Louis Barinka What makes an image realistic? Till Aczel -