I am a machine learning researcher building statistical methods to enhance our grasp of biological systems in neuroscience and the life sciences more broadly. My main interest is to understand how machine learning systems can be used to infer computations and dynamics underlying a phenomenon using observational and interventional data.

In my research group at Helmholtz Munich, we develop machine learning algorithms for representation learning and inference of nonlinear system dynamics, study how large and multi-modal biological datasets can be compressed into foundation models, and study their mechanistic interpretability. If you are interested joining us, have a look at our current openings.

I am also very active in computing education for more than 10 years now. In 2019, I founded KI macht Schule (“AI at schools”), a non-profit organization teaching ML basics to high-school students. We provide teachers with modern teaching materials, AI tools and infrastructure in our open teaching hub, and offer courses on AI for students and teachers in Germany, Switzerland and Austria. We have a great network of volunteers in more than 9 cities who do science outreach directly in schools. We also have a growing team of instructional designers, software engineers and AI trainers to extend our platform, build new teaching materials and conduct teacher trainings. If you want to help educating the next generation of students and make them literate in AI, consider to join our team!

Finally, I am interested in deploying ML solutions in the real world. In 2023, I co-founded Kinematik AI, a company offering customized machine intelligence solutions in the biopharma and animal healthcare sector. If you have a business usecase for our customized behavioral analysis software, please reach out.

Background

I pursued my doctoral studies at the Max Planck International Research School for Intelligent School and the Swiss Federal Institute of Technology Lausanne (EPFL), advised by Matthias Bethge and Mackenzie Mathis in the ELLIS PhD & PostDoc program. My PhD studies were supported by a Google PhD Fellowship in Computational Neural and Cognitive Sciences.

During my PhD, I also worked as a Research Scientist Intern advised by Laurens van der Maaten and Ishan Misra on multimodal representation learning in the FAIR team at Meta NYC, and at Amazon Web Services in Tübingen as an Applied Science Intern where I worked on self-learning and object centric representations with Peter Gehler, Bernhard Schölkopf and Matthias Bethge.

Prior to starting my PhD, I worked on wav2vec and vq-wav2vec, two self-supervised representation learning algorithms for speech processing with Michael Auli, Alexei Baevski and Ronan Collobert at Facebook AI Research in Menlo Park, CA.

Short Bio (for talks etc.)

Steffen Schneider is a tenure-track group leader at Helmholtz Munich and associated faculty at the Graduate School of Systemic Neurosciences at LMU Munich and the Munich Center for Machine Learning. His research focuses on statistical methods for modeling and understanding biological dynamical systems. He earned his PhD in the ELLIS PhD program at EPFL and the Tübingen AI Center, advised by Mackenzie Mathis and Matthias Bethge. Previously, he was an AI Resident at Meta with Michael Auli, working on self-supervised contrastive learning for speech recognition. Beyond research, in 2019 he founded the non-profit KI macht Schule (“AI at Schools”) to advance AI education and statistical literacy. His PhD was supported by a Google PhD Fellowship, and he was recently named Early Career Scientist of the Year 2024 by academics for his contributions to research and education.

Interests

  • Dynamical Systems
  • Self-Supervised Learning
  • Sensorimotor Adaptation
  • AI for Science
  • AI Education and Statistical Literacy
  • Computational Neuroscience

Education & Research

  • Group Leader (Tenure Track)

    Helmholtz Munich

    from 2024

  • PhD, Neuroscience

    Swiss Federal Institute of Technology Lausanne (EPFL)

    2021 - 2024

  • PhD Researcher (via ELLIS)

    Intl. Max Planck Research School, Tübingen

    2019 - 2023

  • Research Scientist Intern

    FAIR at Meta, New York City

    Spring 2022

  • Applied Science Intern

    Amazon Web Services, Tübingen

    Fall 2020

  • AI Resident, Self-Supervised Learning for Speech Recognition

    Facebook AI Research, Menlo Park, CA

    2018 - 2019

  • MSc in Neuroengineering

    Technical University of Munich

    2016 - 2018

  • BSc in Electrical Engineering, Information Technology and Computer Engineering

    RWTH Aachen University

    2013 - 2016

Latest

ConceptScope: Characterizing Dataset Bias via Disentangled Visual Concepts

ConceptScope is a scalable and automated framework for analyzing visual datasets by discovering and quantifying human-interpretable …

Equivariance by Contrast: Identifiable Equivariant Embeddings from Unlabeled Finite Group Actions

EbC learns equivariant embeddings from observation pairs without relying on group-specific inductive biases, with theoretical …

CytoSAE: Interpretable Cell Embeddings for Hematology

CytoSAE is a sparse autoencoder for interpretable cell embeddings in hematology, enabling discovery of morphologically relevant …

xCEBRA: Time-series Attribution Maps with Regularized Contrastive Learning

xCEBRA generates attribution maps with identifiability guarantees for time-series data using regularized contrastive learning and a new …

Sparse Autoencoders Reveal Selective Remapping of Visual Concepts During Adaptation

PatchSAE is a framework to extract interpretable concepts and their spatial attributions in vision transformers.

Neuro-musculoskeletal Modeling Reveals Muscle-level Neural Dynamics of Adaptive Learning in Sensorimotor Cortex

Combining neural recordings with musculoskeletal models to reveal muscle-level neural dynamics during adaptive learning in sensorimotor …

SuperAnimal Pretrained Pose Estimation Models for Behavioral Analysis

SuperAnimal develops unified foundation models for animal pose estimation across 45+ species without additional human labels, with …

RDumb: A Simple Approach that Questions Our Progress in Continual Test-Time Adaptation

RDumb is a simple baseline that periodically resets the model to its pretrained state, yet outperforms state-of-the-art continual …

Learnable Latent Embeddings for Joint Behavioral and Neural Analysis

CEBRA is a contrastive learning algorithm for relating behavior and neural activity.

Multi-animal Pose Estimation, Identification and Tracking with DeepLabCut

A framework for multi-animal pose estimation, identification and tracking built around DeepLabCut, enabling behavioral analysis of …

Unsupervised Object Learning via Common Fate

Unsupervised object learning from videos using the Common Fate Principle, decomposing the problem into motion segmentation, generative …

Contrastive Learning Inverts the Data Generating Process

Contrastive learning with the InfoNCE objective can recover the ground truth latent factors underlying the data.

If Your Data Distribution Shifts, Use Self-Learning

Self-learning techniques like entropy minimization and pseudo-labeling are simple and effective at improving model performance under …

Motor Control: Neural Correlates of Optimal Feedback Control Theory

Recent work is revealing neural correlates of a leading theory of motor control. By linking an elegant series of behavioral experiments …

Pretraining Boosts Out-of-Domain Robustness for Pose Estimation

ImageNet pretraining significantly improves out-of-domain robustness for pose estimation across architectures and species.

A Primer on Motion Capture with Deep Learning: Principles, Pitfalls and Perspectives

We discuss principles of pose estimation algorithms and highlight their potential and pitfalls for experimentalists.

Improving Robustness against Common Corruptions by Covariate Shift Adaptation

Adapting batch norm statistics of trained computer vision models considerably increases robustness at test-time.

vq-wav2vec: Self-Supervised Learning of Discrete Speech Representations

Learning discrete representations of speech yields state-of-the-art recognition performance on TIMIT and WSJ.

Iron-Sequestering Nanocompartments as Multiplexed Electron Microscopy Gene Reporters

We introduce the encapsulin/cargo system of Quasibacillus thermotolerans for multiplexed gene reporter imaging via transmission …

wav2vec: Unsupervised Pre-training for Speech Recognition

Contrastive pre-training on speech at scale reduces the need for labeled data.

Multi-Task Generalization and Adaptation between Noisy Digit Datasets: An Empirical Study

We show that good target performance can be achieved on domain adaptation tasks by adapting only the normalization statistics and …

Salad: A Toolbox for Semi-supervised Adaptive Learning Across Domains

An open source toolbox providing a unified implementation of state-of-the-art methods for transfer learning, semi-supervised learning …

NeuBtracker — imaging neurobehavioral dynamics in freely behaving fish

NeuBtracker is a tracking microscope for simultaneous imaging of neuronal activity and behavior of freely swimming fluorescent reporter …

Context-based Normalization of Histological Stains using Deep Convolutional Features

A fully automated, end-to-end learning-based setup for normalizing histological stains using Feature Aware Normalization, which extends …