Master the building blocks of intelligent systems—from neural networks to LLMs

Explore the essential concepts and architectures that underpin modern machine learning and deep learning. From training neural networks to fine-tuning transformers, this track guides you through practical techniques and state-of-the-art tools to build models that scale, adapt, and perform.

ML & Deep Learning

Learn from Industry Leaders about:

  • Designing and training deep neural networks with modern loss functions and architectures for high-performance models.
  • Supervised learning techniques with best practices for evaluation, generalization, and robust training.
  • Make black-box models interpretable through explainable AI (XAI) methods.
  • Using transfer learning to adapt pre-trained models to new tasks efficiently with minimal data.
  • Explore Transformer architectures and building foundation models.

Frequently Asked Questions

What will I learn in the ML & Deep Learning track?

You’ll learn to design, train, evaluate, and deploy deep learning models using proven architectural patterns. The track balances foundational machine learning principles with cutting-edge deep learning advances like transformers and transfer learning.

How are deep neural networks designed and trained?

You’ll explore how to select architectures, choose loss functions, and tune hyperparameters. Topics include optimization strategies like Adam, batch normalization, and regularization for effective model convergence.

What is supervised learning and why does generalization matter?

Supervised learning uses labeled datasets to train models. You’ll learn how to prevent overfitting, evaluate performance, and ensure your models generalize to unseen data using methods like cross-validation and dropout.

How does explainable AI (XAI) support model trust?

XAI techniques—like SHAP, LIME, and attention visualization—help developers interpret model decisions, improve transparency, and meet compliance requirements.

What role do transformers and LLMs play in deep learning?

You’ll examine the architecture behind transformer-based models, understand how pre-training works, and apply transfer learning to build domain-adapted models with fewer resources.

Track Speakers London 2026

Track Speakers San Diego 2026

San Diego's program will go live soon! Until then, please take a look at

New York's 2024 Program!

TRACK SPEAKERS MUNICH 2026

TRACK SPEAKERS NEW YORK 2026

Track Speakers Amsterdam 2026

Track Speakers Berlin 2025

Track Program London 2026

Track Highlights 2026

Track Program Munich 2026

Track Program New York 2025

Track Program Berlin 2025

Track Program Amsterdam 2026

Track Sessions London 2026

Track Sessions London 2026

View all sessions

Track Sessions Amsterdam 2026

Track Sessions Amsterdam 2026

View all sessions

Track Sessions San Diego 2026

Track Sessions San Diego 2026

View all sessions

Track Sessions MLcon Munich 2026

Track Sessions MLcon Munich 2026

View all sessions

Track Sessions MLCon New York

Track Sessions MLCon New York

View all sessions

Track Sessions MLcon Berlin 2025

Track Sessions MLcon Berlin 2025

View all sessions

Frequently Asked Questions

What will I learn in the ML & Deep Learning track?

You’ll learn to design, train, evaluate, and deploy deep learning models using proven architectural patterns. The track balances foundational machine learning principles with cutting-edge deep learning advances like transformers and transfer learning.

How are deep neural networks designed and trained?

You’ll explore how to select architectures, choose loss functions, and tune hyperparameters. Topics include optimization strategies like Adam, batch normalization, and regularization for effective model convergence

What is supervised learning and why does generalization matter?

Supervised learning uses labeled datasets to train models. You’ll learn how to prevent overfitting, evaluate performance, and ensure your models generalize to unseen data using methods like cross-validation and dropout.

How does explainable AI (XAI) support model trust?

XAI techniques—like SHAP, LIME, and attention visualization—help developers interpret model decisions, improve transparency, and meet compliance requirements.

What role do transformers and LLMs play in deep learning?

You’ll examine the architecture behind transformer-based models, understand how pre-training works, and apply transfer learning to build domain-adapted models with fewer resources.

Enjoying the content?

Get the most out of MLcon by becoming a free community member — curated resources, weekly newsletter, and member-only perks.

Weekly
Articles + tutorials

The reads you'd find if you had time

2× / mo
Live webinars

Experts you can actually ask

Monthly
Magazine + whitepapers

Deep dives worth your weekend

On-demand
Recordings + courses

Past conferences, ready when you are

Behind the Tracks

Generative AI & Agents
Unlock the future of AI with LLM advancements.

MLOps & LLMOps
Bridge the gap between model creation and real-world application.

ML & Deep Learning
From Feature Engineering to Explainable AI.

AI Strategy & Governance
Transform business processes with AI-driven strategies.

AI Developer Tools
Build next-generation machine learning applications.

ML Basics & Principles
Gain a solid foundation in core ML concepts.