Nadav Cohen — Selected Talks
AI Week 2024 (Tel Aviv-Yafo, Israel)
Invited talk
Implicit Biases of Gradient Descent in Offline System Identification and Optimal Control
[slides]
FoCM 2023 (Paris, France)
Invited talk
What Makes Data Suitable for Deep Learning?
[slides]
ICML 2022 (Baltimore, MD, USA)
Invited talk
Continuous vs. Discrete Optimization of Deep Neural Networks
[video|slides]
Workshop on Tensor Methods and their Applications in the Physical and Data Sciences 2021 (virtual)
Invited talk
Implicit Regularization in Deep Learning: Lessons Learned from Matrix and Tensor Factorization
[video|slides]
IMVC 2020 (virtual)
Invited keynote talk
Practical Implications of Theoretical Deep Learning
[video|slides]
NeurIPS 2019 (Vancouver, Canada)
Contributed talk
Implicit Regularization in Deep Matrix Factorization
[video|slides]
AI Week 2019 (Tel Aviv-Yafo, Israel)
Invited talk
Analyzing Optimization and Generalization in Deep Learning via Trajectories of Gradient Descent
[video|slides]
CECAM Workshop on Quantum Computing and Quantum Chemistry 2019 (Tel Aviv-Yafo, Israel)
Invited talk
Expressiveness in Deep Learning via Quantum Entanglement
[slides]
Simons Institute Workshop on Frontiers of Deep Learning 2019 (Berkeley, CA, USA)
Invited talk
Analyzing Optimization and Generalization in Deep Learning via Trajectories of Gradient Descent
[video|slides]
ICERM Workshop on Theory and Practice in Machine Learning and Computer Vision 2019 (Providence, RI, USA)
Invited talk
Analyzing Optimization in Deep Learning via Trajectories
[video|slides]
ICML 2018 (Stockholm, Sweden)
Contributed talk
On the Optimization of Deep Networks: Implicit Acceleration by Overparameterization
[video|slides]
ICLR 2018 (Vancouver, Canada)
Contributed talk
Boosting Dilated Convolutional Networks with Mixed Tensor Decompositions
[video|slides]
Symposium on the Mathematical Theory of Deep Neural Networks 2018 (Princeton, NJ, USA)
Invited talk
On the Optimization of Deep Networks: Implicit Acceleration by Overparameterization
[video|slides]
Symposium on Physics and Machine Learning 2017 (New York City, NY, USA)
Invited talk
Understanding Deep Learning via Physics: The Use of Quantum Entanglement for Studying the Inductive Bias of Convolutional Networks
[slides]
Mathematics of Deep Learning Workshop 2017 (Berlin, Germany)
Invited talk
Expressiveness of Convolutional Networks via Hierarchical Tensor Decompositions
[slides]
AAAI Spring Symposium Series 2017 (Palo Alto, CA, USA)
Invited talk
Expressive Efficiency and Inductive Bias of Convolutional Networks: Analysis & Design via Hierarchical Tensor Decompositions
[slides]
CVPR 2017 (Honolulu, HI, USA)
Invited talk
Expressive Efficiency and Inductive Bias of Convolutional Networks: Analysis & Design via Hierarchical Tensor Decompositions
[slides]
GAMM 2017 (Weimar, Germany)
Invited talk
On the Expressive Power of Deep Learning: A Tensor Analysis
[slides]
NeurIPS 2016 (Barcelona, Spain)
Invited talk
Inductive Bias of Deep Convolutional Networks through Pooling Geometry
[slides]
ICML 2016 (New York City, NY, USA)
Contributed talk
Convolutional Rectifier Networks as Generalized Tensor Decompositions
[video|slides]
COLT 2016 (New York City, NY, USA)
Contributed talk
On the Expressive Power of Deep Learning: A Tensor Analysis
[video|slides]
|