Timezone: »
We propose a simple interpolation-based method for the efficient approximation of gradients in neural ODE models. We compare it with reverse dynamic method (known in literature as “adjoint method”) to train neural ODEs on classification, density estimation and inference approximation tasks. We also propose a theoretical justification of our approach using logarithmic norm formalism. As a result, our method allows faster model training than the reverse dynamic method what was confirmed and validated by extensive numerical experiments for several standard benchmarks.
Author Information
Talgat Daulbaev (Skolkovo Institute of Science and Technology)
Alexandr Katrutsa (Skolkovo Institute of Science and Technology)
Larisa Markeeva (Skolkovo Institute of Science and Technology)
Julia Gusak (Skolkovo Institute of Science and Technology)
Currently, I am a Research Scientist (AI) at Skolkovo Institute of Science and Technology at Tensor networks and deep learning for applications in data mining laboratory, working with Prof. Ivan Oseledets and Prof. Andrzej Cichocki. My recent research deals with compression and acceleration of computer vision models using tensor methods; training time and performance improvement of neural ordinary differential equations; as well as neural networks analysis using low-rank methods. Also, I have participated in some audio-related projects on speech synthesis and voice conversion. Some of my earlier projects were related to medical data processing (EEG, ECG). My research interests include but not limited to: Deep learning (DL), Computer vision, Speech technologies, Multi-modal/Multi-task learning, Semi-supervised/unsupervised learning, One-/few-/low-shot learning, Incremental learning, Continual learning, Domain adaptation, Hyper Networks, Tensor decompositions for DL, Neural Ordinary Differential Equations, Interpretability of DL.
Andrzej Cichocki (Skolkovo Institute of Science and Technology)
Ivan Oseledets (Skolkovo Institute of Science and Technology)
More from the Same Authors
-
2022 Poster: TTOpt: A Maximum Volume Quantized Tensor Train-based Optimization and its Application to Reinforcement Learning »
Konstantin Sozykin · Andrei Chertkov · Roman Schutski · Anh-Huy Phan · Andrzej S CICHOCKI · Ivan Oseledets -
2022 Poster: Smoothed Embeddings for Certified Few-Shot Learning »
Mikhail Pautov · Olesya Kuznetsova · Nurislam Tursynbek · Aleksandr Petiushko · Ivan Oseledets -
2021 Workshop: Second Workshop on Quantum Tensor Networks in Machine Learning »
Xiao-Yang Liu · Qibin Zhao · Ivan Oseledets · Yufei Ding · Guillaume Rabusseau · Jean Kossaifi · Khadijeh Najafi · Anwar Walid · Andrzej Cichocki · Masashi Sugiyama -
2020 : Panel Discussion 1: Theoretical, Algorithmic and Physical »
Jacob Biamonte · Ivan Oseledets · Jens Eisert · Nadav Cohen · Guillaume Rabusseau · Xiao-Yang Liu -
2020 : Invited Talk 4 Q&A by Ivan »
Ivan Oseledets -
2020 : Invited Talk 4: Quantum in ML and ML in Quantum »
Ivan Oseledets -
2018 Poster: Quadrature-based features for kernel approximation »
Marina Munkhoeva · Yermek Kapushev · Evgeny Burnaev · Ivan Oseledets -
2018 Spotlight: Quadrature-based features for kernel approximation »
Marina Munkhoeva · Yermek Kapushev · Evgeny Burnaev · Ivan Oseledets