Arsenii Ashukha

I am a Senior Research Scientist at Isomorphic Labs, an Alphabet subsidiary that aims to solve all disease, led by Demis Hassabis and Max Jaderberg. I work on deep learning models for the interaction of drug and protein molecules. Prior to joining Isomorphic, I was a Research Scientist at Samsung and a PhD student at the Bayesian Methods Research Group under the supervision of Dmitry Vetrov.

Research

LaMa Inpainting
View Project
WACV, 2022

LaMa uses convolutions in the Fourier space to generalize to a high 2k resolutions, despite being trained on 256x256 images. It achieves a strong performance even in challenging scenarios, e.g. completion of periodic structures.

Variational Dropout
View Project

Variational Dropout Sparsifies Deep Neural Networks

Arsenii Ashukha*, Dmitry Molchanov*, Dmitry Vetrov
ICML, 2017

Retrospective

  • SparsesVD works in practice and it was used for network sparsification in leading IT companies. However, future studies showed that careful usage of pruning-based methods can produce better results.
  • Training of deep models with noise is known to be hard and unstable. That is less the case with SparseVD. All variances are initialized with small values and did not change much during training.
  • The sparse solution is just a local optimum, as better values of ELBO can be achieved with a less flexible variational posterior.

We show that variational dropout trains highly sparsified deep neural networks, while a pattern of sparsity is learned jointly with weights during training.

Pitfalls of Uncertainty
View Project

Pitfalls of In-Domain Uncertainty Estimation and Ensembling in Deep Learning

ICLR, 2020

The work shows that i) a simple ensemble of independently trained networks performs significantly better than recent techniques ii) a simple test-time augmentation applied to a conventional network outperforms low-parameters ensembles (e.g. Dropout) and also improves all ensembles for free iii) a comparison of the uncertainty estimation ability of algorithms is often done incorrectly in the literature.

Open Source Implementations

LaMa Inpainting
View Project

Resolution-robust Large Mask Inpainting with Fourier Convolutions. This implementation has gathered over 10,000 stars on GitHub and is widely used in the community for high-quality image editing.

Gradient Boosting
View Project

Simple Gradient Boosting

A very simple and short implementation of gradient boosting in just 18 lines of code, designed for educational purposes to illustrate the core mechanics of the algorithm.

Real NVP
View Project

Real NVP PyTorch

A minimal and clean implementation of Real NVP normalizing flows in just 40 lines of PyTorch code, focusing on clarity and ease of understanding.

Quantile Regression DQN
View Project

Quantile Regression DQN

A minimal working example of Quantile Regression DQN for Distributional Reinforcement Learning, showcasing how to estimate the full distribution of returns.

Equivariant GNN
View Project

Simple Equivariant GNN

A short, 50-line pure PyTorch implementation of E(n) Equivariant Graph Neural Networks for molecule property prediction, matching paper performance without external dependencies.

SimCLR PyTorch
View Project

SimCLR PyTorch

An unofficial reproduction of SimCLR with support for multi-GPU distributed training and performance matching the original paper on ImageNet and CIFAR-10.

PyTorch Ensembles
View Project

PyTorch Ensembles

A collection of state-of-the-art ensemble methods in PyTorch, including Deep Ensembles, Snapshot Ensembles, cSGLD, and FGE, for improved uncertainty estimation.