Evgenii Nikishin
I am a PhD student at Mila, University of Montreal, where I have the good fortune of being advised by Pierre-Luc Bacon and Aaron Courville. I am broadly interested in Reinforcement Learning and using capabilities of Deep Learning for decision making.
Previously, I spent 1.5 years at Cornell University. I received an MS in Computer Science from Higher School of Economics and Skoltech where I worked with Dmitry Vetrov at the Samsung Lab.
I did my undergrad in CS at Lomonosov Moscow State University supervised by Alexander Dyakonov.
During the 2nd half of 2022, I interned in the RL team at DeepMind London working closely with André Barreto, Junhyuk Oh, and Will Dabney.
I spent summer 2020 remotely interning at Mila with Pierre-Luc Bacon and Yoshua Bengio.
In summer 2018, I interned at ETH Zürich with Gunnar Rätsch applying RL for personal treatment.
Email  / 
CV  / 
Google Scholar  / 
Twitter  / 
GitHub
|
|
|
Maxwell’s Demon at Work: Efficient Pruning by Leveraging Saturation of Neurons.
Simon Dufort-Labbé,
Pierluca D’Oro,
Evgenii Nikishin,
Razvan Pascanu,
Pierre-Luc Bacon,
Aristide Baratin
Arxiv
[PDF]
We show that neuron saturation, which traditionally was viewed as undesirable, could instead lead to sparse yet accurate networks.
|
|
The Curse of Diversity in Ensemble-Based Exploration
Zhixuan Lin,
Pierluca D'Oro,
Evgenii Nikishin,
Aaron Courville
ICLR 2024
[PDF]
We demonstrate that individual ensemble members in RL exhibit surprisingly low performance (whilst aggregate returns are adequate) and propose a remedy.
|
|
Deep Reinforcement Learning with Plasticity Injection
Evgenii Nikishin,
Junhyuk Oh,
Georg Ostrovski,
Clare Lyle,
Razvan Pascanu,
Will Dabney,
André Barreto
NeurIPS 2023 (Spotlight); also ICLR 2023 Workshop Track (Spotlight)
[PDF,
Poster]
We propose an intervention for diagnosing the loss of plasticity phenomenon in RL and dynamically growing neural networks in RL for increasing computational efficiency.
|
|
Understanding Plasticity in Neural Networks
Clare Lyle,
Zeyu Zheng,
Evgenii Nikishin,
Bernardo Ávila Pires,
Razvan Pascanu,
Will Dabney
ICML 2023 (Oral)
[PDF]
An analysis of plasticity loss showing its relation to pathological loss landscapes and demonstrating efficiency of layer normalization to mitigate it.
|
|
Sample-Efficient Reinforcement Learning by Breaking the Replay Ratio Barrier
Pierluca D'Oro*,
Max Schwarzer*,
Evgenii Nikishin,
Pierre-Luc Bacon,
Marc G. Bellemare,
Aaron Courville
ICLR 2023 (Oral); also NeurIPS 2022 Workshop Track
[PDF,
Poster,
Code]
Resets unlock increasing sample efficiency by scaling the number of updates per environment step.
|
|
The Primacy Bias in Deep Reinforcement Learning
Evgenii Nikishin*,
Max Schwarzer*,
Pierluca D'Oro*,
Pierre-Luc Bacon,
Aaron Courville
ICML 2022; also RLDM 2022
[PDF,
Short RLDM version,
Code,
Poster,
Blog]
We identify a damaging tendency of deep RL agents to overfit early experiences and propose a simple yet powerful remedy based on periodic resetting of a part of the agent.
|
|
Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation
Evgenii Nikishin,
Romina Abachi,
Rishabh Agarwal,
Pierre-Luc Bacon
AAAI 2022; also ICML 2021 Workshop Track
[PDF,
Code,
Poster]
A model learning method for RL that directly optimizes the sum of rewards instead of likelihood, a proxy to the agent's objective.
|
|
Quantifying and Understanding Adversarial Examples in Discrete Input Spaces
Volodymyr Kuleshov,
Evgenii Nikishin,
Shantanu Thakoor,
Tingfung Lau,
Stefano Ermon
ArXiv
[PDF]
We propose an algorithm to construct discrete token adversarial examples based on the notion of synonyms.
|
|
Unsupervised Domain Adaptation with Shared Latent Dynamics for Reinforcement Learning
Evgenii Nikishin,
Arsenii Ashukha,
Dmitry Vetrov
NeurIPS 2019 Workshop Track
[PDF,
Code,
Poster]
Domain adaptation via learning shared dynamics in a latent space with adversarial matching of latent states.
|
|
Improving Stability in Deep Reinforcement Learning with Weight Averaging
Evgenii Nikishin,
Pavel Izmailov,
Ben Athiwaratkun,
Dmitrii Podoprikhin,
Timur Garipov,
Pavel Shvechikov,
Dmitry Vetrov,
Andrew Gordon Wilson
UAI 2018 Workshop Track
[PDF,
Poster]
Averaging weights during training of an RL agent stabilizes the achieved performance.
|
|