Evgenii Nikishin

I am a PhD student at Mila, University of Montreal, where I have the good fortune of being advised by Pierre-Luc Bacon and Aaron Courville. I am broadly interested in Reinforcement Learning and using capabilities of Deep Learning for decision making.

Previously, I spent 1.5 years at Cornell University. I received an MS in Computer Science from Higher School of Economics and Skoltech where I worked with Dmitry Vetrov at the Samsung Lab. I did my undergrad in CS at Lomonosov Moscow State University supervised by Alexander Dyakonov.

During the 2nd half of 2022, I interned in the RL team at DeepMind London working closely with André Barreto, Junhyuk Oh, and Will Dabney. I spent summer 2020 remotely interning at Mila with Pierre-Luc Bacon and Yoshua Bengio. In summer 2018, I interned at ETH Zürich with Gunnar Rätsch applying RL for personal treatment.

Email  /  CV  /  Google Scholar  /  Twitter  /  GitHub

News

  • The DeepMind internship paper on plasticity injection got a spotlight at NeurIPS 2023 and a spotlight at the Reincarnating RL workshop at ICLR 2023
  • Clare's paper on analysis of plasticity loss got an oral at ICML 2023
  • I've got an AI scholarship from the University of Montreal
  • Our paper showing how resets enable scaling computations in RL got an oral at ICLR 2023

  • Research
    Maxwell’s Demon at Work: Efficient Pruning by Leveraging Saturation of Neurons.
    Simon Dufort-Labbé, Pierluca D’Oro, Evgenii Nikishin, Razvan Pascanu, Pierre-Luc Bacon, Aristide Baratin
    Arxiv
    [PDF]

    We show that neuron saturation, which traditionally was viewed as undesirable, could instead lead to sparse yet accurate networks.

    The Curse of Diversity in Ensemble-Based Exploration
    Zhixuan Lin, Pierluca D'Oro, Evgenii Nikishin, Aaron Courville
    ICLR 2024
    [PDF]

    We demonstrate that individual ensemble members in RL exhibit surprisingly low performance (whilst aggregate returns are adequate) and propose a remedy.

    Deep Reinforcement Learning with Plasticity Injection
    Evgenii Nikishin, Junhyuk Oh, Georg Ostrovski, Clare Lyle, Razvan Pascanu, Will Dabney, André Barreto
    NeurIPS 2023 (Spotlight); also ICLR 2023 Workshop Track (Spotlight)
    [PDF, Poster]

    We propose an intervention for diagnosing the loss of plasticity phenomenon in RL and dynamically growing neural networks in RL for increasing computational efficiency.

    Understanding Plasticity in Neural Networks
    Clare Lyle, Zeyu Zheng, Evgenii Nikishin, Bernardo Ávila Pires, Razvan Pascanu, Will Dabney
    ICML 2023 (Oral)
    [PDF]

    An analysis of plasticity loss showing its relation to pathological loss landscapes and demonstrating efficiency of layer normalization to mitigate it.

    Sample-Efficient Reinforcement Learning by Breaking the Replay Ratio Barrier
    Pierluca D'Oro*, Max Schwarzer*, Evgenii Nikishin, Pierre-Luc Bacon, Marc G. Bellemare, Aaron Courville
    ICLR 2023 (Oral); also NeurIPS 2022 Workshop Track
    [PDF, Poster, Code]

    Resets unlock increasing sample efficiency by scaling the number of updates per environment step.

    The Primacy Bias in Deep Reinforcement Learning
    Evgenii Nikishin*, Max Schwarzer*, Pierluca D'Oro*, Pierre-Luc Bacon, Aaron Courville
    ICML 2022; also RLDM 2022
    [PDF, Short RLDM version, Code, Poster, Blog]

    We identify a damaging tendency of deep RL agents to overfit early experiences and propose a simple yet powerful remedy based on periodic resetting of a part of the agent.

    Control-Oriented Model-Based Reinforcement Learning with Implicit Differentiation
    Evgenii Nikishin, Romina Abachi, Rishabh Agarwal, Pierre-Luc Bacon
    AAAI 2022; also ICML 2021 Workshop Track
    [PDF, Code, Poster]

    A model learning method for RL that directly optimizes the sum of rewards instead of likelihood, a proxy to the agent's objective.

    Quantifying and Understanding Adversarial Examples in Discrete Input Spaces
    Volodymyr Kuleshov, Evgenii Nikishin, Shantanu Thakoor, Tingfung Lau, Stefano Ermon
    ArXiv
    [PDF]

    We propose an algorithm to construct discrete token adversarial examples based on the notion of synonyms.

    Unsupervised Domain Adaptation with Shared Latent Dynamics for Reinforcement Learning
    Evgenii Nikishin, Arsenii Ashukha, Dmitry Vetrov
    NeurIPS 2019 Workshop Track
    [PDF, Code, Poster]

    Domain adaptation via learning shared dynamics in a latent space with adversarial matching of latent states.

    Improving Stability in Deep Reinforcement Learning with Weight Averaging
    Evgenii Nikishin, Pavel Izmailov, Ben Athiwaratkun, Dmitrii Podoprikhin, Timur Garipov, Pavel Shvechikov, Dmitry Vetrov, Andrew Gordon Wilson
    UAI 2018 Workshop Track
    [PDF, Poster]

    Averaging weights during training of an RL agent stabilizes the achieved performance.

    Last update: Mar 2024.
    Credit for the template to Jon Barron.