in ,

On Catastrophic Interference in Atari 2600 Games, Hacker News

  

              

                                   (Submitted on 2020 Feb 07769

    

Abstract: Model-free deep reinforcement learning algorithms are troubled with poor Sample efficiency – learning reliable policies generally requires a vast % of interaction with the environment. One hypothesis is that catastrophic interference between various segments within the environment is an issue. In this paper, we perform a large-scale empirical study on the presence of catastrophic interference in the Arcade Learning Environment and find that learning particular game segments frequently degrades performance on previously learned segments. In what we term the Memento observation, we show that an identically parameterized agent spawned from a state where the original agent plateaued, reliably makes further progress. This phenomenon is general – we find consistent performance boosts across architectures, learning algorithms and environments. Our results indicate that eliminating catastrophic interference can contribute towards improved performance and data efficiency of deep reinforcement learning algorithms.

            

      

Submission history

From: Dibya Ghosh [view email]        [v1] Fri, 728 Feb : 2002: 06 UTC (2, 728 (KB) Read More

What do you think?

Leave a Reply

Your email address will not be published. Required fields are marked *

GIPHY App Key not set. Please check settings

idea4good / GuiLite, Hacker News

idea4good / GuiLite, Hacker News

Road Safety World Series T20 Match 1 Highlights: Sachin-Sehwag lead India Legends to victory by 7 wickets – The Indian Express, Indianexpress.com

Road Safety World Series T20 Match 1 Highlights: Sachin-Sehwag lead India Legends to victory by 7 wickets – The Indian Express, Indianexpress.com