High-dimensional always-changing environments constitute a hard challenge for current reinforcement learning techniques. Artificial agents, nowadays, are often trained off-line in very static and controlled conditions in simulation such that training observations can be thought as sampled i.i.d. from the entire observations space. However, in real world settings, the environment is often non-stationary and subject to unpredictable, frequent changes. In this paper we propose and openly release CRLMaze, a new benchmark for learning continually through reinforcement in a complex 3D non-stationary task based on ViZDoom and subject to several environmental changes. Then, we introduce an end-to-end model-free continual reinforcement learning strategy showing competitive results with respect to four different baselines and not requiring any access to additional supervised signals, previously encountered environmental conditions or observations

Continual Reinforcement Learning in 3D Non-stationary Environments / Vincenzo Lomonaco, Karan Desai, Eugenio Culurciello, Davide Maltoni. - ELETTRONICO. - (2020), pp. 999-1008. (Intervento presentato al convegno IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) tenutosi a Las Vegas nel 14-19 June 2020) [10.1109/CVPRW50498.2020.00132].

Continual Reinforcement Learning in 3D Non-stationary Environments

Vincenzo Lomonaco
;
Davide Maltoni
2020

Abstract

High-dimensional always-changing environments constitute a hard challenge for current reinforcement learning techniques. Artificial agents, nowadays, are often trained off-line in very static and controlled conditions in simulation such that training observations can be thought as sampled i.i.d. from the entire observations space. However, in real world settings, the environment is often non-stationary and subject to unpredictable, frequent changes. In this paper we propose and openly release CRLMaze, a new benchmark for learning continually through reinforcement in a complex 3D non-stationary task based on ViZDoom and subject to several environmental changes. Then, we introduce an end-to-end model-free continual reinforcement learning strategy showing competitive results with respect to four different baselines and not requiring any access to additional supervised signals, previously encountered environmental conditions or observations
2020
2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
999
1008
Continual Reinforcement Learning in 3D Non-stationary Environments / Vincenzo Lomonaco, Karan Desai, Eugenio Culurciello, Davide Maltoni. - ELETTRONICO. - (2020), pp. 999-1008. (Intervento presentato al convegno IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) tenutosi a Las Vegas nel 14-19 June 2020) [10.1109/CVPRW50498.2020.00132].
Vincenzo Lomonaco, Karan Desai, Eugenio Culurciello, Davide Maltoni
File in questo prodotto:
File Dimensione Formato  
Lomonaco_Continual_Reinforcement_Learning_in_3D_Non-Stationary_Environments_CVPRW_2020_paper (3).pdf

accesso aperto

Tipo: Postprint
Licenza: Licenza per accesso libero gratuito
Dimensione 1.12 MB
Formato Adobe PDF
1.12 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/769495
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 20
  • ???jsp.display-item.citation.isi??? 3
social impact