In this paper, an algorithm, for in-parallel, greedy experience generator (briefly IPE, In Parallel Experiences), has been crafted, and added to the Double Deep Q-Learning algorithm. The algorithm aims to perturbs the weights of the online network, and as results, the network, trying to recover from the perturbed weights, escapes from the local minima. DDQN with IPE takes about the double of time of the previous to compute, but even if it slows down the learning rate in terms of wall clock time, the solution converges faster in terms of number of epochs.

Double Deep Q Network with In-Parallel Experience Generator

Dentamaro V.;Impedovo D.;Pirlo G.;Gattulli V.
2020-01-01

Abstract

In this paper, an algorithm, for in-parallel, greedy experience generator (briefly IPE, In Parallel Experiences), has been crafted, and added to the Double Deep Q-Learning algorithm. The algorithm aims to perturbs the weights of the online network, and as results, the network, trying to recover from the perturbed weights, escapes from the local minima. DDQN with IPE takes about the double of time of the previous to compute, but even if it slows down the learning rate in terms of wall clock time, the solution converges faster in terms of number of epochs.
2020
978-1-7281-4384-2
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11586/310645
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact