Show simple item record

dc.contributor.authorMeng, Li
dc.contributor.authorGoodwin, Morten
dc.contributor.authorYazidi, Anis
dc.contributor.authorEngelstad, Paal
dc.date.accessioned2023-02-21T13:11:37Z
dc.date.available2023-02-21T13:11:37Z
dc.date.created2023-01-19T11:57:07Z
dc.date.issued2022
dc.identifier.citationMeng, L., Goodwin, M., Yazidi, A. & Engelstad, P. (2022). Improving the Diversity of Bootstrapped DQN by Replacing Priors with Noise. IEEE Transactions on Games (TG), 1-10.en_US
dc.identifier.issn2475-1510
dc.identifier.urihttps://hdl.handle.net/11250/3052785
dc.descriptionAuthors accepted manuscripten_US
dc.descriptionPersonal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
dc.description.abstractQ-learning is one of the most well-known Reinforcement Learning algorithms. There have been tremendous efforts to develop this algorithm using neural networks. Bootstrapped Deep Q-Learning Network is amongst them. It utilizes multiple neural network heads to introduce diversity into Q-learning. Diversity can sometimes be viewed as the amount of reasonable moves an agent can take at a given state, analogous to the definition of the exploration ratio in RL. Thus, the performance of Bootstrapped Deep Q-Learning Network is deeply connected with the level of diversity within the algorithm. In the original research, it was pointed out that a random prior could improve the performance of the model. In this article, we further explore the possibility of replacing priors with noise and sample the noise from a Gaussian distribution to introduce more diversity into this algorithm. We conduct our experiment on the Atari benchmark and compare our algorithm to both the original and other related algorithms. The results show that our modification of the Bootstrapped Deep Q-Learning algorithm achieves significantly higher evaluation scores across different types of Atari games. Thus, we conclude that replacing priors with noise can improve Bootstrapped Deep Q-Learning’s performance by ensuring the integrity of diversities.en_US
dc.language.isoengen_US
dc.publisherIEEEen_US
dc.titleImproving the Diversity of Bootstrapped DQN by Replacing Priors With Noiseen_US
dc.typePeer revieweden_US
dc.typeJournal articleen_US
dc.description.versionacceptedVersionen_US
dc.rights.holder© 2022 IEEEen_US
dc.subject.nsiVDP::Teknologi: 500en_US
dc.source.pagenumber1-10en_US
dc.source.journalIEEE Transactions on Games (TG)en_US
dc.identifier.doihttps://doi.org/10.1109/TG.2022.3185330
dc.identifier.cristin2110227
cristin.qualitycode1


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record