Volltext-Downloads (blau) und Frontdoor-Views (grau)

Introducing natural adversarial observations to a Deep Reinforcement Learning agent for Atari Games

  • Deep Learning methods are known to be vulnerable to adversarial attacks. Since Deep Reinforcement Learning agents are based on these methods, they are prone to tiny input data changes. Three methods for adversarial example generation will be introduced and applied to agents trained to play Atari games. The attacks target either single inputs or can be applied universally to all possible inputs of the agents. They were able to successfully shift the predictions towards a single action or to lower the agent’s confidence in certain actions, respectively. All proposed methods had a severe impact on the agent’s performance while producing invisible adversarial perturbations. Since natural-looking adversarial observations should be completely hidden from a human evaluator, the negative impact on the performance of the agents should additionally be undetectable. Several variants of the proposed methods were tested to fulfil all posed criteria. Overall, seven generated observations for two of three Atari games are classified as natural-looking adversarial observations.

Download full text files

  • Master_Thesis_Hanfeld.pdf

Export metadata

Additional Services

Share in Twitter Search Google Scholar


Author:Pia Hanfeld
Advisor:Thomas Villmann, Jan Chorowski
Document Type:Master's Thesis
Year of Completion:2021
Granting Institution:Hochschule Mittweida
Release Date:2022/05/10
GND Keyword:Deep learning
Institutes:Angewandte Computer‐ und Bio­wissen­schaften
Open Access:Innerhalb der Hochschule
Licence (German):License LogoUrheberrechtlich geschützt