kth.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Investigating Vulnerabilities of Deep Neural Policies
KTH, School of Electrical Engineering and Computer Science (EECS), Computer Science, Network and Systems Engineering.
2021 (English)In: 37th Conference on Uncertainty in Artificial Intelligence, UAI 2021, Association For Uncertainty in Artificial Intelligence (AUAI) , 2021, p. 1661-1670Conference paper, Published paper (Refereed)
Abstract [en]

Reinforcement learning policies based on deep neural networks are vulnerable to imperceptible adversarial perturbations to their inputs, in much the same way as neural network image classifiers. Recent work has proposed several methods to improve the robustness of deep reinforcement learning agents to adversarial perturbations based on training in the presence of these imperceptible perturbations (i.e. adversarial training). In this paper, we study the effects of adversarial training on the neural policy learned by the agent. In particular, we follow two distinct parallel approaches to investigate the outcomes of adversarial training on deep neural policies based on worst-case distributional shift and feature sensitivity. For the first approach, we compare the Fourier spectrum of minimal perturbations computed for both adversarially trained and vanilla trained neural policies. Via experiments in the OpenAI Atari environments we show that minimal perturbations computed for adversarially trained policies are more focused on lower frequencies in the Fourier domain, indicating a higher sensitivity of these policies to low frequency perturbations. For the second approach, we propose a novel method to measure the feature sensitivities of deep neural policies and we compare these feature sensitivity differences in state-of-the-art adversarially trained deep neural policies and vanilla trained deep neural policies. We believe our results can be an initial step towards understanding the relationship between adversarial training and different notions of robustness for neural policies. 

Place, publisher, year, edition, pages
Association For Uncertainty in Artificial Intelligence (AUAI) , 2021. p. 1661-1670
Keywords [en]
Deep neural networks, Feature sensitivity, Fourier spectra, Fourier-domain, High sensitivity, Image Classifiers, Lower frequencies, Neural-networks, Policy-based, Reinforcement learning agent, Shift-and, Reinforcement learning
National Category
Physiology and Anatomy
Identifiers
URN: urn:nbn:se:kth:diva-317513Scopus ID: 2-s2.0-85121686550OAI: oai:DiVA.org:kth-317513DiVA, id: diva2:1695547
Conference
37th Conference on Uncertainty in Artificial Intelligence, UAI 2021, Virtual Online, 27-30 July 2021
Note

Duplicate in Scopus 2-s2.0-85163377135

QC 20220914

Available from: 2022-09-14 Created: 2022-09-14 Last updated: 2025-02-10Bibliographically approved

Open Access in DiVA

No full text in DiVA

Scopus

Authority records

Korkmaz, Ezgi

Search in DiVA

By author/editor
Korkmaz, Ezgi
By organisation
Network and Systems Engineering
Physiology and Anatomy

Search outside of DiVA

GoogleGoogle Scholar

urn-nbn

Altmetric score

urn-nbn
Total: 105 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf