Strategies to architect AI Safety: Defense to guard AI from Adversaries
Rajagopal. A1 , Nirmala. V2
Section:Research Paper, Product Type: Journal Paper
Volume-7 ,
Issue-5 , Page no. 451-456, May-2019
CrossRef-DOI: https://doi.org/10.26438/ijcse/v7i5.451456
Online published on May 31, 2019
Copyright © Rajagopal. A, Nirmala. V . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
View this paper at Google Scholar | DPI Digital Library
How to Cite this Paper
- IEEE Citation
- MLA Citation
- APA Citation
- BibTex Citation
- RIS Citation
IEEE Style Citation: Rajagopal. A, Nirmala. V, “Strategies to architect AI Safety: Defense to guard AI from Adversaries,” International Journal of Computer Sciences and Engineering, Vol.7, Issue.5, pp.451-456, 2019.
MLA Style Citation: Rajagopal. A, Nirmala. V "Strategies to architect AI Safety: Defense to guard AI from Adversaries." International Journal of Computer Sciences and Engineering 7.5 (2019): 451-456.
APA Style Citation: Rajagopal. A, Nirmala. V, (2019). Strategies to architect AI Safety: Defense to guard AI from Adversaries. International Journal of Computer Sciences and Engineering, 7(5), 451-456.
BibTex Style Citation:
@article{A_2019,
author = {Rajagopal. A, Nirmala. V},
title = {Strategies to architect AI Safety: Defense to guard AI from Adversaries},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {5 2019},
volume = {7},
Issue = {5},
month = {5},
year = {2019},
issn = {2347-2693},
pages = {451-456},
url = {https://www.ijcseonline.org/full_paper_view.php?paper_id=4263},
doi = {https://doi.org/10.26438/ijcse/v7i5.451456}
publisher = {IJCSE, Indore, INDIA},
}
RIS Style Citation:
TY - JOUR
DO = {https://doi.org/10.26438/ijcse/v7i5.451456}
UR - https://www.ijcseonline.org/full_paper_view.php?paper_id=4263
TI - Strategies to architect AI Safety: Defense to guard AI from Adversaries
T2 - International Journal of Computer Sciences and Engineering
AU - Rajagopal. A, Nirmala. V
PY - 2019
DA - 2019/05/31
PB - IJCSE, Indore, INDIA
SP - 451-456
IS - 5
VL - 7
SN - 2347-2693
ER -
![]() |
![]() |
![]() |
346 | 230 downloads | 123 downloads |




Abstract
The impact of designing for safety of AI is critical for humanity in the AI era. With humans increasingly becoming dependent of AI, there is a need for neural networks that work reliably, inspite of Adversarial attacks. Attacks can be one of 3 types: I) Similar looking adversarial images that aim to deceive both human and computer intelligence, II) Adversarial attacks such as evasion and exploratory attacks, III) Hacker introduced occlusions/perturbations to misguide AI. The vision for Safe and secure AI for popular use is achievable. To achieve safety of AI, this paper contributes both a strategy and a novel deep learning architecture. To guard AI from adversaries, paper proposes 3 strategies: 1) Introduce randomness at inference time to hide the representation learning from adversaries/attackers, 2) Detect presence of adversaries by analyzing the input sequence to AI, 3) Exploit visual similarity against adversarial perturbations. To realize these strategies, this paper proposes a novel architecture, Dynamic Neural Defense (DND). This defense has 3 deep learning architectural features: I) By hiding the way a neural network learns from exploratory attacks using a random computation graph, DND evades attack. II) By analyzing input sequence to cloud AI inference engine with CNN-LSTM, DND detects fast gradient sign attack sequence. III) By inferring with visual similar inputs generated by VAE, any AI defended by DND approach doesn’t succumb to hackers. Thus, a roadmap to develop reliable, safe & secure AI is presented.
Key-Words / Index Term
AI, Deep Learning, AI Safety, AI Security, Neural Networks, Adversarial Attacks and Defences, autonomous AI
References
[1] Chakraborty, A., Alam, M., Dey, V., Chattopadhyay, A., & Mukhopadhyay, D, “Adversarial Attacks and Defences: A Survey”, CoRR, arXiv:1810.00069, 2018.
[2] Szegedy, C at el., “Intriguing properties of neural networks”, arXiv:1312.6199, 2013.
[3] Papernot, N., Goodfellow, I., Sheatsley, R., Feinman, R. and McDaniel, P., “cleverhans v1. 0.0: an adversarial machine learning library”, arXiv:1610.00768, 2016.
[4] Biggio, B at el, “Evasion attacks against machine learning at test time”, Joint European conference on machine learning and knowledge discovery in databases, Springer, pp. 387-402, 2013.
[5] Sitawarin, C., Bhagoji, A.N., Mosenia, A., Chiang, M., Mittal, P., “Darts: Deceiving autonomous cars with toxic signs”, arXiv:1802.06430, 2018.
[6] Kurakin, Alexey, I. Goodfellow, and S. Bengio. "Adversarial machine learning at scale." arXiv:1611.01236, 2016
[7] Yuan, Xiaoyong, Pan He, Qile Zhu, Xiaolin Li., "Adversarial examples: Attacks and defenses for deep learning." IEEE transactions on neural networks and learning systems, 2019.
[8] Amodei, Dario, Chris O, Jacob S, Paul C, John S, Dan M. "Concrete problems in AI safety", arXiv:1606.06565, 2016.
[9] Liu, G, Issa K, Abdallah K. "GanDef", arXiv:1903.02585, 2019.
[10] Carlini, Nicholas. "Is AmI Robust to Adversarial Examples?.", arXiv:1902.02322, 2019.
[11] Carlini, Nicholas, David W. "Defensive distillation is not robust to adversarial examples.", arXiv:1607.04311, 2016
[12] Mahmood S, Sruti B, Lujo B, and Michael K. “Accessorize to a Crime: Real and Stealthy Attacks on State-of-the-Art Face Recognition”, Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp. 1528-1540, 2016.
[13] Tramer, Florian, A Kurakin, N Papernot, I Goodfellow, D Boneh, P McDaniel. “Ensemble adversarial training: Attacks and defenses” arXiv:1705.07204 , 2017
[14] U.Kaur, Mahajan, Singh, “Trust Models in Cloud Computing”, International Journal of Scientific Research in Network Security and Communication, Vol.6, Issue.2, pp.19-23, 2018
[15] Arora, Sharma, ”Synthesis of Cryptography and Security Attacks", International Journal of Scientific Research in Network Security and Communication, Vol.5, Issue.5, pp.1-5, 2017
[16] Das at el., "Shield: Fast, practical defense & vaccination for deep learning", 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, ACM, pp. 196-204, 2018