Trustworthy AI Principles to Face Adversarial Machine Learning: A Novel Study
PDF

Keywords

Trust
AI
ML
Adversarial
Attacks

How to Cite

Priya, CH.E.N. Sai, and Manas Kumar Yogi. 2023. “Trustworthy AI Principles to Face Adversarial Machine Learning: A Novel Study”. Journal of Artificial Intelligence and Capsule Networks 5 (3): 227-45. https://doi.org/10.36548/jaicn.2023.3.002.

Abstract

Artificial Intelligence (AI) has witnessed significant advancements in recent years, enabling its widespread adoption across various domains. However, this progress has also given rise to new challenges, particularly in the context of adversarial machine learning. Adversarial attacks exploit vulnerabilities in AI models, resulting in their misclassification or misbehaviour. To address this critical issue, it is crucial to develop trustworthy AI systems that can withstand such adversarial threats. This paper presents a comprehensive study that covers the types of adversarial machine learning cyber-attacks, methods employed by adversaries to launch such attacks, effective defence mechanisms, and potential future directions in the field. It starts by exploring various types of adversarial ML attacks, characteristics and potential consequences of each attack type, emphasizing the risks they pose to privacy, security, and fairness in AI systems and delving into the methods employed by adversaries to launch adversarial ML attacks. By understanding the tactics used by adversaries, researchers and practitioners can develop robust defence mechanisms that can withstand these attacks. Building upon this understanding, a range of defence strategies can be invented for defending against adversarial ML attacks and emerging research areas, such as the integration of secure multi-party computation, differential privacy, and federated learning are used to enhance the resilience of AI models. By understanding the nature of adversarial attacks and implementing effective defence strategies, AI systems can be fortified against malicious manipulations. The findings of this study contribute to the development of trustworthy AI systems, ensuring their resilience, transparency, and fairness.

PDF

References

Pierazzi, Fabio, et al. "Intriguing properties of adversarial ml attacks in the problem space." 2020 IEEE symposium on security and privacy (SP). IEEE, 2020.

Liu, Jinxin, et al. "Adversarial machine learning: A multilayer review of the state-of-the-art and challenges for wireless and mobile systems." IEEE Communications Surveys & Tutorials 24.1 (2021): 123-159.

Kumar, Ram Shankar Siva, et al. "Adversarial machine learning-industry perspectives." 2020 IEEE Security and Privacy Workshops (SPW). IEEE, 2020.

Newaz, AKM Iqtidar, et al. "Adversarial attacks to machine learning-based smart healthcare systems." GLOBECOM 2020-2020 IEEE Global Communications Conference. IEEE, 2020.

Sadeghi, Koosha, Ayan Banerjee, and Sandeep KS Gupta. "A system-driven taxonomy of attacks and defenses in adversarial machine learning." IEEE transactions on emerging topics in computational intelligence 4.4 (2020): 450-467.

Wang, Xianmin, et al. "The security of machine learning in an adversarial setting: A survey." Journal of Parallel and Distributed Computing 130 (2019): 12-23.

Hartl, Alexander, et al. "Explainability and adversarial robustness for rnns." 2020 IEEE Sixth International Conference on Big Data Computing Service and Applications (BigDataService). IEEE, 2020.

Vorobeychik, Yevgeniy. "The Many Faces of Adversarial Machine Learning." Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 37. No. 13. 2023.

Eshete, Birhanu. "Making machine learning trustworthy." Science 373.6556 (2021): 743-744.

Adversarial Attacks on Graph Neural Network: Techniques and Countermeasures

A Systematic Literature Review on Malicious Use of Reinforcement Learning

Wild patterns: Ten years after the rise of adversarial machine learning

Huang, Ling, Anthony D. Joseph, Blaine Nelson, Benjamin IP Rubinstein, and J. Doug Tygar. "Adversarial machine learning." In Proceedings of the 4th ACM workshop on Security and artificial intelligence, pp. 43-58. 2011.

Udi Weinsberg, Smriti Bhagat, Stratis Ioannidis, and Nina Taft. Blurme: Inferring and obfuscating user gender based on ratings. In RecSys, 2012.

Arvind Narayanan, Hristo Paskov, Neil Zhenqiang Gong, John Bethencourt, Emil Stefanov, Eui Chul Richard Shin, and Dawn Song. On the feasibility of internet-scale author identification. In IEEE S&P, 2012.