Applied Ethical and Explainable AI in Adversarial Deepfake Detection: From Theory to Real-World Systems
Main Article Content
Abstract
Deepfake technology is advancing by the minute. This gives rise to increased privacy, trust and security risks. Such technology can be used for malicious activities like manipulating public opinion and spreading misinformation using social media. Adversarial machine learning techniques seem to be a strong defense in detecting and flagging deepfake content. But the challenge with practical use of many deepfake detection models is that they operate as black-boxes with little transparency or accountability in their decisions. This paper proposes a framework and guidelines to integrate ethical AI and explainable AI (XAI) - specifically techniques like SHAP and LIME, to make deepfake detection systems more transparent and trustworthy. We will propose guidelines to incorporate techniques which will make deepfake detection systems more accountable and explainable such that they make deepfake detection systems seamlessly deployable in the real world.
Article Details

This work is licensed under a Creative Commons Attribution 4.0 International License.
©2024 All rights reserved by the respective authors and JAIGC