Emotional Deepfake Detection Via Voice Stress Analysis
  • Author(s): Asfiya Khanum ; Soubiya Siddiqua
  • Paper ID: 1711237
  • Page: 890-893
  • Published Date: 17-10-2025
  • Published In: Iconic Research And Engineering Journals
  • Publisher: IRE Journals
  • e-ISSN: 2456-8880
  • Volume/Issue: Volume 9 Issue 4 October-2025
Abstract

The rapid advancement of generative artificial intelligence has enabled the creation of highly convincing audio deepfakes, where synthetic voices can mimic real speakers with near-human accuracy Posing new threats in fraud, misinformation, and security. Current detection techniques largely rely on acoustic artifacts or signal irregularities, which are increasingly difficult to identify as synthesis models improve. This paper introduces a novel approach for emotional deepfake detection via voice stress analysis. By examining subtle stress and emotion-related cues—such as pitch fluctuations, jitter, shimmer, rhythm, and speech rate— we capture inconsistencies that synthetic voices struggle to replicate. Using emotional speech datasets alongside AI-generated voice samples, we train deep learning models to distinguish authentic from synthetic speech. Results highlight stress-based analysis as a promising defense against evolving deepfake audio attacks

Citations

IRE Journals:
Asfiya Khanum , Soubiya Siddiqua "Emotional Deepfake Detection Via Voice Stress Analysis" Iconic Research And Engineering Journals Volume 9 Issue 4 2025 Page 890-893

IEEE:
Asfiya Khanum , Soubiya Siddiqua "Emotional Deepfake Detection Via Voice Stress Analysis" Iconic Research And Engineering Journals, 9(4)