Recent advances in the democratization of AI have been enabling the widespread use of generative models, causing the exponential rise of fake content. Nudification of over 680.000 women by a social bot, impersonation scams worth millions of dollars, or spreading political misinformation through synthetic politicians are just the footfall of the deep fake dystopia.
As every technology is simultaneously built with its counterpart to neutralize it, this is the perfect time to fortify our eyes with deep fake detectors. Deep fakes depend on photorealism to disable our natural detectors: we cannot simply look at a video to decide if it is real. On the other hand, this realism is not preserved in physiological, biological, and physical signals of deep fakes, yet. In this talk, I will begin with presenting our renowned FakeCatcher, which detects synthetic content in portrait videos using heart beats, as a preventive solution for the emerging threat of deep fakes. Detectors blindly utilizing deep learning are not as effective in catching fake content, as generative models keep producing formidably realistic results. My key assertion follows that such signals hidden in portrait videos can be used as an implicit descriptor of authenticity, like a generalizable watermark of humans, because they are neither spatially nor temporally preserved in deep fakes. Building robust and accurate deep detectors by exhaustively analyzing heartbeats, PPG signals, eye vergence, and gaze movements of deep fake actors reinforce our perception of reality.
Moreover, we also innovate novel models to detect the source generator of any deep fake by exploiting its heart beats to unveil residuals of different generative models. Achieving leading results over both existing datasets and our recently introduced in-the-wild dataset justifies our approaches and pioneers a new dimension in deep fake research.