\
insight
\
\
Insight
\
Reality Defender
Team
A new deepfake video making the rounds showcases the need to label deepfakes as generative/synthetic media, while also showing a live demonstration of a labeling technology developed by a consortium of technology companies.
Unlike other deepfake detection (and labeling) solutions, which predominantly use a single model in an attempt to detect deepfakes, Reality Defender uses an “ensemble of models” approach, measuring deepfakes against known (existing) deepfake models and theoretical (unknown and/or yet to exist) deepfake models. This not only provides comprehensive and more accurate detection than existing solutions; it also creates a way of detecting the latest deepfaking/GenAI methods and potentially those that have yet to exist.
We believe labeling deepfakes is a crucial part of the future, and we’re nearing a day where all platforms, apps, enterprises, and everyone potentially dealing with generated/deepfaked media institutes media labeling and removal protocols — regardless of whether or not legislation calls for it. Due to the advanced and ever-evolving nature of deepfakes, our multifaceted approach allows our platform to outperform any similar detection platforms while consistently remaining several steps ahead of bad actors.
Stanford University released their 2023 AI Index Report. The worthwhile read contains information on all facets of AI development, use, proliferation, and monetization. Buried within the text is the above graph, demonstrating the meteoric rise of AI incidents over the years. The latest number in the report comes from 2021, not accounting for last year’s introduction of ChatGPT, rise in popularity of Midjourney, and all related incidents. You can find the full report here.
Programmer Jessica Card hooked up a deconstructed Furby to ChatGPT to give the much-maligned ‘90s toy some AI-generated oomph. The results are as terrifying as you would expect.
"AI can help deal with some very difficult challenges like disease and climate change, but we also have to address the potential risks to our society, to our economy, to our national security." These were President Biden’s remarks during a recent speech, recognizing the impact weaponized deepfakes and generative content have on a societal level. Ars Technica has the full rundown from the speech.
Reality Defender's deepfake detection technology helped a national public broadcaster maintain journalistic integrity and audience trust by accurately identifying and preventing the spread of manipulated content.
Get an in-depth analysis of the benefits gained by the broadcaster from Reality Defender’s detection platform in our latest case study.
Thank you for reading the Reality Defender Newsletter. If you have any questions about Reality Defender, or if you would like to see anything in future issues, please reach out to us here.