Media Forensics

Trustworthy media content

Fraunhofer IDMT is focusing on research and development of various media forensics technologies to analyze, detect and localize manipulation, decontextualization, and fabrication in media content (audio, visual, text), using a combination of different techniques and competencies, including signal analysis, machine learning, reasoning and IT security methods. The goal is to support e. g. journalists, law enforcement, or media platforms in the process of content verification. By doing so, it helps avoiding potential negative impacts of misinformation, deepfakes, and other malicious uses of manipulated media.

News and upcoming events

 

Workshop / 30.6.2025

MAD'25

The 4th ACM International Workshop on Multimedia AI against Disinformation welcomes your contributions.

 

29.10.2024

EBU webinar on audio analysis

Luca Cuccovillo and Milica Gerhardt showcased AI-based audio tools to verify information developed in veraAI.

 

Project Results / 1.10.2024

Highlighted innovation of AI4Media

High-potential innovation of Fraunhofer IDMT and NISV “Improvement of search and retrieval of audio content” featured in the European's commission innovation radar

 

Research focus

Disinformation as a challenge

Thanks to the availability of ever-growing amounts of content, low-cost editing tools, advanced synthesis techniques for content generation, and an abundance of distribution and communication channels, the creation and distribution of disinformation in all forms has become cheap and easy, and increasingly common.

Obvious forms of disinformation include decontextualization, i.e. presentation of authentic material in a misleading or inappropriate context; manipulation, i.e. modification of existing material; and fabrication, where material is made up from scratch. Two terms that are commonly used in this context are

  • Shallowfakes/cheapfakes, a term that refers to media content created through transformation or editing of genuine content, for example, through deletion, splicing or doctoring, with the aim of manipulation or decontextualisation. Until now, most fakes have fallen into this category, as they are simple to create and yet can be very effective and convincing.
  • Deepfakes, a term that refers to media content that is fabricated using AI. Until now, these remain less common than shallowfakes/cheapfakes, but ease of use, pervasiveness and availability of technologies to create them improve by the day, and it is clear that they pose serious challenge for disinformation detection.

Content verification – the search for truth

The process of content verification can be considered a "search for truth" that applies falsification, similar to how scientific theories and hypotheses are or should be tested according to "critical rationalism": To answer an overall question (typically something along the lines of ‘does the material at hand capture a real event and its appropriate context?’), there must be falsifiable claims about the material. For an audio recording, for instance, this could look like this:

"This file was recorded on Dec 6, 2022 in Amsterdam, NL, using an iPhone 6 and its standard recording app. The recording was not processed afterwards. The SHA-512 hash of the original file is 9f86d081 … The file was uploaded to cloud service XYZ… no transcoding or other modifications were applied."

Content verification is the process of testing such claims (which can also be implicit) against facts and findings using human assessment and various tools. The more and the “richer” the claims provided and verified (i. e. not rejected), the more trustworthy the content. However, human capabilities are limited with respect to perception and speed necessary to conduct the testing - therefore, our goal is to develop approaches and tools that support this process in the best possible manner, focusing on a broad technological coverage for the audio domain. The objective is to provide solutions and methods to support verification, including

  • Technologies for the analysis of acquisition, editing and synthesis traces within A/V material, to understand whether and how it was recorded, encoded, edited or synthesized, and then use it for the falsification process, and especially for manipulation and synthesis detection and localization.
  • Technologies for content provenance analysis, i. e. detect relationships between A/V content items, to understand whether and how they were reused and transformed, and in which order they were created (including the detection of “root” items).
  • Technologies for automatic annotation of A/V material, to quickly research relevant material for content verification, i. e. for a specific event, a specific person, or to retrieve information about circumstances that can be used for the verification process, i. e. acoustic scene classification and event detection.

We focus primarily on broad technological coverage for the audio domain and collaborate with other organizations that specialize in other tasks and modalities. The aim is to provide a comprehensive set of tools that can enhance and accelerate the verification of content.

In addition, our research also includes development of technologies for active media authentication, which are based on a combination of digital signatures and signal analysis. The idea is that content providers can use this to proactively sign and “mark” content and related metadata, including synthetic content, to allow other stakeholders to check its authenticity afterwards. Both approaches, (passive) falsification and (active) authentication, have distinct advantages and disadvantages, and we believe the two approaches are not mutually exclusive. On the contrary, they are complementary and should be considered and used together wherever possible. 

 

How to proceed

Media forensics research includes various disciplines, such as signal analysis, machine learning, but also security and adversarial thinking. At Fraunhofer IDMT, we address the topic with a focus on audio, and a combination of signal analysis and machine learning. Both provide specific and somewhat complementary advantages and disadvantages regarding interpretability / explainability, robustness, and performance for several tasks.

We believe that there are particular challenges in media forensic research that need to be addressed:

  • The need to design and develop technologies that users can work with, to enable them to make the best possible decisions within the verification process; this also includes addressing trust aspects, especially explainability, bias (most importantly sample bias, by ensuring suitable selection of training data wherever applicable), adversarial robustness and generalizability, all of which needs to be supported by systematic evaluation.
  • The need to establish a “falsification culture” that ensures that whoever provides content to be verified, also provides enough information to enable a proper verification process.
  • The need to cooperate with many other disciplines related to disinformation analysis, including textual analysis, visual analysis, social network analysis, legal analysis, and others; similarly, disinformation analysis needs to be understood as a complex interplay between technology, market, law and norms.
  • The understanding that media forensics is a cat-and-mouse game, which requires continuous research and development and sustainable business models as well as cooperation between companies and research institutions – financing such activities solely through publicly funded projects will not be enough.
  • The understanding that not only machine bias, but also human bias is a topic for content verification, and that organizational and technical measures are needed to address this.