Open Media Forensics Challenge

Overview
The Open Media Forensics Challenge (OpenMFC) is a media forensics evaluation to facilitate development of systems that can automatically detect and locate manipulations in imagery (i.e., images and videos).
What
The NIST OpenMFC evaluation is being conducted to examine the performance of system’s accuracy and robustness over diverse datasets collected under controlled environments.
Who
The NIST OpenMFC is open worldwide. We invite all organizations including past DARPA MediFor Program participants to submit their results using their technologies to the OpenMFC evaluation server. Participation is free. NIST does not provide funds to participants.
How
To take part in the OpenMFC evaluation you need to register on this website and complete the data license to download the data. Once your system is functional you will be able to upload your outputs to the challenge website and see your results displayed on the leaderboard.
Evaluation Plan
OpenMFC 2022 Evaluation Plan [Download Link]
Task coordinator
If you have any question, please email to the NIST MFC team: [email protected]
OpenMFC Tasks Summary

In general, there are multiple tasks in media forensic applications. For example, manipulation detection and localization, Generative Adversarial Network (GAN) detection, image splice detection and localization, event verification, camera verification, and provenance history analysis etc.

The OpenMFC initially focuses on manipulation detection and deepfake tasks. In future, challenges may be expanded with community interest. The OpenMFC 2022 has following three task categories: Manipulation Detection (MD), Deepfakes Detection (DD), and Steganography Detection (StegD).

  • MD: Manipulation Detection
  • DD: Deepfakes Detection
  • StegD: Steganography Detection

A brief summary of each category and their tasks are described below. In the summary, the evaluation media is described in the following way: A ‘base’ indicates original media with high provenance, while a ‘probe’ indicates a test media. A ‘donor’ indicates another media whose content was donated into the base media and generated the probe media. For a full description of the evaluation tasks, please refer to the OpenMFC 2022 Evaluation Plan [Download Link].

  • Manipulation Detection (MD)

    The objective for Manipulation Detection (MD) is to detect if a probe has been manipulated, and if so, to spatially localize the edits. Manipulation in this context is defined as deliberate modifications of media (e.g., splicing and cloning etc.) and localization is encouraged but not required for OpenMFC.

    The MD task includes three tasks, namely,

    • Image_MD (or IMD): Image Manipulation Detection
    • The Image Manipulation Detection task is to detect if the image has been manipulated, and then to spatially localize the manipulated region. For detection, the IMD system provides a confidence score for all probe (i.e., a test image) with higher numbers indicating the image is more likely to have been manipulated. The target probes (i.e., probes that should be detected as manipulated) included potentially any image manipulations while the non-target probes (i.e., probes not containing image manipulations) include only high provenance images that are known to be original. Systems are required to process and report a confidence score for every probe.

      For the localization part of the task, the system provides an image bit-plane mask (either binary or greyscale) that indicates the manipulated pixels. Only local manipula­tions (e.g., clone) require a mask output while global manipulations (e.g., blur) affecting the entire image do not require a mask.

    • ImageSplice_MD (or ISMD): Image Splice Manipulation Detection
    • The new task, Image Splice Manipulation Detection, is added in the OpenMFC 2022 to support entry-level public participants. The ISMD is designed for 'splice' manipulation operation only. The testing dataset is a small-size dataset (2K images), which contains either original images without any manipulation, or spliced images. The ISMD task will detect if a probe image has been spliced.

    • Video_MD (or VMD): Video Manipulation Detection
    • The Video Manipulation Detection (VMD) task is to detect if the video has been manipulated. In this task, the localization of spatial/temporal-spatial manipulated regions is not addressed. For detection, the VMD system provides a confidence score for all probes (i.e, a test video) with higher numbers indicating the video is more likely to have been manipulated. For VMD, target probes (i.e., probes that should be detected as manipulated) included potentially any video manipulations while the non-target probes (i.e., probes not containing video manipulations) include only high provenance videos that are known to be original. Systems are required to process and report a confidence score for every probe.

  • Deepfakes Detection (DD)

    With recent advances in DeepFakes techniques and GAN (Generative Adversarial Network), imagery producers are able to generate realistic fake objects in media. The objective for Deepfakes Detection (DD) is to detect if a probe has been Deepfakes or GAN manipulated.

    The DD task includes two tasks based on testing media type, namely,

    • Image_DD (or IDD): Image Deepfakes Detection
    • The Image Deepfakes Detection task evaluates if a system can detect Deepfaked images (e.g. a face video is manipulated by a Deepfakes tool, then deepfaked face image frames are extracted as image files etc.) or GAN-manipulated images (e.g. created by a GAN model, locally/globally modified by a GAN filter/operation, etc.) specifically while not detecting other forms of manipulations. In the testing datasets, the non-target probes are high provenance images (or cropped high provenance images) and the target images are the ones manipulated with Deepfakes or GAN techniques.
    • Video_DD (or VDD): Video Deepfakes Detection
    • The Video Deepfakes Detection task evaluates if a system can detect Deepfaked videos. The target probes include Deepfaked videos while the non-target probes include high provenance original videos or video clips.
  • Steganography Detection (StegD)
    • StegD: Steganography Detection
    • The Steganography Detection task evaluates if a probe is a stego image, which contains the hidden message either in pixel values or in optimally selected coefficients. For each StegD trial, which consists of a single probe image, the StegD system must render a confidence score with higher numbers indicating the probe image is more likely to be a stego image.
Conditions

All probes must be processed independently of each other within a given task and across all tasks, meaning content extracted from probe data must not affect another probe.

For the OpenMFC 2022 evaluation, all tasks should run under the following conditions:

  • Image Only (IO)

    For the image tasks, the system is only allowed to use the pixel-based content for images as input to the system. No image header or other information should be used.

  • Video Only (VO)

    For the video tasks, the system is only allowed to use the pixel-based content for videos and audio (if audio exists) as input. No video header or other information should be used.

Metrics

For detection performance assessment, system performance is measured by Area Under Curve (AUC) which is the primary metric and the Correct Detection Rate at a False Alarm Rate of 5% (CDR@FAR = 0.05) from the Receiver Operating Characteristic (ROC) as shown Figure (a) below. This applies to both image and video tasks.

For the image localization performance assessment, the Optimum Matthews Correlation Coefficient (MCC) is the primary metric. The optimum MCC is calculated using an ideal mask-specific threshold found by computing metric scores over all pixel thresholds. Figure (b) below shows a visualization of the different mask regions used for mask image evaluations.

  • TP (True Positive) is an overlap area (green) of the reference mask and system output mask as manipulated at the pixel threshold.
  • FN (False Negative) is the reference mask indicates as manipulated, but the system did not detect it as manipulated at the threshold (red).
  • FP (False Positive) is the reference mask indicates not-manipulated, but the system detected it as manipulated at the threshold (orange).
  • TN (True Negative) is the reference mask indicates not-manipulated, and the system also detects it as not-manipulated at the threshold (white).
  • NS (No-Score) is the region of the reference mask not scored, the result of the dilation and erosion operations (purple).
  • If the denominator is zero, then we set MCC_o = 0.
  • If MCC_o = 1, there is a perfect correlation between the reference and system output masks.
  • If MCC_o = 0, there is no correlation between the reference mask and the system output mask.
  • If MCC_o = -1, there is perfect anti-correlation.

ROC and AUC

Figure 1. Detection System Performance Metrics: Receiver Operating Characteristic (ROC) and Area Under the Curve (AUC)

Localization Metrics

Figure 2. Localization System Performance Metrics: Optimum Matthews Correlation Coefficient (MCC)

Localization Metrics

Figure 3. An Example of Localization System Evaluation Report


Overview

Registered participants will get access to datasets created by the DARPA Media Forensics (MediFor) Program [Website Link]. During the registration process, registrants will get the data access credentials.

There will be both development data sets (those which include reference material) and evaluation data sets (which consist of only probe images to test systems). Each data set is structured similarly as described on the “MFC Data Set Structure Summary” section below.

Evaluation Datasets

Manipulation Detection (MD) Task
  • Image Manipulation Detection (IMD): OpenMFC20_Image_MD: previous MFC19 Image Data, generated from over 700 image journals, with more than 16K test images.
  • Video Manipulation Detection (VMD): OpenMFC20_Video_MD: previous MFC19 Video Data, generated from over 100 video journals, with about 1K test videos.
  • (NEW) Image Splice Manipulation Detection (ISMD): OpenMFC22_SpliceImage_MD: about 500 test images.
  • Deepfake Detection (DD) Task
  • Image Deepfake Detection (IDD): OpenMFC20_Image_DD: previous MFC18 GAN Full Image Data, generated from over 200 image journals, with more than 1K test images.
  • Video Deepfake Detection (VDD): OpenMFC20_Video_DD: previous MFC18 GAN Video Data, over 100 test videos.
  • Steganography Detection (StegD) Task
  • (NEW) Steganography Detection (StegD): OpenMFC22_Image_StegD
  • Development Datasets

    • DARPA MediFor NC16 kickoff image dataset: 1200 images, size is about 4GB.
    • DARPA MediFor NC17 development dataset: 3.5K images from about 400 journals, over 200 videos from over 20 journals, total size is about 441GB.
    • DARPA MediFor NC17 Evaluation Part 1 image dataset: 4K images from over 400 journals, size is about 41GB.
    • DARPA MediFor NC17 Evaluation Part 1 video dataset: 360 videos from 45 journals, size is about 117GB.
    • DARPA MediFor MFC18 development 1 image dataset: 5.6K images from over 177 journals, size is about 80GB.
    • DARPA MediFor MFC18 development 2 video dataset: 231 videos from 36 journals, size is about 57GB.

    MFC Dataset Structure Summary
    • Summary

      NIST OpenMFC dataset is designed and used for NIST OpenMFC evaluation. The datasets include the following items:

      • Original high-provenance image or video
      • Manipulated image or video
      • (optional) The reference ground-truth information for detection and localization

    • Dataset Structure

      • README.txt
      • /probe - Directory of images/videos to be analyzed for specified task
      • /indexes - Directory of index files indicating which images should be analyzed
      • /references - Directory of reference ground-truth information. For evaluation datasets, this directory is sequestered for evaluation purposes. For resource datasets, this directory could be released to public.

    • MFC Dataset Index File

      The index files are pipe-separated CSV formatted files. The index file for the Manipulation task will have the columns:

      • TaskID: Detection task (e.g. "manipulation")
      • ProbeFileID: Label of the probe image (e.g. NC2016_9397)
      • ProbeFileName: Full filename and relative path of the probe image (e.g. /probe/NC2016_9397.jpg)
      • ProbeWidth: Width of the probe image (e.g. 4032)
      • ProbeHeight: Height of the probe image (e.g. 3024)
      • ProbeFileSize: File size of probe (e.g. 4049990) (manipulation task only)
      • HPDeviceID: Camera device ID. If "UNDEF", the data is not provided for training. (e.g. PAR-A077)

    MFC Data Visual examples
    OpenMFC Evaluation Tentative Schedule

    Date Event
    Nov. 14-15, 2023 OpenMFC 2023 workshop
    Oct. 25, 2023 OpenMFC 2023 submission deadline
    Dec. 6-7, 2022 OpenMFC 2022 workshop
    Nov. 15, 2022 OpenMFC 2022 submission deadline
    Aug. 1 - Aug. 30, 2022 OpenMFC 2022 participant pre-challenge phase (QC testing)
    July 29, 2022 OpenMFC STEG challenge dataset available
    July 28, 2022 OpenMFC 2022 Leaderboard open for the next evaluation cycle
    Jul. 26, 2022 (New) OpenMFC dataset resource website
    Mar. 03, 2022 OpenMFC2022 Eval Plan available
    Feb. 15, 2022 OpenMFC2021 Workshop Talks and Slides available
    Dec. 7- 10, 2021 OpenMFC/TRECVID 2021 Virtual Workshop
    Nov. 1, 2021 OpenMFC 2021 Virtual Workshop agenda finalization
    Oct. 30, 2021 OpenMFC 2020-2021 submission deadline
    May 15, 2021 OpenMFC 2020-2021 submission open
    April 23, 2021 - May 09, 2021
    • OpenMFC 2020-2021 participant pre-challenge phase (QC testing)
      • Participant dry-run submission
      • NIST leaderboard testing/validation result
    August 31, 2020 OpenMFC evaluation GAN image and video dataset available
    August 21, 2020 OpenMFC evaluation image and video dataset available
    August 17, 2020 OpenMFC development datasets resource available

    Image Manipulation Detection (IMD)

    IMD-IO (Image Only)

    Updated: 2024-06-13 11:45:04 -0400
    RANK SUBMISSION ID SUBMISSION DATE TEAM NAME SYSTEM NAME AUC [email protected] ROC CURVE AVERAGE OPTIMAL MCC
    1 63 2021-06-07 11:26:58 Mayachitra test1june6 0.993707 0.972
    2 10 2020-11-05 21:53:02 UIIA naive-efficient 0.616186 0.071351
    3 67 2021-06-08 00:51:16 UIIA testIMDL 0.5 0.05
    4 81 2021-06-26 00:31:16 UIIA testIMDL 0.0553688699305928

    Updated:

    IMD-IM (Image and Metadata)

    Updated: 2024-06-13 11:45:04 -0400
    RANK SUBMISSION ID SUBMISSION DATE TEAM NAME SYSTEM NAME AUC [email protected] ROC CURVE AVERAGE OPTIMAL MCC

    Updated: