Search for glossary terms (regular expression allowed)


Term Definition

A "deepfake" is a type of synthetic media that uses artificial intelligence to manipulate images, audio, or video to make it appear as if someone said or did something they didn't. This is achieved by training deep learning models on large datasets of real images and audio, allowing them to create highly realistic and convincing forgeries.

Here are some key points about deepfakes:


  • Deepfakes use various techniques like facial mapping, voice cloning, and video editing to seamlessly splice someone's face, voice, or body onto another person's.
  • Different algorithms and levels of sophistication exist, ranging from simple face swaps to complex, full-body deepfakes with manipulated expressions and dialogue.


  • Deepfakes can be used for creative purposes like humor, satire, or entertainment.
  • However, malicious applications raise significant concerns, such as:
    • Spreading misinformation and propaganda.
    • Defamation and identity theft.
    • Blackmail and financial scams.

Concerns and risks:

  • Deepfakes challenge our trust in visual and audio evidence, making it difficult to discern real from fake.
  • They can have serious consequences for individuals and society, impacting reputations, elections, and social order.

Current landscape:

  • Technology to detect deepfakes is constantly evolving, but it remains a challenge to identify sophisticated forgeries.
  • Addressing the ethical and societal implications of deepfakes requires collaboration between researchers, policymakers, and technology companies.

To learn more:

  • You can find numerous examples of deepfakes online, highlighting both their creative potential and potential harms.
  • Look for resources from reputable organizations investigating deepfakes and their impact, such as the Deepfake Detection Challenge and the Partnership on AI.