SlideShare a Scribd company logo
1 of 21
B.E (ISE)-VI Semester-B SEC 2020
DEEPFAKES
NEURAL NETWORKS
ANALYZING THE TECHNOLOGY
PROCESS
APPLICATION
3D HEADPOSE ESTIMATION
INCONSISTENT HEAD POSES IN DEEP FAKES
CLASSIFICATION BASED ON HEAD POSES
CONCLUSION
Well, what you see is not always what you see!
The word deepfake has been around only for a
couple years. It is a combination of “deep
learning” – which is a subset of AI that uses
neural networks – and “fake.” It is a technique
for human image synthesis that is used to combine
and superimpose existing images and videos onto
source images or videos using a machine learning
technique known generative adversarial network.
The phrase deep fakes was coined in 2017. The term
is named for a Reddit user known as deep fake who,
in December 2017, used deep fake technology to edit
the faces of celebrities onto people in pornographic
video clips. These videos and audios look and sound
just like the real thing. Deep fakes are lies disguised to
look like truth.
A deep neural network is a concept of deep learning and it is what artificial
intelligence researchers call computer systems that have been trained to do
specific tasks, in this case, recognize altered images. These networks are
organized in connected layers. Deep neural network architecture can identify
manipulated images at the pixel level with high precision. These neural
networks are also used in our snapchat and Instagram filters.
Here are the specific parameters for defining what constitutes a
successful deepfake.
The following criteria have been defined to evaluate these
requirements:
 Number of images
 Lighting conditions
 Size/quality of the source material
 Angle of the source material
 Differing facial structures
 Overlapping objects
PROCESS
At the moment there are two main applications used to create deep fakes: FakeApp and faceswap. It
requires three steps: extraction, training and creation.
Extraction
The deep- in deep fakes comes from the fact that this face-swap technology uses Deep Learning. It
often requires large amounts of data. Without hundreds of face pictures or some videos, you will
not be able to create a deepfake video.
A way to get around this is to collect a number of video clips which feature the people you want to
face-swap. The extraction process refers to the process of extracting all frames from these video
clips, identifying the faces and aligning them. The alignment is critical, since the neural network
that performs the face swap requires all faces to have the same size (usually 256×256 pixels) and
features aligned. Detecting and aligning faces is a problem that is considered mostly solved, and is
done by most applications very efficiently.
Training
Training is a technical term borrowed from Machine Learning. In this case, it
refers to the process which allows a neural network to convert a face into
another. Although it takes several hours, the training phase needs to be done only
once. Once completed, it can convert a face from person A into person B. This is
the most obscure part of the entire process.
Creation
Once the training is complete, it is finally time to create a deepfake. Starting from a video, all frames
are extracted and all faces are aligned. Then, each one is converted using the trained neural network.
The final step is to merge the converted face back into the original frame. While this sounds like an
easy task, it is actually where most face-swap applications go wrong.
The creation process is the only one which does not use any Machine Learning. This is a phase where
most of the mistakes are detected. Also, each frame is processed independently; there is no temporal
correlation between them, meaning that the final video might have some flickering.
EXAMPLES
Obama Deepfake
Jordan Peele
In, this deepfake, a false image or video seems
deceptively real, by the American actor and director
Jordan Peele, shows former US president Barack Obama
speaking about the dangers of false information and fake
news. Jordan Peele transferred his own facial movements
to Obama’s facial characteristics using deepfake
technology.
Mark Zuckerberg Deepfake
This particular deepfake manipulates the audio to make Facebook
CEO Zuckerberg sound like a psychopath talking to CBS News
about the "truth of Facebook and who really owns the future."
This video was widely circulated on Instagram and ultimately
went viral.
APPLICATIONS
FAKE APP
In January 2018, a proprietary desktop application called Fake App was launched. The app allows users to easily
create and share videos with faces swapped. The app uses an artificial neural network, a GPU, and three to four
gigabytes of storage space to generate the fake video. For detailed information, the program needs a lot of visual
material from the person to be inserted in order to learn which image aspects have to be exchanged, based on the
video sequences and images.
FACE SWAP
When applied correctly, this technique is uncannily good at swapping faces. But it has a major disadvantage: it
only works on pre-existing pictures.
It relies on neural networks, computational models that are loosely inspired by the way real brains process
information. This novel technique allows generating so-called deepfakes, which actually morph a person’s face
to mimic someone else’s features, although preserving the original facial expression.
3D HEAD POSE ESTIMATION
The 3D head pose corresponds to the rotation and translation of the world coordinates to the corresponding camera
coordinates. Specifically, denote [U,V,W]T as the world coordinates of one facial landmark, [X,Y,Z]T be its camera
coordinates, and (x,y)T be its image coordinates. The transformation between the world and the camera coordinate
systems can be formulated as
where R is the 3 × 3 rotation matrix, ~t is 3 × 1 translation vector. The transformation between camera and image
coordinate systems is defined as
where fx and fy are the focal lengths in the x- and y directions and (cx,cy) is the optical center, and s is an unknown scaling
factor.
In 3D head pose estimation, we need to solve the reverse problem, i.e, estimating s, R and ~t using the 2D image
coordinates and 3D world coordinates of the same set of facial landmarks obtained from a standard model, e.g, a 3D average
face model, assuming we know the camera parameter. Specifically, for a set of n facial landmark points, this can be
formulated as an optimization problem, as
that can be solved efficiently using the Levenberg-Marquardt algorithm [15]. The estimated R is the camera pose which is
the rotation of the camera with regards to the world coordinate, and the head pose is obtained by reversing it as RT (as R is an
orthornormal matrix).
INCONSISTENT HEAD POSES IN DEEP FAKES
As a result of swapping faces in the central face region in the Deep Fake process in Fig. 1, the
landmark locations of fake faces often deviate from those of the original faces. As shown in Fig.
1(c), a landmark in the central face region P0 is firstly affine-transformed into P0 in = MP0. After
the generative neural network, its corresponding landmark on the faked face is Q0 out.
As the configuration of the generative neural network in Deep Fake does not guarantee
landmark matching, and people have different facial structures, this landmark Q0 out on
generated face could have different locations to P0 in. Based on the comparing 51 central
region landmarks of 795 pairs of images in 64 × 64 pixels, the mean shifting of a landmark from
the input (Fig. 1(d)) to the output (Fig. 1(e)) of the generative neural network is 1.540 pixels,
and its standard deviation is 0.921 pixel. After an inverse transformation Q0 = M−1Q0 out, the
landmark locations Q0 in the faked faces will differ from the corresponding landmarks P0 in the
original face..
Fig. 1 Distribution of the cosine distance between ~vc and
~va for fake and real face images
CLASSIFICATION BASED ON
HEAD POSES
We further trained SVM classifiers based on the differences between head poses estimated using the full set of facial
landmarks and those in the central face regions to differentiate Deep Fakes from real images or videos. The features
are extracted in following procedures: (1) For each image or video frame, we run a face detector and extract 68 facial
landmarks using software package DLib [16]. (2) Then, with the standard 3D facial landmark model of the same 68
points from OpenFace2 [17], the head poses from central face region (Rc and tc) and whole face (Ra and ta) are
estimated with landmarks 18 − 36,49,55 (red in Fig. 2) and 1 − 36,49,55 (red and blue in Fig. 2), respectively. Here, we
approximate the camera focal length as the image width, camera center as image center, and ignore the effect of lens
distortion. (3) The differences between the obtained rotation matrices (Ra −Rc) and translation vectors are flattened
into a vector, which is standardized by subtracting its mean and divided by its standard deviation for classification.
Fig. 2. ROC curves of the SVM classification results, see texts for
details.
Experimental evaluations of our methods on a set of real face
images and Deep Fakes.
CONCLUSION
In this paper, we propose a new method to expose AIgenerated fake face images or videos
(commonly known as the Deep Fakes). Our method is based on observations that such Deep
Fakes are created by splicing a synthesized face region into the original image, and in doing so,
introducing errors that can be revealed when 3D head poses are estimated from the face
images. We perform experiments to demonstrate this phenomenon and further develop a
classification
REFERENCES
[1]https://en.wikipedia.org/wiki/Deepfake
[2]https://whatis.techtarget.com/definition/deepfake
[3]https://www.popularmechanics.com/technology/security/a28691128/deepfake-technology/
[4]https://www.csoonline.com/article/3293002/deepfake-videos-how-and-why-they-work.html
[5]https://edition.cnn.com/interactive/2019/01/business/pentagons-race-against-deepfakes/
[6]https://internethealthreport.org/2019/deepfakes-are-here-now-what/
[7]https://www.alanzucconi.com/2018/03/14/Introduction-to-deepfakes/
[8]https://www.forbes.com/sites/bernardmarr/2019/07/22/the-best-and-scariest-examples-of-ai-
enabled- deepfakes/#78a0783c2eaf
ANY QUERIES?

More Related Content

Similar to DeepFakes presentation : brief idea of DeepFakes

Real time multi face detection using deep learning
Real time multi face detection using deep learningReal time multi face detection using deep learning
Real time multi face detection using deep learning
Reallykul Kuul
 
Image–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlabImage–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlab
Ijcem Journal
 
PBL presentation p2.pptx
PBL presentation p2.pptxPBL presentation p2.pptx
PBL presentation p2.pptx
Tony383416
 

Similar to DeepFakes presentation : brief idea of DeepFakes (20)

International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Report face recognition : ArganRecogn
Report face recognition :  ArganRecognReport face recognition :  ArganRecogn
Report face recognition : ArganRecogn
 
N010226872
N010226872N010226872
N010226872
 
Face Scope.pptx
Face Scope.pptxFace Scope.pptx
Face Scope.pptx
 
Real time facial expression analysis using pca
Real time facial expression analysis using pcaReal time facial expression analysis using pca
Real time facial expression analysis using pca
 
Real time multi face detection using deep learning
Real time multi face detection using deep learningReal time multi face detection using deep learning
Real time multi face detection using deep learning
 
computer vision.pdf
computer vision.pdfcomputer vision.pdf
computer vision.pdf
 
IRJET- IoT based Door Lock and Unlock System using Face Recognition
IRJET- IoT based Door Lock and Unlock System using Face RecognitionIRJET- IoT based Door Lock and Unlock System using Face Recognition
IRJET- IoT based Door Lock and Unlock System using Face Recognition
 
IRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for BlindIRJET- Wearable AI Device for Blind
IRJET- Wearable AI Device for Blind
 
vision correcting display
vision correcting displayvision correcting display
vision correcting display
 
Python Project.pptx
Python Project.pptxPython Project.pptx
Python Project.pptx
 
Antispoofing techniques in Facial recognition
Antispoofing techniques in Facial recognitionAntispoofing techniques in Facial recognition
Antispoofing techniques in Facial recognition
 
Deepfakesの生成および検出
Deepfakesの生成および検出Deepfakesの生成および検出
Deepfakesの生成および検出
 
Face Recognition
Face RecognitionFace Recognition
Face Recognition
 
Iaetsd multi-view and multi band face recognition
Iaetsd multi-view and multi band face recognitionIaetsd multi-view and multi band face recognition
Iaetsd multi-view and multi band face recognition
 
Face recognition
Face recognitionFace recognition
Face recognition
 
Image–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlabImage–based face-detection-and-recognition-using-matlab
Image–based face-detection-and-recognition-using-matlab
 
Image_recognition.pptx
Image_recognition.pptxImage_recognition.pptx
Image_recognition.pptx
 
PBL presentation p2.pptx
PBL presentation p2.pptxPBL presentation p2.pptx
PBL presentation p2.pptx
 
ppt[1].pptx
ppt[1].pptxppt[1].pptx
ppt[1].pptx
 

More from MayuraD1

SDP_May2023:student developement program
SDP_May2023:student developement programSDP_May2023:student developement program
SDP_May2023:student developement program
MayuraD1
 
cyber_security_brochure details of workshop
cyber_security_brochure details of workshopcyber_security_brochure details of workshop
cyber_security_brochure details of workshop
MayuraD1
 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester Elective
MayuraD1
 
Introduction to Machine Learning Elective Course
Introduction to Machine Learning Elective CourseIntroduction to Machine Learning Elective Course
Introduction to Machine Learning Elective Course
MayuraD1
 
Cloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptxCloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptx
MayuraD1
 
Introduction to Computer Forensics for all streams.
Introduction to Computer Forensics for all streams.Introduction to Computer Forensics for all streams.
Introduction to Computer Forensics for all streams.
MayuraD1
 
Introduction about research methodology explained in simple language.
Introduction about research methodology explained in simple language.Introduction about research methodology explained in simple language.
Introduction about research methodology explained in simple language.
MayuraD1
 

More from MayuraD1 (13)

SDP_May2023:student developement program
SDP_May2023:student developement programSDP_May2023:student developement program
SDP_May2023:student developement program
 
cyber_security_brochure details of workshop
cyber_security_brochure details of workshopcyber_security_brochure details of workshop
cyber_security_brochure details of workshop
 
Machine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester ElectiveMachine learning Module-2, 6th Semester Elective
Machine learning Module-2, 6th Semester Elective
 
Introduction to Machine Learning Elective Course
Introduction to Machine Learning Elective CourseIntroduction to Machine Learning Elective Course
Introduction to Machine Learning Elective Course
 
Module1 of Introduction to Machine Learning
Module1 of Introduction to Machine LearningModule1 of Introduction to Machine Learning
Module1 of Introduction to Machine Learning
 
Research Methodology Module 1 useful for any course work
Research Methodology Module 1 useful for any course workResearch Methodology Module 1 useful for any course work
Research Methodology Module 1 useful for any course work
 
Cloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptxCloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptx
 
Introduction to Computer Forensics for all streams.
Introduction to Computer Forensics for all streams.Introduction to Computer Forensics for all streams.
Introduction to Computer Forensics for all streams.
 
Cloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptxCloud computing_Applications and paradigams.pptx
Cloud computing_Applications and paradigams.pptx
 
Introduction about research methodology explained in simple language.
Introduction about research methodology explained in simple language.Introduction about research methodology explained in simple language.
Introduction about research methodology explained in simple language.
 
Cloud Computing Introduction. Engineering seventh Semester
Cloud Computing Introduction. Engineering seventh SemesterCloud Computing Introduction. Engineering seventh Semester
Cloud Computing Introduction. Engineering seventh Semester
 
M2.pptx
M2.pptxM2.pptx
M2.pptx
 
M5.pptx
M5.pptxM5.pptx
M5.pptx
 

Recently uploaded

ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdfONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
Kamal Acharya
 
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
Lovely Professional University
 
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdfDR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
DrGurudutt
 
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
MohammadAliNayeem
 

Recently uploaded (20)

Lab Manual Arduino UNO Microcontrollar.docx
Lab Manual Arduino UNO Microcontrollar.docxLab Manual Arduino UNO Microcontrollar.docx
Lab Manual Arduino UNO Microcontrollar.docx
 
15-Minute City: A Completely New Horizon
15-Minute City: A Completely New Horizon15-Minute City: A Completely New Horizon
15-Minute City: A Completely New Horizon
 
Online book store management system project.pdf
Online book store management system project.pdfOnline book store management system project.pdf
Online book store management system project.pdf
 
ROAD CONSTRUCTION PRESENTATION.PPTX.pptx
ROAD CONSTRUCTION PRESENTATION.PPTX.pptxROAD CONSTRUCTION PRESENTATION.PPTX.pptx
ROAD CONSTRUCTION PRESENTATION.PPTX.pptx
 
ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdfONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
ONLINE VEHICLE RENTAL SYSTEM PROJECT REPORT.pdf
 
How to Design and spec harmonic filter.pdf
How to Design and spec harmonic filter.pdfHow to Design and spec harmonic filter.pdf
How to Design and spec harmonic filter.pdf
 
Operating System chapter 9 (Virtual Memory)
Operating System chapter 9 (Virtual Memory)Operating System chapter 9 (Virtual Memory)
Operating System chapter 9 (Virtual Memory)
 
Research Methodolgy & Intellectual Property Rights Series 2
Research Methodolgy & Intellectual Property Rights Series 2Research Methodolgy & Intellectual Property Rights Series 2
Research Methodolgy & Intellectual Property Rights Series 2
 
RM&IPR M5 notes.pdfResearch Methodolgy & Intellectual Property Rights Series 5
RM&IPR M5 notes.pdfResearch Methodolgy & Intellectual Property Rights Series 5RM&IPR M5 notes.pdfResearch Methodolgy & Intellectual Property Rights Series 5
RM&IPR M5 notes.pdfResearch Methodolgy & Intellectual Property Rights Series 5
 
Multivibrator and its types defination and usges.pptx
Multivibrator and its types defination and usges.pptxMultivibrator and its types defination and usges.pptx
Multivibrator and its types defination and usges.pptx
 
NEWLETTER FRANCE HELICES/ SDS SURFACE DRIVES - MAY 2024
NEWLETTER FRANCE HELICES/ SDS SURFACE DRIVES - MAY 2024NEWLETTER FRANCE HELICES/ SDS SURFACE DRIVES - MAY 2024
NEWLETTER FRANCE HELICES/ SDS SURFACE DRIVES - MAY 2024
 
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
Activity Planning: Objectives, Project Schedule, Network Planning Model. Time...
 
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdfDR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
DR PROF ING GURUDUTT SAHNI WIKIPEDIA.pdf
 
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
Complex plane, Modulus, Argument, Graphical representation of a complex numbe...
 
Fabrication Of Automatic Star Delta Starter Using Relay And GSM Module By Utk...
Fabrication Of Automatic Star Delta Starter Using Relay And GSM Module By Utk...Fabrication Of Automatic Star Delta Starter Using Relay And GSM Module By Utk...
Fabrication Of Automatic Star Delta Starter Using Relay And GSM Module By Utk...
 
ANSI(ST)-III_Manufacturing-I_05052020.pdf
ANSI(ST)-III_Manufacturing-I_05052020.pdfANSI(ST)-III_Manufacturing-I_05052020.pdf
ANSI(ST)-III_Manufacturing-I_05052020.pdf
 
Quiz application system project report..pdf
Quiz application system project report..pdfQuiz application system project report..pdf
Quiz application system project report..pdf
 
"United Nations Park" Site Visit Report.
"United Nations Park" Site  Visit Report."United Nations Park" Site  Visit Report.
"United Nations Park" Site Visit Report.
 
Electrical shop management system project report.pdf
Electrical shop management system project report.pdfElectrical shop management system project report.pdf
Electrical shop management system project report.pdf
 
Instruct Nirmaana 24-Smart and Lean Construction Through Technology.pdf
Instruct Nirmaana 24-Smart and Lean Construction Through Technology.pdfInstruct Nirmaana 24-Smart and Lean Construction Through Technology.pdf
Instruct Nirmaana 24-Smart and Lean Construction Through Technology.pdf
 

DeepFakes presentation : brief idea of DeepFakes

  • 2. DEEPFAKES NEURAL NETWORKS ANALYZING THE TECHNOLOGY PROCESS APPLICATION 3D HEADPOSE ESTIMATION INCONSISTENT HEAD POSES IN DEEP FAKES CLASSIFICATION BASED ON HEAD POSES CONCLUSION
  • 3. Well, what you see is not always what you see!
  • 4.
  • 5. The word deepfake has been around only for a couple years. It is a combination of “deep learning” – which is a subset of AI that uses neural networks – and “fake.” It is a technique for human image synthesis that is used to combine and superimpose existing images and videos onto source images or videos using a machine learning technique known generative adversarial network. The phrase deep fakes was coined in 2017. The term is named for a Reddit user known as deep fake who, in December 2017, used deep fake technology to edit the faces of celebrities onto people in pornographic video clips. These videos and audios look and sound just like the real thing. Deep fakes are lies disguised to look like truth.
  • 6. A deep neural network is a concept of deep learning and it is what artificial intelligence researchers call computer systems that have been trained to do specific tasks, in this case, recognize altered images. These networks are organized in connected layers. Deep neural network architecture can identify manipulated images at the pixel level with high precision. These neural networks are also used in our snapchat and Instagram filters.
  • 7. Here are the specific parameters for defining what constitutes a successful deepfake. The following criteria have been defined to evaluate these requirements:  Number of images  Lighting conditions  Size/quality of the source material  Angle of the source material  Differing facial structures  Overlapping objects
  • 8. PROCESS At the moment there are two main applications used to create deep fakes: FakeApp and faceswap. It requires three steps: extraction, training and creation. Extraction The deep- in deep fakes comes from the fact that this face-swap technology uses Deep Learning. It often requires large amounts of data. Without hundreds of face pictures or some videos, you will not be able to create a deepfake video. A way to get around this is to collect a number of video clips which feature the people you want to face-swap. The extraction process refers to the process of extracting all frames from these video clips, identifying the faces and aligning them. The alignment is critical, since the neural network that performs the face swap requires all faces to have the same size (usually 256×256 pixels) and features aligned. Detecting and aligning faces is a problem that is considered mostly solved, and is done by most applications very efficiently.
  • 9. Training Training is a technical term borrowed from Machine Learning. In this case, it refers to the process which allows a neural network to convert a face into another. Although it takes several hours, the training phase needs to be done only once. Once completed, it can convert a face from person A into person B. This is the most obscure part of the entire process.
  • 10. Creation Once the training is complete, it is finally time to create a deepfake. Starting from a video, all frames are extracted and all faces are aligned. Then, each one is converted using the trained neural network. The final step is to merge the converted face back into the original frame. While this sounds like an easy task, it is actually where most face-swap applications go wrong. The creation process is the only one which does not use any Machine Learning. This is a phase where most of the mistakes are detected. Also, each frame is processed independently; there is no temporal correlation between them, meaning that the final video might have some flickering.
  • 11. EXAMPLES Obama Deepfake Jordan Peele In, this deepfake, a false image or video seems deceptively real, by the American actor and director Jordan Peele, shows former US president Barack Obama speaking about the dangers of false information and fake news. Jordan Peele transferred his own facial movements to Obama’s facial characteristics using deepfake technology. Mark Zuckerberg Deepfake This particular deepfake manipulates the audio to make Facebook CEO Zuckerberg sound like a psychopath talking to CBS News about the "truth of Facebook and who really owns the future." This video was widely circulated on Instagram and ultimately went viral.
  • 12. APPLICATIONS FAKE APP In January 2018, a proprietary desktop application called Fake App was launched. The app allows users to easily create and share videos with faces swapped. The app uses an artificial neural network, a GPU, and three to four gigabytes of storage space to generate the fake video. For detailed information, the program needs a lot of visual material from the person to be inserted in order to learn which image aspects have to be exchanged, based on the video sequences and images. FACE SWAP When applied correctly, this technique is uncannily good at swapping faces. But it has a major disadvantage: it only works on pre-existing pictures. It relies on neural networks, computational models that are loosely inspired by the way real brains process information. This novel technique allows generating so-called deepfakes, which actually morph a person’s face to mimic someone else’s features, although preserving the original facial expression.
  • 13. 3D HEAD POSE ESTIMATION The 3D head pose corresponds to the rotation and translation of the world coordinates to the corresponding camera coordinates. Specifically, denote [U,V,W]T as the world coordinates of one facial landmark, [X,Y,Z]T be its camera coordinates, and (x,y)T be its image coordinates. The transformation between the world and the camera coordinate systems can be formulated as where R is the 3 × 3 rotation matrix, ~t is 3 × 1 translation vector. The transformation between camera and image coordinate systems is defined as where fx and fy are the focal lengths in the x- and y directions and (cx,cy) is the optical center, and s is an unknown scaling factor.
  • 14. In 3D head pose estimation, we need to solve the reverse problem, i.e, estimating s, R and ~t using the 2D image coordinates and 3D world coordinates of the same set of facial landmarks obtained from a standard model, e.g, a 3D average face model, assuming we know the camera parameter. Specifically, for a set of n facial landmark points, this can be formulated as an optimization problem, as that can be solved efficiently using the Levenberg-Marquardt algorithm [15]. The estimated R is the camera pose which is the rotation of the camera with regards to the world coordinate, and the head pose is obtained by reversing it as RT (as R is an orthornormal matrix).
  • 15. INCONSISTENT HEAD POSES IN DEEP FAKES As a result of swapping faces in the central face region in the Deep Fake process in Fig. 1, the landmark locations of fake faces often deviate from those of the original faces. As shown in Fig. 1(c), a landmark in the central face region P0 is firstly affine-transformed into P0 in = MP0. After the generative neural network, its corresponding landmark on the faked face is Q0 out. As the configuration of the generative neural network in Deep Fake does not guarantee landmark matching, and people have different facial structures, this landmark Q0 out on generated face could have different locations to P0 in. Based on the comparing 51 central region landmarks of 795 pairs of images in 64 × 64 pixels, the mean shifting of a landmark from the input (Fig. 1(d)) to the output (Fig. 1(e)) of the generative neural network is 1.540 pixels, and its standard deviation is 0.921 pixel. After an inverse transformation Q0 = M−1Q0 out, the landmark locations Q0 in the faked faces will differ from the corresponding landmarks P0 in the original face..
  • 16. Fig. 1 Distribution of the cosine distance between ~vc and ~va for fake and real face images
  • 17. CLASSIFICATION BASED ON HEAD POSES We further trained SVM classifiers based on the differences between head poses estimated using the full set of facial landmarks and those in the central face regions to differentiate Deep Fakes from real images or videos. The features are extracted in following procedures: (1) For each image or video frame, we run a face detector and extract 68 facial landmarks using software package DLib [16]. (2) Then, with the standard 3D facial landmark model of the same 68 points from OpenFace2 [17], the head poses from central face region (Rc and tc) and whole face (Ra and ta) are estimated with landmarks 18 − 36,49,55 (red in Fig. 2) and 1 − 36,49,55 (red and blue in Fig. 2), respectively. Here, we approximate the camera focal length as the image width, camera center as image center, and ignore the effect of lens distortion. (3) The differences between the obtained rotation matrices (Ra −Rc) and translation vectors are flattened into a vector, which is standardized by subtracting its mean and divided by its standard deviation for classification.
  • 18. Fig. 2. ROC curves of the SVM classification results, see texts for details. Experimental evaluations of our methods on a set of real face images and Deep Fakes.
  • 19. CONCLUSION In this paper, we propose a new method to expose AIgenerated fake face images or videos (commonly known as the Deep Fakes). Our method is based on observations that such Deep Fakes are created by splicing a synthesized face region into the original image, and in doing so, introducing errors that can be revealed when 3D head poses are estimated from the face images. We perform experiments to demonstrate this phenomenon and further develop a classification