Peng, B., Fan, H., Wang, W., Dong, J., & Lyu, S.. (2021). A Unified Framework for High Fidelity Face Swap and Expression Reenactment. IEEE Transactions on Circuits and Systems for Video Technology
“Face manipulation techniques improve fast with the development of powerful image generation models. two particular face manipulation methods, namely face swap and expression reenactment attract much attention for their flexibility and ease to generate high quality synthesis results. recently, these two subjects are actively studied. however, most existing methods treat the two tasks separately, ignoring their underlying similarity. in this paper, we propose to tackle the two problems within a unified framework that achieves high quality synthesis results. the enabling component for our unified framework is the clean disentanglement of 3d pose, shape, and expression factors and then recombining them for different tasks accordingly. we then use the same set of 2d representations for face swap and expression reenactment tasks that are input to a common image translation model to directly generate the final synthetic images. once trained, the proposed model can accomplish both face swap and expression reenactment tasks for previously unseen subjects. comprehensive experiments and comparisons show that the proposed method achieves high fidelity results in multiple aspects, and it is especially good at faithfully preserving source facial shape in the face swap task, and accurately transferring facial movements in the expression reenactment task.”
Zhang, W., Zhao, C., & Li, Y.. (2020). A novel counterfeit feature extraction technique for exposing face-swap images based on deep learning and error level analysis. Entropy
“The quality and efficiency of generating face-swap images have been markedly strengthened by deep learning. for instance, the face-swap manipulations by deepfake are so real that it is tricky to distinguish authenticity through automatic or manual detection. to augment the efficiency of distinguishing face-swap images generated by deepfake from real facial ones, a novel counterfeit feature extraction technique was developed based on deep learning and error level analysis (ela). it is related to entropy and information theory such as cross-entropy loss function in the final softmax layer. the deepfake algorithm is only able to generate limited resolutions. therefore, this algorithm results in two different image compression ratios between the fake face area as the foreground and the original area as the background, which would leave distinctive counterfeit traces. through the ela method, we can detect whether there are different image compression ratios. convolution neural network (cnn), one of the representative technologies of deep learning, can extract the counterfeit feature and detect whether images are fake. experiments show that the training efficiency of the cnn model can be significantly improved by the ela method. in addition, the proposed technique can accurately extract the counterfeit feature, and therefore achieves outperformance in simplicity and efficiency compared with direct detection methods. specifically, without loss of accuracy, the amount of computation can be significantly reduced (where the required floating-point computing power is reduced by more than 90%).”
Korshunova, I., Shi, W., Dambre, J., & Theis, L.. (2017). Fast Face-Swap Using Convolutional Neural Networks. In Proceedings of the IEEE International Conference on Computer Vision
“We consider the problem of face swapping in images, where an input identity is transformed into a target identity while preserving pose, facial expression and lighting. to perform this mapping, we use convolutional neural networks trained to capture the appearance of the target identity from an unstructured collection of his/her photographs. this approach is enabled by framing the face swapping problem in terms of style transfer, where the goal is to render an image in the style of another one. building on recent advances in this area, we devise a new loss function that enables the network to produce highly photorealistic results. by combining neural networks with simple pre- and post-processing steps, we aim at making face swap work in real-time with no input from the user.”
Senses, M., & Topal, C.. (2019). Real time face swap based on patch warping. In 27th Signal Processing and Communications Applications Conference, SIU 2019
“Face swap is one of the popular machine vision problems recently. important problems in real-time face swap are the need for high computation power and the decrease in success of applications in case of face pose variations. in this paper, an original face swap algorithm which can work in real time is proposed. after detecting two faces in the input image, 68 landmark points of each face are localized. using these points, a facial model has been formed which separates the faces into 23 quadrilateral planar regions. homographies are calculated between the quadrilaterals formed by the same points on both sides in order not to disrupt the holistic structure of the selected regions. face swap is performed by warping the face patches with calculated homographies. with the proposed method, the calculation resources used in the algorithm have been used efficiently and identity information such as eye, mustache and eyebrow has been preserved. the symmetries of the visible regions are used for the invisible parts in the viewing angles that the camera cannot see the entire face. the success of our study has been tested with different camera resolutions and faces at different angles and qualitative results are given.”
Mahajan, S., Chen, L. J., & Tsai, T. C.. (2017). SwapItUp: A face swap application for privacy protection. In Proceedings – International Conference on Advanced Information Networking and Applications, AINA
“There is a growing concern over the issues related to online privacy due to large availability of high quality images. to tackle the privacy concerns a face swapping application is proposed. there is a library of face images, which is created by downloading images from different sources on internet. for any given image, first of all facial landmarks are detected. the second image is rotated and scaled so that it can properly fit over the input image. to make sure that the new image looks natural, color balance adjustment is done. after that blending of features from the second image onto the input image is done. it is also shown how this system can be used to creating appealing and funny photographs for entertainment purposes. we conclude with a study that shows the high quality of images produced by this system as compared to existing face swap applications and also limitations of this system.”
Sadu, C., & Das, P. K.. (2020). Swapping face images based on augmented facial landmarks and its detection. In IEEE Region 10 Annual International Conference, Proceedings/TENCON
“Facial landmark points that are precisely extracted from the face images improve the performance of many applications in the domains of computer vision and graphics. face swapping is one of such applications. with the availability of sophisticated image editing tools and the use of deep learning models, it is easy to create swapped face images or face swap attacks in images or videos even for non-professionals. face swapping transfers a face from a source to a destination image, while preserving photo realism. it has potential applications in computer games, privacy protection, etc. however, it could also be used for fraudulent purposes. in this paper, we propose an approach to create face swap attacks and detect them from the original images. the augmented 81-facial landmark points are extracted for creating the face swap attacks. the feature descriptors weighted local magnitude patterns (wlmp) and support vector machines (svm) are utilized for the swapped face images detection. the performance of the proposed approach is demonstrated by different types of svm classifiers on a real-world dataset. experimental results show that the proposed system effectively does face swapping and detection with an accuracy of 95%.”
Zhao, Y., Tang, F., Dong, W., Huang, F., & Zhang, X.. (2019). Joint face alignment and segmentation via deep multi-task learning. Multimedia Tools and Applications
“Face alignment and segmentation are challenging problems which have been extensively studied in the field of multimedia. these two tasks are closely related and their learning processes are supposed to benefit each other. hence, we present a joint multi-task learning algorithm for both face alignment and segmentation using deep convolutional neural network (cnn). the proposed multi-task learning approach allows cnn model to simultaneously share visual knowledge between different tasks. with a carefully designed refinement residual module, the cross-layer features are fused in a collaborative manner. to the best of our knowledge, this is the first time that face alignment and segmentation are learned together via deep multi-task learning. our experiments show that learning these two related tasks simultaneously builds a synergy between them, improves the performance of each individual task, and rivals recent approaches. furthermore, we demonstrate the effectiveness of our model in two practical applications: virtual makeup and face swap.”
Chawla, R.. (2019). Deepfakes: How a pervert shook the world. International Journal of Advance Research and Development
Show/hide publication abstract
“Recently a machine learning based open source software (i.e. a free to use the software) tool has made it easy to create hyper-realistic face swaps in videos that leave little to no traces of manipulation, in what is known as ‘deepfake’ videos. scenarios, where these ai manipulated/generated videos, are used for political distress, blackmail or even terrorism are easily envisioned as a near dystopia. this paper explores the various aspects of deepfake videos including its consequences and newly developed innovations in detecting deepfakes.”
Jiang, J., Li, B., Wei, B., Li, G., Liu, C., Huang, W., … Yu, M.. (2021). FakeFilter: A cross-distribution Deepfake detection system with domain adaptation. Journal of Computer Security
“Abuse of face swap techniques poses serious threats to the integrity and authenticity of digital visual media. more alarmingly, fake images or videos created by deep learning technologies, also known as deepfakes, are more realistic, high-quality, and reveal few tampering traces, which attracts great attention in digital multimedia forensics research. to address those threats imposed by deepfakes, previous work attempted to classify real and fake faces by discriminative visual features, which is subjected to various objective conditions such as the angle or posture of a face. differently, some research devises deep neural networks to discriminate deepfakes at the microscopic-level semantics of images, which achieves promising results. nevertheless, such methods show limited success as encountering unseen deepfakes created with different methods from the training sets. therefore, we propose a novel deepfake detection system, named fakefilter, in which we formulate the challenge of unseen deepfake detection into a problem of cross-distribution data classification, and address the issue with a strategy of domain adaptation. by mapping different distributions of deepfakes into similar features in a certain space, the detection system achieves comparable performance on both seen and unseen deepfakes. further evaluation and comparison results indicate that the challenge has been successfully addressed by fakefilter.”
Mirsky, Y., & Lee, W.. (2021). The Creation and Detection of Deepfakes. ACM Computing Surveys
“Generative deep learning algorithms have progressed to a point where it is difficult to tell the difference between what is real and what is fake. in 2018, it was discovered how easy it is to use this technology for unethical and malicious applications, such as the spread of misinformation, impersonation of political leaders, and the defamation of innocent individuals. since then, these ‘deepfakes’have advanced significantly. in this article, we explore the creation and detection of deepfakes and provide an in-depth view as to how these architectures work. the purpose of this survey is to provide the reader with a deeper understanding of (1) how deepfakes are created and detected, (2) the current trends and advancements in this domain, (3) the shortcomings of the current defense solutions, and (4) the areas that require further research and attention.”
Zhang, W., & Zhao, C.. (2019). Exposing Face-Swap Images Based on Deep Learning and ELA Detection. Proceedings
“New developments in artificial intelligence (ai) have significantly improved the quality and efficiency in generating fake face images; for example, the face manipulations by deepfake are so realistic that it is difficult to distinguish their authenticity—either automatically or by humans. in order to enhance the efficiency of distinguishing facial images generated by ai from real facial images, a novel model has been developed based on deep learning and error level analysis (ela) detection, which is related to entropy and information theory, such as cross-entropy loss function in the final softmax layer, normalized mutual information in image preprocessing, and some applications of an encoder based on information theory. due to the limitations of computing resources and production time, the deepfake algorithm can only generate limited resolutions, resulting in two different image compression ratios between the fake face area as the foreground and the original area as the background, which leaves distinctive artifacts. by using the error level analysis detection method, we can detect the presence or absence of different image compression ratios and then use convolution neural network (cnn) to detect whether the image is fake. experiments show that the training efficiency of the cnn model can be significantly improved by using the ela method. and the detection accuracy rate can reach more than 97% based on cnn architecture of this method. compared to the state-of-the-art models, the proposed model has the advantages such as fewer layers, shorter training time, and higher efficiency.”
Yan, S., He, S., Lei, X., Ye, G., & Xie, Z.. (2018). Video Face Swap Based on Autoencoder Generation Network. In ICALIP 2018 – 6th International Conference on Audio, Language and Image Processing
“Video facial swap usually has strong entertainment applications, and it is also applicable for the post-production of films and has great application value. at present, the popular face swap is done manually by the ps software, and the synthetic effect of the automatic face changing technology is not good. in order to make up for the lack of these features, this paper proposes a method of video face swap based on autoencoder generation network. the network learns the mapping relationship between distorted face and original face: the encoder can distinguish and extract facial information, and the decoder can restore face separately. first, the local information of tow face is sent to the network to get the initial model; then, the global information is put into the network for fine-tuning; finally, the face exchange between a and b is completed with face alignment and alpha fusion. the experimental results show that the quality of the method is improved significantly.”
“Deepfakes are a recent off-the-shelf manipulation technique that allows anyone to swap two identities in a single video. in addition to deepfakes, a variety of gan-based face swapping methods have also been published with accompanying code. to counter this emerging threat, we have constructed an extremely large face swap video dataset to enable the training of detection models, and organized the accompanying deepfake detection challenge (dfdc) kaggle competition. importantly, all recorded subjects agreed to participate in and have their likenesses modified during the construction of the face-swapped dataset. the dfdc dataset is by far the largest currently and publicly available face swap video dataset, with over 100,000 total clips sourced from 3,426 paid actors, produced with several deepfake, gan-based, and non-learned methods. in addition to describing the methods used to construct the dataset, we provide a detailed analysis of the top submissions from the kaggle contest. we show although deepfake detection is extremely difficult and still an unsolved problem, a deepfake detection model trained only on the dfdc can generalize to real ‘in-the-wild’ deepfake videos, and such a model can be a valuable analysis tool when analyzing potentially deepfaked videos. training, validation and testing corpuses can be downloaded from ai.facebook.com/datasets/dfdc.”
Wöhler, L., Henningson, J. O., Castillo, S., & Magnor, M.. (2020). PEFS: A Validated Dataset for Perceptual Experiments on Face Swap Portrait Videos. In Communications in Computer and Information Science
“Videos obtained by current face swapping techniques can contain artifacts potentially detectable, yet unobtrusive to human observers. however, the perceptual differences between real and altered videos, as well as properties leading humans to classify a video as manipulated, are still unclear. thus, to support the research on perceived realism and conveyed emotions in face swap videos, this paper introduces a high-resolution dataset providing the community with the necessary sophisticated stimuli. our recording process has been specifically designed to focus on human perception research and entails three scenarios (text-reading, emotion-triggering, and free-speech). we assess the perceived realness of our dataset through a series of experiments. the results indicate that our stimuli are overall convincing, even for long video sequences. furthermore, we partially annotate the dataset with noticeable facial distortions and artifacts reported by participants.”
Bode, L., Lees, D., & Golding, D.. (2021). The Digital Face and Deepfakes on Screen. Convergence
Cole, S.. (2018). We Are Truly Fucked: Everyone Is Making AI-Generated Fake Porn Now. Motherboard
Show/hide publication abstract
“A user-friendly application has resulted in an explosion of convincing face-swap porn.”
Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A., & Ortega-Garcia, J.. (2020). Deepfakes and beyond: A Survey of face manipulation and fake detection. Information Fusion
“The free access to large-scale public databases, together with the fast progress of deep learning techniques, in particular generative adversarial networks, have led to the generation of very realistic fake content with its corresponding implications towards society in this era of fake news. this survey provides a thorough review of techniques for manipulating face images including deepfake methods, and methods to detect such manipulations. in particular, four types of facial manipulation are reviewed: i) entire face synthesis, ii) identity swap (deepfakes), iii) attribute manipulation, and iv) expression swap. for each manipulation group, we provide details regarding manipulation techniques, existing public databases, and key benchmarks for technology evaluation of fake detection methods, including a summary of results from those evaluations. among all the aspects discussed in the survey, we pay special attention to the latest generation of deepfakes, highlighting its improvements and challenges for fake detection. in addition to the survey information, we also discuss open issues and future trends that should be considered to advance in the field.”
Baek, J. Y., Yoo, Y. S., & Bae, S. H.. (2020). Generative Adversarial Ensemble Learning for Face Forensics. IEEE Access
“The recent advance of synthetic image generation and manipulation methods allows us to generate synthetic face images close to real images. on the other hand, the importance of identifying the synthetic face images increases more and more to protect personal privacy from those. although some deep learning-based image forensic methods have been developed recently, it is still challenging to distinguish synthetic images generated by recent image generation and manipulation methods such as the deep fake, face2face, and face swap. to resolve this challenge, we propose a novel generative adversarial ensemble learning method. we train multiple discriminative and generative networks based on the adversarial learning. compared to the conventional adversarial learning, our method is however more focused on improving the discrimination ability rather than image generation one. to this end, we improve the discriminabilty by ensembling outputs from different two discriminators. in addition, we train two generators in order to generate general and hard synthetic images. by ensemble learning of all the generators and discriminators, we improve the discriminators by using the generated synthetic face images, and improve the generators by passing the combined feedback of the discriminators. on the faceforensics benchmark challenge, we thoroughly evaluate our methods by comparing the recent methods. we also provide the ablation study to prove the effectiveness and usefulness of our method.”
Kaur, S., Kumar, P., & Kumaraguru, P.. (2020). Deepfakes: temporal sequential analysis to detect face-swapped video clips using convolutional long short-term memory. Journal of Electronic Imaging
“Deepfake (a bag of ‘deep learning’ and ‘fake’) is a technique for human image synthesis based on artificial intelligence, i.e., to superimpose the existing (source) images or videos onto destination images or videos using neural networks (nns). deepfake enthusiasts have been using nns to produce convincing face swaps. deepfakes are a type of video or image forgery developed to spread misinformation, invade privacy, and mask the truth using advanced technologies such as trained algorithms, deep learning applications, and artificial intelligence. they have become a nuisance to social media users by publishing fake videos created by fusing a celebrity’s face over an explicit video. the impact of deepfakes is alarming, with politicians, senior corporate officers, and world leaders being targeted by nefarious actors. an approach to detect deepfake videos of politicians using temporal sequential frames is proposed. the proposed approach uses the forged video to extract the frames at the first level followed by a deep depth-based convolutional long short-term memory model to identify the fake frames at the second level. also the proposed model is evaluated on our newly collected ground truth dataset of forged videos using source and destination video frames of famous politicians. experimental results demonstrate the effectiveness of our method.”
Gu, S., Bao, J., Yang, H., Chen, D., Wen, F., & Yuan, L.. (2019). Mask-guided portrait editing with conditional gans. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
“Portrait editing is a popular subject in photo manipulation.the generative adversarial network (gan) advances the generating of realistic faces and allows more face editing. in this paper, we argue about three issues in existing techniques: diversity, quality, and controllability for portrait synthesis and editing. to address these issues, we propose a novel end-to-end learning framework that leverages conditional gans guided by provided face masks for generating faces. the framework learns feature embeddings for every face component (e.g., mouth, hair, eye), separately, contributing to better correspondences for image translation, and local face editing. with the mask, our network is available to many applications, like face synthesis driven by mask, face swap+ (including hair in swapping), and local manipulation. it can also boost the performance of face parsing a bit as an option of data augmentation.”
Wen, L., & Xu, D.. (2019). Face Image Manipulation Detection. In IOP Conference Series: Materials Science and Engineering
“This paper proposes a cnn-based (convolutional neural network based) network to detect altered face picture, which can cover the most common face swap methods. the network uses an autoencoder which is pre-trained on the original images to reconstruct the input images. the reconstructed one and the input image are then processed by the srm filter which can extract the noise distribution of images. we then feed the minus result of two processed results into a cnn architecture to predict whether the input image is original or tampered. the model was trained and evaluated in faceforensics dataset and state-of-art face swap method. experimental results demonstrate the effectiveness of our network.”
Guera, D., & Delp, E. J.. (2019). Deepfake Video Detection Using Recurrent Neural Networks. In Proceedings of AVSS 2018 – 2018 15th IEEE International Conference on Advanced Video and Signal-Based Surveillance
“In recent months a machine learning based free software tool has made it easy to create believable face swaps in videos that leaves few traces of manipulation, in what are known as deepfake videos. scenarios where these realistic fake videos are used to create political distress, blackmail someone or fake terrorism events are easily envisioned. this paper proposes a temporal-aware pipeline to automatically detect deepfake videos. our system uses a convolutional neural network (cnn) to extract frame-level features. these features are then used to train a recurrent neural network (rnn) that learns to classify if a video has been subject to manipulation or not. we evaluate our method against a large set of deepfake videos collected from multiple video websites. we show how our system can achieve competitive results in this task while using a simple architecture.”
Hashmi, M. F., Ashish, B. K. K., Keskar, A. G., Bokde, N. D., Yoon, J. H., & Geem, Z. W.. (2020). An Exploratory Analysis on Visual Counterfeits Using Conv-LSTM Hybrid Architecture. IEEE Access
“In recent years, with the advancements in the deep learning realm, it has been easy to create and generate synthetically the face swaps from gans and other tools, which are very realistic, leaving few traces which are unclassifiable by human eyes. these are known as ‘Deepfakes’ and most of them are anchored in video formats. such realistic fake videos and images are used to create a ruckus and affect the quality of public discourse on sensitive issues; defaming one’s profile, political distress, blackmailing and many more fake cyber terrorisms are envisioned. this work proposes a microscopic-typo comparison of video frames. this temporal-detection pipeline compares very minute visual traces on the faces of real and fake frames using convolutional neural network (cnn) and stores the abnormal features for training. a total of 512 facial landmarks were extracted and compared. parameters such as eye-blinking lip-synch; eyebrows movement, and position, are few main deciding factors that classify into real or counterfeit visual data. the recurrent neural network (rnn) pipeline learns based on these features-fed inputs and then evaluates the visual data. the model was trained with the network of videos consisting of their real and fake, collected from multiple websites. the proposed algorithm and designed network set a new benchmark for detecting the visual counterfeits and show how this system can achieve competitive results on any fake generated video or image.”
Nagarajan, A., & Soghrati, S.. (2018). Conforming to interface structured adaptive mesh refinement: 3D algorithm and implementation. Computational Mechanics
“A new non-iterative mesh generation algorithm named conforming to interface structured adaptive mesh refinement (cisamr) is introduced for creating 3d finite element models of problems with complex geometries. cisamr transforms a structured mesh composed of tetrahedral elements into a conforming mesh with low element aspect ratios. the construction of the mesh begins with the structured adaptive mesh refinement of elements in the vicinity of material interfaces. an r-adaptivity algorithm is then employed to relocate selected nodes of nonconforming elements, followed by face-swapping a small fraction of them to eliminate tetrahedrons with high aspect ratios. the final conforming mesh is constructed by sub-tetrahedralizing remaining nonconforming elements, as well as tetrahedrons with hanging nodes. in addition to studying the convergence and analyzing element-wise errors in meshes generated using cisamr, several example problems are presented to show the ability of this method for modeling 3d problems with intricate morphologies.”
Gerstner, E.. (2020). Face/off:” DeepFake” face swaps and privacy laws. Def. Counsel J.
Show/hide publication abstract
“… detect various prohibited content such as copyright violations, it would not be too complicated to add face swap detection as well to the bots which scan each and every post uploaded to their servers.the final potential target of legislation would be the software used to create …”
In this article, we analyze newspaper articles and advertisements mentioning vaccination from 1915 to 1922 and refer to historical studies of vaccination practices and attitudes in the early…