• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Multi-Branch Deepfake Detection Algorithm Based on Fine-Grained Features

    2023-12-12 15:50:08WenkaiQinTianliangLuLuZhangShufanPengandDaWan
    Computers Materials&Continua 2023年10期

    Wenkai Qin,Tianliang Lu,?,Lu Zhang,Shufan Peng and Da Wan

    1School of Information Network Security,People’s Public Security University of China,Beijing,100038,China

    2Department of Investigation,Shandong Police College,Jinan,250200,China

    ABSTRACT With the rapid development of deepfake technology,the authenticity of various types of fake synthetic content is increasing rapidly,which brings potential security threats to people’s daily life and social stability.Currently,most algorithms define deepfake detection as a binary classification problem,i.e.,global features are first extracted using a backbone network and then fed into a binary classifier to discriminate true or false.However,the differences between real and fake samples are often subtle and local,and such global feature-based detection algorithms are not optimal in efficiency and accuracy.To this end,to enhance the extraction of forgery details in deep forgery samples,we propose a multi-branch deepfake detection algorithm based on fine-grained features from the perspective of fine-grained classification.First,to address the critical problem in locating discriminative feature regions in fine-grained classification tasks,we investigate a method for locating multiple different discriminative regions and design a lightweight feature localization module to obtain crucial feature representations by augmenting the most significant parts of the feature map.Second,using information complementation,we introduce a correlation-guided fusion module to enhance the discriminative feature information of different branches.Finally,we use the global attention module in the multi-branch model to improve the cross-dimensional interaction of spatial domain and channel domain information and increase the weights of crucial feature regions and feature channels.We conduct sufficient ablation experiments and comparative experiments.The experimental results show that the algorithm outperforms the detection accuracy and effectiveness on the FaceForensics++and Celeb-DF-v2 datasets compared with the representative detection algorithms in recent years,which can achieve better detection results.

    KEYWORDS Deepfake detection;fine-grained classification;multi-branch;global attention

    1 Introduction

    In recent years,with the continuous deep learning research,deepfake techniques have also made rapid development,especially the generative models such as autoencoder and Generative Adversarial Networks(GAN)[1]have been applied to the technology,which makes it easy to achieve high realism generation and replacement of video faces.This technology which uses deep learning methods to generate high-quality forged videos and images is called“deepfake”[2].People can generate realistic fake face images using deepfake that are difficult to detect with traditional techniques.Compared with conventional forgery methods,deepfake techniques enrich the details of face forgery and greatly enhance the realism of forged faces,bringing a new challenge to detecting and identifying forged content.These forged video images are widely disseminated on the Internet through social media,which not only violate personal privacy and reputation but also mislead social opinion,confuse the public,and even cause political disputes,posing a significant threat to national security[3].Therefore,to cope with the security risks brought by deepfake technology,researchers have carried out a series of explorations in the academic community,and the design of a general and efficient deepfake detection algorithm has become one of the research hotspots[4–8].

    The deep generative model is at the heart of deep face video forgery techniques represented by “deepfake”.The deep generative model can be simply described as a model that uses a deep neural network for data generation to generate new data that is not included in the training dataset.Currently,forgery generation algorithms are divided into two categories.One replaces the face to realize the tampering of identity information,and the other replaces only the expression of the face without changing person’s identity information.However,these generation techniques also have obvious drawbacks,such as the presence of noticeable visual artifacts in the generated forged video and the difficulty of maintaining frame-to-frame consistency between consecutive video frames.This is because the characteristics of its probabilistic model,the face generation process often a certain degree of randomness,which makes the synthesized face video has a large jitter,and the final forgery effect still needs to be further improved.Therefore,with the continuous improvement of the generation technology,the new forgery method focuses on solving the various shortcomings existing in the existing techniques,and the quality of the fake face produced in the face forgery process is also continuously improved,making the generated image or video more realistic.

    Various deepfake detection methods have been proposed by researchers from different research perspectives,some of which can achieve high identification accuracy on open-source forgery datasets.Most of them model deepfake detection as a binary classification problem,and they tend to extract the global features of an image using a backbone network and then feed them into a binary classifier to distinguish between true and false.Depending on the type of feature selection,these detection methods can be classified as temporal feature-based and spatial feature-based.The temporal featurebased detection methods mainly use Recurrent Neural Networks(RNN)to learn the temporal features between forgery video frames.However,these methods rely heavily on the video’s preprocessing and the frame sequence’s length,resulting in poor generalization for cross-library detection.At the same time,it can not meet the detection needs in the face of highly compressed videos and complex video backgrounds due to less attention to the local forgery details within frames.Therefore,the detection methods based on spatial features have become the mainstream in deepfake detection tasks due to their better robustness and higher accuracy.These methods transform video detection into an image detection problem by extracting spatial features using Convolutional Neural Networks(CNN)after randomly selecting videos frames.However,in real scenarios,the differences between real and fake samples often exist in subtle ways,such as changes in the curvature of the lips of a face,folds in the facial skin texture,etc.These differences can be easily ignored by these in-frame global featurebased detection algorithms.These detection algorithms based on intra-frame global features can easily overlook these differences,resulting in the model ignoring many discriminative features in the feature extraction phase,which induces the model to make wrong judgments about the authenticity of the samples.In deep forgery detection,due to the extensive presence of forgery traces in different regions of the sample,researchers need to pay more attention to the influence of interactions between various features to avoid monolithic feature learning.In addition,we examined that the commonly used channel or spatial attention mechanisms can only effectively capture local information and cannot establish the functional dependency between space and channel,making it difficult to fully utilize the global information across dimensions.Therefore,this paper makes the detection model focus on critical features more effectively by incorporating global attention mechanism into the backbone network.

    Therefore,to solve the difficulties and problems in the detection mentioned above,this paper investigates the method of introducing localized fine-grained features in the backbone network inspired by the idea of fine-grained classification.And to enhance the cross-dimensional interactions of attention information,we embed the global attention mechanism in the backbone and design a multi-branch deepfake detection algorithm.Specifically,the improvements in this paper are mainly carried out in the following three aspects:

    1.The recent detection algorithms based on intra-frame features directly extract global features as the source of classification.It is often difficult to focus on the subtle differences between authentic and fake samples.Therefore,to mine these fine-grained features hidden in different regions,we design a feature localization module and use it to transform the backbone into a multi-branch detection model,thus locating multiple discriminative regions and obtaining multi-scale fine-grained information.

    2.In the training phase,we introduce the global attention module [9] in the second branch to calculate the model attention weights.The global attention module can enhance the cross-dimensional interaction of spatial and channel information while increasing the weights of crucial feature regions and channels to assign more appropriate attention weights to the feature maps and improve the accuracy of localization and detection.And we validate the effectiveness of embedding the global attention module in the backbone through various ablation experiments.

    3.To explicitly explore the relationship and calculate the complementary information between the features of different branch regions,we employ the correlation-guided fusion module in the feature integration phase.And then,under its guidance,we fuse the information of different branch features to provide a diverse basis for identifying genuine and forged samples.

    4.In the final integrated decision-making phase,to improve the correlation between standard and forged samples,we formulate a multi-branch loss function,which uses cross-entropy loss as an end-to-end loss function in all three branches.

    The structure of our treatise is organized as follows.Section 2 combs through the existing work and relevant references in the field of deepfake detection in recent years and introduces the sources and limitations that inspired us to make algorithmic improvements.Section 3 is the basic methodology and implementation details of our approach,including the principles and model construction for the improvement module.Section 4 provides the details of the parameter settings,environment configurations,dataset processing,and various results of our experiments,which validate the effectiveness of our model.Section 5 concludes with a summary of contributions and proposes future research directions to advance the field of deep forgery detection.

    2 Related Work

    2.1 Deepfake Detection Based on Deep Learning

    To cope with the impact of deepfake techniques on public security,researchers have proposed various deepfake detection methods,which can be divided into detection based on video intraframe features and detection based on video interframe features.

    2.1.1 Detection Based on Interframe Features

    The detection methods based on interframe timing features to focus on mining the timing features between consecutive frames of the video and classifying the video authenticity by detecting the interframe continuity and integrity.Since deepfake techniques can not forge the video as a whole but tamper with specific regions of the face frame by frame,it is difficult to take into account the sequence of previously forged frames when forging the current frame,which leads to discrepancies in the spatiotemporal distribution of successive frames such as differences in the expressions of the characters,inconsistent spatiotemporal states of the tampered regions,visual artifacts and noise inside successive frames.Such discrepancies can be captured by RNN or other methods based on sequence feature extraction and used to detect actual and forged videos.

    Agarwal et al.[10] encoded face expressions,head movements,and other physiological signals,then used a support vector machine(SVM)for detection and classification.Lima et al.[11]employed spatiotemporal convolution to detect temporal and spatial artifacts of forged videos as well as utilized multiple layers of coding units based on the self-attention mechanism to enable each frame feature to aggregate other frame information,thus enabling effective detection of current mainstream forgery datasets.Cozzolino et al.[12]trained an ID-Reveal model using a self-supervised learning paradigm to mine interframe temporal features through ID networks and perform adversarial learning using 3DMM generative networks to identify the authenticity of face videos.Sun et al.[13] proposed a robust framework LRNet to detect forged videos by modeling accurate geometric features.

    2.1.2 Detection Based on Intraframe Features

    The main idea of the detection methods based on intraframe spatial features is to randomly extract image frames from the video stream,input them into CNN for feature extraction after preprocessing,such as face alignment and face cropping,and finally use the feature information for classification.These approaches first transform the detection object from video to image and use classical models such as ResNet[14],Xception[15],and EfficientNet[16]as backbone networks for classification.

    Currently,most deepfake detection is carried out based on interframe spatial features.Afchar et al.[17] found that the artifacts of deepfake images are often present in the middle layer semantic information and proposed using MesoNet combined with the Inception module to extract image middle layer features for detection.Nguyen et al.[18] designed to use the VGG16 model to extract features and then employ a capsule network to learn more detailed features of the face.Nirkin et al.[19]used three encoders based on the Xception structure to encode the whole image,face region,and background region after preprocessing.Finally,they combined the three feature vectors to achieve the effect of detection and localization.Shang et al.[20]constructed a pixel-region relationship network(PRRNet)and used the pixel relationship module and the region relationship module to detect spatial association and inconsistent forgery traces in images.

    However,as the forged content becomes more and more realistic,the above binary classification models based on global features have poor detection performance.The reason is that the highly distinguishable discrepancies between forgery samples and original samples often exist in subtleties of the image,such as lip curvature changes and skin texture differences,which are often not easily captured by the detection algorithm based on global features,resulting in reduced detection accuracy.Such subtle and located differences are similar to the fine-grained classification.Therefore,to improve the detection based on spatial features,researchers have started to analyze the deepfake detection task expressed as a fine-grained classification problem.

    Zhou et al.[21]designed a detection network based on multi-attention and used bilinear pooling to aggregate low-level texture features and high-level semantic features under the guidance of attention graphs.Guarnera et al.[22]utilized an expectation-maximization algorithm to extract convolutional traces from specific local pixels.Chen et al.[23] used a pixelhop++module to extract features and reduce the dimensionality of local regions separately in multiple regions of face images and finally integrated and classified the information of each region.Liu et al.[24] divided the original image into several blocks of the same size and randomly shuffled pixels within blocks and positions between blocks,forcing the model to extract more discriminative forgery traces.Based on the lack of physiological constraints in the current generative models,Hu et al.[25] proposed to explore whether the number,shape,and relative positions of the reflected glare of forged eyes are consistent for detection.Guo et al.[26] first located the human eyes’oval pupil mask and then used the IoU algorithm to calculate the difference between the fake pupil shape and the real pupil shape to determine the authenticity of the input image.

    In recent years,with the rapid development of Transformers[27],Vision Transformer algorithms[28]that can be used in computer vision have been derived.These algorithms can extract the global association relationship of pixels,and the number of operations required to compute the association between two positions does not increase with distance compared with the CNN.Moreover,the internal self-attention mechanism can generate more interpretable models with stronger modeling ability and detection performance,which have been gradually applied in the study of deepfake detection.

    2.2 Fine-Grained Classification

    Fine-grained classification is a challenging research task in computer vision,which distinguishes different fine-grained categories by capturing local discriminative features [29,30].Compared with traditional classification tasks,fine-grained classification provides a more detailed classification of images.Since the discrepancies between categories are minor,the essence of this classification task is to locate the target and local regions and perform feature extraction and processing to complete the training and detection of the classifier.

    The key to fine-grained classification is how to locate the subtle discrepancies that exist in local regions.The main research directions can be divided into strongly supervised and weakly supervised approaches For different ways of extracting fine-grained features.The strongly supervised approach uses additional information,such as annotation frames and part annotation points,in addition to category labels in classification.Therefore,it requires more manual effort and time consumption.The weakly supervised approach only uses category labels to complete the model’s training.Currently,research in this field mainly focuses on locating discriminative regions in a weakly supervised manner[31–34].In this paper,we define deepfake detection as a particular fine-grained classification problem,and both have the same characteristics in extracting subtle and discriminative features.However,deepfake detection only involves two categories,i.e.,real and fake.

    2.3 Attention Mechanism

    In deep neural networks,the attention mechanism can mainly be divided into channel attention mechanism and hybrid attention mechanism.

    The channel attention mechanism focuses on the correlation between different channels of the feature graph and automatically obtains the importance of each feature channel through network learning,and finally assigns different weight coefficients to each channel to reinforce the critical features and suppress the non-important features.The commonly used are the squeeze-and-excitation(SE)module[35]and the efficient channel attention(ECA)module[36].The SE first compresses the spatial dimension of the graph by global average pooling operation and then extracts the channel attention information using a fully connected layer.The ECA,on the other hand,utilizes a 1 × 1 convolution instead of the fully connected layer to achieve information interaction across channels based on the SE.

    The hybrid attention mechanism is a more integrated attention method that combines the advantages of channel dimension attention and spatial dimension attention.The commonly used are the bottleneck attention module(BAM)[37],the convolutional block attention module(CBAM)[38],and the coordinate attention(CA)module[39].The BAM and CBAM can extract the attention information of both channel and spatial dimensions and fuse different attention information.The CA extracts channel attention information from the width and height directions,combining the position information and channel attention information.In 2021,Liu et al.[9]proposed the global attention module(GAM),which still uses the channel-space attention arrangement order.But they redesigned the channel attention submodule and the space attention submodule.When the GAM is embedded in the backbone model,the model cannot only consider the attention weight information extraction in both channel dimension and spatial dimension but also ensure the cross-dimensional information interaction to improve the detection accuracy further.

    3 Methods

    In this paper,we propose a multi-branch deepfake detection algorithm based on fine-grained features,and the algorithm architecture is shown in Fig.1.

    Figure 1:Architecture of multi-branch deepfake detection algorithm based on fine-grained features

    As aforementioned,the discrepancies between real and fake faces are usually subtle and occur in local regions,which are not easily captured by the models based on global features.Therefore,to collect local features from different regions more effectively,we design a multi-branch detection framework.In this framework,three main components are embedded in the backbone network:(1)We use the feature localization module (FLM) to locate the subtle discrepancy in different regions.(2) We employ the correlation-guided fusion module(CGFM)to aggregate complementary information from different branches.(3) We introduce the global attention module (GAM) to increase the weights of critical regions and channels and enhance the cross-dimensional interaction.Besides,to learn effectively the fine-grained features of the sample,we specially design a multi-branch loss.

    3.1 Feature Localization Module

    To locate forgery traces in different regions,we introduce the feature localization module to obtain significant feature representations by mining the most prominent parts of the feature map and then suppressing them to force the model to explore other potential discriminative features in subsequent branches.Moreover,by embedding the feature localization module into different layers of the backbone network,we construct a multi-branch fine-grained feature extraction model,which can obtain feature representations of multiple regions.The specific structure of the feature localization module is shown in Fig.2.

    Figure 2:Structure of feature localization module

    In the feature localization module,the feature map from a particular layer is first defined asF∈RC×W×H,whereC,W,Hdenotes the number of channels,width,and height of the feature map,respectively.In the feature mapF,it is uniformly partitioned intokstriped sections along the width dimension,and each striped section is represented asF(i)∈RC×(W/k)×H,i∈[1,k].We defineA(i)as a high attention activation map for each part as shown in Eq.(1):

    where the nonlinear activation function Reluis employed to remove the negative activation parts.φrepresents the 1×1 convolution,which is used to calculate the importance of each striped section and share the parameters among different parts.In the activation mapA(i),to measure the contribution of each striped section to the detection,we use Eq.(2)to gain its average value:

    whereGAPdenotes the global average pooling.biis used as an importance factor,and its value reflects the contribution of the striped sections.If thebivalue of a part is larger,it indicates that it is more likely to belong to the critical regional features.The normalization ofbiis performed by theSoftmax function,which is expressed in Eq.(3):

    whereis the normalized importance factor,which are combined to obtain the feature vectorB′=The vector has a global sensory field,which can obtain the statistical information of the feature mapFon different striped sections.The most significant region on the feature mapFcan be represented using the feature vector,and the enhanced featureFbis obtained by feature fusion.The feature fusion process is shown in Eq.(4):

    where the magnitude of the weight parameterαreflects the proportion of the most significant region in the feature fusion,and the degree of feature enhancement can be effectively controlled by adjustingα.?denotes element-by-element multiplication.Finally,as shown in Eq.(5),through applying a convolution layerφ′onFb,we can obtain a specific high-concentration featureFpas follows:

    In order for the subsequent modules to continue mining discriminative features in other regions,the suppression factorsican be obtained by Eq.(6),which suppress the high attention regions on the current branch and force the subsequent modules of the model to focus more on other regions of the feature mapF,and then learn different scale features.

    where the magnitude of the weight parameterβreflects the degree of suppression of the most significant region in the feature mapF.The combination of the suppression factorssiyields the feature vectorS=(s1,s2,...,sk)T.In the current branch,the most significant region of the feature mapFis suppressed byS,as shown in Eq.(7),and we can obtain the suppressed potential feature mapFs.

    The feature mapFsstill contains many discriminative features in other regions except the suppressed high-concentration features.Therefore,to locate these discriminative features,Fsis continued to be fed into the feature localization module of the subsequent branches to capture the multi-scale discriminative features more effectively.

    3.2 Correlation-Guided Fusion Module

    Although the feature localization module can locate the highly focused part of the sample,the relationship among the features of each branch can be easily ignored because the branch structures are relatively independent of each other.To extract a more robust feature representation,we introduce a correlation-guided fusion module for the feature information learned from different branches.The correlation-guided fusion module consists of several feature complementary modules(FCM),which first explore the complementary information among the features of each branch to avoid ignoring other discriminative features due to the focus on individual local features,and then enhance the feature representation of the current branch through feature fusion guided by the complementary information.The structure of the feature complementation module is shown in Fig.3.

    In the feature complementation module,the regional feature mapsFp1output from any two branches are defined aswhereCdenotes the number of channels,andW1×H1andW2×H2denote the spatial size of each feature map,respectively.To explore the relationship among the features of different branches,we represent the feature vector of each spatial position along the channel dimension of the feature mapFas a pixel,and characterizes the correlation among the feature maps by the correlation among pixels,as shown in Eq.(8):

    wherepixcel(F,i)denotes the feature vector of theithregion along the channel dimension on the feature mapF,andFj,idenotes theithregion located on thejthchannel of the feature mapF.After obtaining the feature vectors along the channel dimension for each region onFp1andFp2,respectively,we calculate the vector inner product to reflect the similarity among the feature image elements more concisely,as shown in Eqs.(9)and(10):

    where the matrixMrepresents the similarity matrix between feature mapsFp1andFp2,and the elementMi,jrepresents the similarity between theithpixel of feature mapFp1and thejthpixel of feature mapFp2.If theMi,jbetween two pixels is smaller,the less similar and more complementary these two pixels are to each other.Therefore,based on the negative correlation between similarity and complementarity,to reflect the complementarity more intuitively,we define the complementarity matrixCas the negative matrix of the similarity matrixM,and normalize the columns and rows of the complementarity matrix through Eqs.(11)and(12)as follows:

    wheredenotes the complementary information extracted fromFpjagainstFpi,which is used as the basis for enhancingFpiin the subsequent aggregation of branch features.To further explore the complementary relationship between the two feature maps,the mining of complementary information operation is transformed as pixel-specific through Eqs.(15)and(16)as follows:

    where each pixel inis enhanced by all pixels inFp2,and each pixel inis enhanced by all pixels inFp1.If the complementarity among the pixels inFp1andFp2is stronger,it indicates that the pixels in one of the feature maps contribute more to the pixels in the other feature map.Then each pixel in both can mine semantic complementary information from different regions of the other feature.Therefore,the complementary information of regional features in each branch can be obtained according to Eq.(17):

    whereXpjpican be obtained by Eqs.(9),(13)and(14).In addition,as shown in Fig.3,for the regional feature informationFp1andFp2in any two branches,the feature complementation module can calculate both the complementary informationandat the same time.Combining the complementary information from different branches,we can enhance the regional feature information of the current branch through feature fusion to prompt the model focus on different fine-grained features,as Eq.(18)shown in:

    whereYpidenotes the enhanced regional features,which contain discriminative information at different scales.The magnitude of the weight parameterγreflects the proportion of complementary information from different branch features in the feature fusion process.

    Figure 3:Structure of feature complementary module

    3.3 Global Attention Module

    In this paper,we choose the Xception as the backbone model,which has a good effect in image classification,and design a multi-branch structure.We find that the suppressed potential featureFsgenerated in the previous branch still not only contains a large number of potential discriminative features but also includes some background and noise information.These interferences can impact the localization of subsequent potential discriminative features and eventually lead to a drift between feature localization and the target region.

    Therefore,this paper introduces the global attention module in the Middle Flow of Xception to build a more robust network structure by enhancing the model’s ability to extract key features of forged images.The global attention module utilizes spatial and channel information at different scales on the feature map,increases the weights of crucial feature regions and feature channels,and enhances the global cross-dimensional interaction of spatial and channel information so that the network can obtain more effective feature representation capability.The structure of GAM is shown in Fig.4.

    Figure 4:Structure of global attention module

    The GAM comprises two submodules: the channel attention module (CAM) and the spatial attention module (SAM).Similar to the CBAM,the GAM belongs to the hybrid attention module that incorporates channel and spatial attention.

    In the CAM,the input feature map is given asF1∈RC×W×H,whereC,W,Hdenote the number of channels,width,and height of the feature map,respectively.To obtain global cross-dimensional interaction information,the channel attention module uses a 3D permutation to retain information across three dimensions and get the feature map∈RW×H×C.Thenis input to a two-layer multilayer perceptron(MLP),which first compresses the channels asC/rand then reduces them to amplify the cross-dimensional dependence of channel and space.Finally,the feature map dimensions are reduced using another 3D permutation,and the result is element-by-element multiplied by the original featureF1after the activation functionsigmiodto obtain the outputF2∈RC×W×H.The calculation process is shown in Eqs.(19)and(20):

    wheresigmoiddenotes the activation function.The parametersW0∈RC×C/randW1∈RC/r×Cdenote the parameter-sharing multilayer perceptron weights.rdenotes the compression rate of the hidden layer,andMc(F1)denotes the output channel feature vector.The structure of the CAM is shown in Fig.5.

    Figure 5:Structure of channel attention module

    In the SAM,the input feature map is the outputF2∈RC×W×Hof the channel attention module,whereC,W,Hdenote the number of feature map channels,width,and height,respectively.To focus on spatial information more precisely,the spatial attention module employs two convolutional layers of 7×7 for compression of channel dimension and fusion of spatial information.Meanwhile,since max pooling reduces the amount of information,the spatial attention module removes the pooling layer to further preserve the feature information.Besides,the spatial attention module uses grouped convolution on the channel dimension to reduce the negative effect of the significant increase in the number of parameters due to removing of the pooling layer.Finally,the result is element-by-element multiplied with the original featureF2after the activation functionsigmoidto obtain the outputF3∈RC×W×Hof the spatial attention module,i.e.,the final output of the global attention module.The calculation process is shown in Eqs.(21)and(22):

    wheresigmoiddenotes the activation function.Conv7×7denotes the convolutional layer using a convolution kernel of size 7×7.The compression rate of the channel between the two convolution layers is represented asr′.Ms(F2)denotes the output spatial feature vector.The structure of the SAM is shown in Fig.6.

    Figure 6:Structure of spatial attention module

    In this paper,we embed the GAM into the Block of Xception as a kind of data enhancement module,which makes it possible to modify the Block without affecting the branch structure,as shown in Fig.7.The model reconstructs the extracted feature maps by embedding the global attention mechanism.It can improve the attention of subsequent branches on essential features and eliminate the interference of background information as much as possible.In addition,the global attention module achieves cross-dimension interaction without reducing dimensionality,which avoids information loss while reducing the number of parameters and computational cost.

    Figure 7:Block structure embedded in GAM

    3.4 Multi-Branch Loss Function

    To assess the validity of the model classification results,this paper designs a multi-branch loss function that uses cross-entropy loss as an end-to-end loss function in all three branches.It is expressed in Eqs.(23)–(25)as:

    whereLidenotes the loss of theithbranch.yiis the true label of the input sample which is represented by the one-hot vector.clsidenotes the classifier of theithbranch.Pidenotes the classification probability calculated by theithbranch.Ltotaldenotes the final loss.

    In the training phase,the final lossLtotalis obtained through calculating the classification loss of the enhanced regional featuresYpiin each branch.We use the AdamW algorithm to optimize the final loss and accelerate the convergence of the model.The three branches together finally make the model learn the various fine-grained features of the forged sample.In the testing phase,the model takes the average prediction scores of all branches as the final prediction result.

    4 Experimental Results and Analysis

    4.1 Experiment Environment

    The experimental platform in this paper is a 64-bit Linux operating system with Ubuntu 18.04 LTS,an NVIDIA GeForce RTX 2080Ti graphics card with 11 GB of video memory,an Intel(R)Xeon(R)CPU E5-2678 v3@2.50 GHz,and 62 GB of RAM.The experimental code is implemented in the Pytorch deep learning framework with version 1.11.0,Anaconda version 4.12.0,and Python version 3.8.

    4.2 Introductions of Datasets

    In this paper,we select two mainstream deepfake datasets for experiments,i.e.,FaceForensics++[40]and Celeb-DF[41].Some samples of datasets are shown in Fig.8.

    Figure 8:Samples of the three datasets

    The FaceForensics++(FF++) dataset is one of the most significant and widely used datasets.The dataset captures 1000 short videos containing unobstructed faces from YouTube and ensures that the faces are contained in consecutive frames.Then it is manipulated by five forgery techniques(DeepFakes,Face2Face,FaceSwap,Neural Textures,and FaceShifer)to generate a total of 5000 forged videos with three different compression ratios(lossless compression:Raw,high-quality compression:c23,and low-quality compression: c40).Among them,DeepFakes is based on an autoencoder for face replacement;FaceSwap is a graphics-based approach.To verify the effectiveness of the proposed algorithm for the deepfake detection task,we select the DeepFakes and FaceSwap datasets with a compression rate of c23 in this paper.

    The Celeb-DF dataset captures 590 real videos of 59 celebrities from YouTube,which consider people of different genders,ages,and races.It generates 5639 forged videos in MPEG4.0 format with an average length of 13 s using a single DeepFakes approach.Since this dataset can fuse the boundary between the forged and original region by enhancing the face resolution and establishing a color conversion algorithm for the faces in the forged and original videos,it dramatically improves the quality of the forged data.The Celeb-DF dataset is currently recognized as a high-quality deepfake dataset.In this paper,we use the extended Celeb-DF-v2 dataset for our experiments.

    4.3 Experiment Setting

    Since the forgery regions are primarily concentrated in the facial regions,to make the model focus more on the features of the forgery traces,the experiments in this paper firstly intercept 30 frames at equal intervals for each video in the data preprocessing stage.Second,we use the face detection algorithm RetinaFace to locate the critical points of the five facial features in each frame to determine the facial rectangle.After the face alignment,the images are cropped to 320 × 320 input samples.Then,we employ the sharpening operation and Gaussian blur on the samples to refine the images and reduce image noise.We standardize and normalize the images to enhance the model’s generalization ability.Finally,the datasets are divided into train-sets and test-sets in the ratio of 7:3,and the division results are shown in Table 1.

    Table 1:Division results of datasets

    In addition,the specific settings of other hyperparameters are shown in Table 2.

    Table 2:Hyperparameters setting

    4.4 Evaluation Index

    In this paper,we use the Accuracy(Acc)and Area under Receiver Operating Characteristic(ROC)Curve(AUC)to evaluate the model effect comprehensively.

    Acc is used to describe the classification accuracy of the classifier for genuine and fake samples,and a larger value of Acc indicates a higher correct classification rate of the model.AUC is defined as the probability that the predicted probability value of getting real samples is greater than the probability of fake samples,and a larger value of AUC indicates the better performance of the detection model.The calculation formula of Acc and AUC is shown in(26)and(27):

    whereTPdenotes the real face image predicted as true by the model.TNdenotes the fake face image predicted as false by the model.FNdenotes the real face image predicted as false by the model.FPdenotes the fake face image predicted as true by the model.Andndenotes the total number of positive and negative samples.

    4.5 Analysis of Experimental Results

    4.5.1 The Ablation Experiment of Each Improved Strategy in Terms of Performance Gain

    In this experiment,we use the Xception as the baseline model and design a set of ablation experiments on the DeepFakes dataset and FaceSwap dataset to verify the performance gains brought by the feature localization module(FLM),correlation-guided fusion module(CGFM),global attention module(GAM)and their combinations to the model detection.The evaluation metrics include Acc value and AUC value.The experimental results are shown in Table 3 and Fig.9.

    Table 3:The classification effect performance gain generated by each improved strategy

    Figure 9:The performance gains brought by different improvement strategies on two datasets

    Based on Xception,when the FLM is used only,the detection accuracy of the model is improved by 0.94%and 0.67%on the two datasets,respectively,proving FLM’s effectiveness.After embedding the FLM in the baseline model,the Xception is transformed into a multi-branch structure,allowing the model to focus on the subtle discrepancies among different regions of the real and fake images so that more potential feature information can be mined,providing more possible bases for subsequent classification.

    When the FLM and CGFM are used together,the detection accuracy of the model is improved by 0.28%and 0.39%,respectively,compared with using the FLM alone.Furthermore,the AUC and loss are also optimized.It indicates that the model’s performance can be further improved by using the CGFM and FLM together.The reason is that the submodule FCM in CGFM can effectively explore the complementary information between the features of two branches and enhance the regional features in each branch through feature fusion guided by the complementary information to make them more diverse and discriminative.

    After introducing the GAM,compared with the baseline model,using the GAM alone leads to 0.76% and 0.45% improvement in detection accuracy on the two datasets,respectively.Compared with using FLM and CGFM,the simultaneous introduction of the three improvement strategies leads to 0.64% and 0.32% improvement in model detection accuracy,respectively.The experimental comparison results show that the global attention module enables the model to better represent critical information through the global cross-dimensional interaction.At the same time,it further demonstrates that the combination of the three improvement strategies can obtain the best detection results,which verifies the effectiveness of the proposed method in deepfake detection in this paper.

    4.5.2 The Ablation Experiment of Global Attention Module Embedding

    To explore the best embedding position of the global attention module in the multi-branch model,this paper designs seven types of attention module embedding methods and inserts them into the Blocks of different branches,respectively.Acc and AUC are used for the experimental indexes.The specific embedding methods and detection results are shown in Table 4 and Fig.10.

    Table 4:The classification effect performance gain generated by each embedding method

    Figure 10:The classification effect and the number of model parameter change on two datasets

    According to the experimental setting,the results in the table can be classified as unembedded GAM,single-branch embedded GAM,and multi-branch embedded GAM.Method (1) does not embed GAM.Methods (2) (3) (4) embed GAM in different branches,respectively.Methods (5) (6)(7)(8)belong to multi-branch embedded GAM.

    The model achieved the highest Acc and AUC values on both datasets when the GAM is embedded in the second branch only.Compared with method (1),embedding GAM in the second branch increased the Acc values by 0.61%and 0.64%,and embedding GAM in the third branch increased the Acc values by 0.2%and 0.39%.However,there is a slight decrease in the model detection accuracy after embedding the GAM in the first branch.In the first branch,the model still mines the global features of the sample,which contains both forgery information and a lot of background information.Therefore,embedding the GAM in the first branch tends to make the model focus on irrelevant background information,which causes bias in detection.

    At the same time,compared with the case of method(1),the model detection accuracy in the case of embedding GAM in multiple branches does not improve significantly with the increase of attention module numbers.Therefore,considering model detection accuracy and the number of parameters brought by the introduction of the attention modules,in this paper,we choose to insert the GAM in the Block of the second branch.

    4.5.3 The Comparative Experiment of Different Attention Modules

    To further verify the effectiveness of GAM for model performance improvement in the deepfake detection task,we replace the GAM embedded in the backbone network with the SE[29],ECA[30],BAM[31],CBAM[32],and CA[33].Meanwhile,the improved model with the FLM and CGFM is defined as the baseline to compare the detection performance of the model after embedding different attention modules on two kinds of datasets.Experimental metrics use Acc and AUC.The experimental results are shown in Table 5 and Fig.11.

    According to the comparison results in Table 5,it can be found that the model performance does not improve significantly and even decreases after embedding the SE and the ECA in the branches,respectively.The analysis of the reason for this phenomenon may be related to the lack of spatial information.In extracting channel attention weights,the feature map’s spatial dimension is compressed,which makes it easy to ignore the spatial information in the process of feature fusion among channels,leading to a decrease in the model detection accuracy.In addition,although the CA injects the position information into the channel attention,it focuses mainly on the channel information of images and cannot capture spatially long-distance dependencies.

    When using the BAM and CBAM,the final results are still unsatisfactory despite the slight improvement in model performance compared with the baseline model.Due to the lack of global cross-dimensional interaction of spatial and channel information,the network cannot obtain a more accurate feature representation capability.

    Figure 11:The classification effect of using different attention modules on two datasets

    After embedding the GAM,the detection performance of the model is significantly improved,indicating that compared with other mainstream attention modules,the GAM provides more effective attention guidance in deepfake detection tasks,which is conducive to improving model detection accuracy.

    4.5.4 The Comparative Experiment with Other Algorithms

    To improve the effectiveness of the model in this paper for the deepfake detection task,nine mainstream deepfake detection models are selected for comparison in this experiment,and the evaluation metrics include Acc and AUC.The comparison results are shown in Table 6 and Fig.12.

    Table 6:Comparison with other models

    Figure 12:Comparison with other models

    Based on the comparison results,it can be seen that the proposed algorithm in this paper has significant improvements in detection performance compared with the mainstream detection models.On the DeepFakes and FaceSwap datasets,the proposed algorithm achieves the best results regarding Acc and AUC values,with 98.67%and 98.55%for Acc and 99.80%and 99.72%for AUC,respectively.Meanwhile,to test the ability of the algorithm to deal with samples with more complex forgery degree,the Celeb-DF-v2 dataset with higher forgery quality is selected experimentally,and the Acc value of 98.37%and the AUC value of 99.58%are obtained,which have better improvements in the detection accuracy and effect than most current algorithms.

    In summary,the detection algorithm proposed in this paper is more accurate in detecting forged samples,which is mainly due to the multi-branch structure that enables the model to mine the finegrained feature information in different regions of the samples and fuses the features among different branches through the correlation-guided fusion module,to learn more discriminative fine-grained features.

    4.5.5 Visualization of Results

    To represent the gain effect of the improvement module on the multi-branch network more intuitively,we use the Grad-CAM graph[42]to visualize the regions of the detection model in different branches focusing on samples,as shown in Fig.13.

    The Grad-CAM is a deep network visualization method based on gradient localization.Specifically,the activation map is obtained as a cumulative weighted average of activation values across channel dimensions given a feature map.The red part in Fig.8 represents the region of interest for the model.For the forged samples chosen from the three datasets in the experiment,the activation maps in the first column to the third column correspond to the attention maps drawn from different branches of the model,respectively,and the three rows represent the changes in the activation maps in the case of different improvement methods.

    Figure 13:Visualization of the forged region

    By comparing the discrepancies among the proposed different improvement methods in the class activation attention graph,it can be found that: (1) After the introduction of FLM,the model can locate different regions of the sample and starts to mine feature information from these fine-grained regions.(2)After the introduction of CGFM,the feature information learned in all branches is focused on the diverse regional information mined by different branches.(3)After the introduction of GAM,different weights are assigned to different channels and regions,improving the feature map’s multiscale feature representation capability.The visualization experiments demonstrate the contribution of FLM,CGFM,and GAM to the model in the deepfake detection task,which gives the model better detection capability.

    5 Conclusions

    In this paper,we express the deepfake detection task as the fine-grained image classification problem and propose a multi-branch deepfake detection algorithm based on fine-grained features.The algorithm constructs a multi-branch network structure that can focus on the subtleties of different sample regions to learn multi-scale detailed features,which effectively solves the shortage of detection accuracy due to coarse-grained features.We introduce the feature localization module and correlationguided fusion module to complete the forgery traces’localization and detection.In addition,embedding the global attention module in the backbone network enhances the cross-dimensional interaction of spatial and channel information and reduces the influence of irrelevant background regions in the sample.These improved strategies provide a new idea for deepfake detection.

    In terms of detection accuracy,the proposed algorithm in this paper investigates the effectiveness of the improved strategies through a variety of ablation experiments,which has been confirmed to be more comprehensive and accurate in the detection task.

    In future work,we can continue to conduct research from the following two aspects:(1)First,we should explore more efficient fine-grained feature extraction methods.(2) Second,we research how to construct richer deepfake samples to improve the cross-library testing capability of fine-grained features for deepfake detection.

    Acknowledgement:All the face images used in this paper are from publicly available deepfake datasets,including FaceForensics++and Celeb-DF-V2.

    Funding Statement:This work was supported by the 2023 Open Project of Key Laboratory of Ministry of Public Security for Artificial Intelligence Security(RGZNAQ-2304)and the Fundamental Research Funds for the Central Universities of PPSUC(2023JKF01ZK08).

    Author Contributions:Study conception and design,W.K.Qin and T.L.Lu;methodology,W.K.Qin;validation,L.Zhang,S.F.Peng and D.Wan;formal analysis,W.K.Qin and T.L.Lu;investigation,W.K.Qin and S.F.Peng;data collection,W.K.Qin and T.L.Lu;draft manuscript preparation,W.K.Qin;writing review and editing,W.K.Qin and L.Zhang;visualization,W.K.Qin and S.F.Peng;supervision,T.L.Lu and L.Zhang.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国产亚洲精品久久久久久毛片| 亚洲国产高清在线一区二区三| 日本色播在线视频| 99久久中文字幕三级久久日本| 男女那种视频在线观看| 久久久久久久久久黄片| 亚洲成人久久性| 日韩欧美国产在线观看| 狂野欧美激情性xxxx在线观看| 亚洲av中文av极速乱| 精品国内亚洲2022精品成人| 亚洲av第一区精品v没综合| 国产精品一区二区性色av| 99热精品在线国产| 国产私拍福利视频在线观看| 中文字幕av成人在线电影| 欧美日韩综合久久久久久| 久久久欧美国产精品| 乱码一卡2卡4卡精品| 性欧美人与动物交配| 能在线免费观看的黄片| 禁无遮挡网站| 亚洲av第一区精品v没综合| 99在线人妻在线中文字幕| 又黄又爽又刺激的免费视频.| 日本一二三区视频观看| 亚洲国产欧美人成| 精品一区二区三区人妻视频| 舔av片在线| 国产av不卡久久| 亚洲成人久久爱视频| 搡老妇女老女人老熟妇| 男女啪啪激烈高潮av片| 欧美成人免费av一区二区三区| 九九热线精品视视频播放| 欧美日韩综合久久久久久| 热99在线观看视频| 不卡视频在线观看欧美| 免费看av在线观看网站| 日韩 亚洲 欧美在线| 国产一区二区亚洲精品在线观看| 99视频精品全部免费 在线| 一级毛片我不卡| 三级国产精品欧美在线观看| 日韩强制内射视频| 精品久久久噜噜| 久久中文看片网| 一级毛片aaaaaa免费看小| 麻豆乱淫一区二区| 美女内射精品一级片tv| 亚洲国产欧美人成| 99久久精品热视频| 国产av在哪里看| 精品一区二区三区人妻视频| 国产黄片美女视频| 欧美三级亚洲精品| 国产三级在线视频| 久久婷婷人人爽人人干人人爱| av在线观看视频网站免费| 我的女老师完整版在线观看| 欧美色视频一区免费| 日韩欧美国产在线观看| av.在线天堂| 久久久久久伊人网av| 亚洲精品影视一区二区三区av| 中文字幕制服av| 免费看光身美女| 国产亚洲5aaaaa淫片| 国产精品三级大全| 国产亚洲欧美98| 精品少妇黑人巨大在线播放 | 亚洲最大成人中文| 草草在线视频免费看| 亚洲欧美日韩卡通动漫| 精品日产1卡2卡| 成人高潮视频无遮挡免费网站| 亚洲成人精品中文字幕电影| 91久久精品国产一区二区成人| 亚洲欧美精品综合久久99| 亚洲性久久影院| 自拍偷自拍亚洲精品老妇| 欧美一区二区亚洲| 亚洲精品456在线播放app| 亚洲欧美日韩东京热| 深夜a级毛片| 日日干狠狠操夜夜爽| 如何舔出高潮| 日本黄色视频三级网站网址| 国产一区二区在线av高清观看| 在线免费十八禁| 99精品在免费线老司机午夜| 12—13女人毛片做爰片一| 精品一区二区免费观看| 精品久久久久久久末码| 欧美一级a爱片免费观看看| 日本-黄色视频高清免费观看| 爱豆传媒免费全集在线观看| 日日撸夜夜添| 天美传媒精品一区二区| 高清毛片免费看| 国产精品一区二区在线观看99 | 老熟妇乱子伦视频在线观看| 99久久中文字幕三级久久日本| 国产精华一区二区三区| 亚洲丝袜综合中文字幕| 亚洲自拍偷在线| 日韩成人av中文字幕在线观看| 国产v大片淫在线免费观看| 成人高潮视频无遮挡免费网站| 久久精品夜夜夜夜夜久久蜜豆| 日日干狠狠操夜夜爽| 日本三级黄在线观看| 国产高清不卡午夜福利| 亚洲精品粉嫩美女一区| 又爽又黄无遮挡网站| 如何舔出高潮| 蜜桃久久精品国产亚洲av| 日本在线视频免费播放| 一卡2卡三卡四卡精品乱码亚洲| 国产精品永久免费网站| 久久久久久久亚洲中文字幕| 老女人水多毛片| a级一级毛片免费在线观看| 欧美激情久久久久久爽电影| 不卡视频在线观看欧美| 高清午夜精品一区二区三区 | 男插女下体视频免费在线播放| 成人av在线播放网站| 国产一级毛片在线| 欧美高清成人免费视频www| 国产精品电影一区二区三区| 在线国产一区二区在线| 成年av动漫网址| 日本黄色片子视频| 欧美另类亚洲清纯唯美| 国产精品永久免费网站| 婷婷亚洲欧美| 亚洲欧美精品自产自拍| 亚洲精品成人久久久久久| 亚洲综合色惰| 一级毛片久久久久久久久女| 日韩亚洲欧美综合| 人妻久久中文字幕网| 中文字幕av在线有码专区| 美女脱内裤让男人舔精品视频 | 在线观看美女被高潮喷水网站| 久久久久网色| 日本三级黄在线观看| 日韩,欧美,国产一区二区三区 | 免费av不卡在线播放| 日产精品乱码卡一卡2卡三| 99久久中文字幕三级久久日本| 国产一区二区三区在线臀色熟女| 久久久国产成人精品二区| 亚洲国产精品合色在线| 国产av不卡久久| 久久午夜福利片| 十八禁国产超污无遮挡网站| 天堂中文最新版在线下载 | 亚洲欧美日韩高清在线视频| 欧美又色又爽又黄视频| 嫩草影院新地址| 久久久久久久久久久丰满| 日韩国内少妇激情av| 亚洲精品日韩av片在线观看| 午夜精品在线福利| 18+在线观看网站| 免费观看a级毛片全部| 美女内射精品一级片tv| 尾随美女入室| 亚洲一级一片aⅴ在线观看| 成人av在线播放网站| 熟妇人妻久久中文字幕3abv| 99热网站在线观看| 欧美色欧美亚洲另类二区| 岛国毛片在线播放| 亚洲av成人av| 日韩高清综合在线| 久久精品国产亚洲av天美| 亚洲中文字幕日韩| 少妇裸体淫交视频免费看高清| 国产久久久一区二区三区| 身体一侧抽搐| 国产黄片视频在线免费观看| 亚洲av成人精品一区久久| av天堂在线播放| 天天躁日日操中文字幕| 国产午夜精品论理片| 亚洲综合色惰| 国产亚洲欧美98| 精品不卡国产一区二区三区| 嫩草影院新地址| 国产精品麻豆人妻色哟哟久久 | 我的老师免费观看完整版| 青春草亚洲视频在线观看| 九九爱精品视频在线观看| 国产一级毛片七仙女欲春2| 国产精品女同一区二区软件| 日韩视频在线欧美| 少妇丰满av| 精品欧美国产一区二区三| 欧美激情国产日韩精品一区| 国产一区二区亚洲精品在线观看| 中文字幕人妻熟人妻熟丝袜美| 在线观看av片永久免费下载| 日韩av在线大香蕉| 大又大粗又爽又黄少妇毛片口| 国产精品久久久久久亚洲av鲁大| 亚洲精品国产av成人精品| 日韩欧美在线乱码| 国产成人精品婷婷| 99久久人妻综合| 日韩人妻高清精品专区| 午夜老司机福利剧场| 小蜜桃在线观看免费完整版高清| 欧美高清性xxxxhd video| 色5月婷婷丁香| 亚洲av.av天堂| 国产视频内射| 日本-黄色视频高清免费观看| 欧美日韩乱码在线| 联通29元200g的流量卡| 日韩制服骚丝袜av| 最近2019中文字幕mv第一页| 插逼视频在线观看| 亚洲av熟女| 大又大粗又爽又黄少妇毛片口| 黄片wwwwww| 精品一区二区三区视频在线| 欧美最新免费一区二区三区| 我要搜黄色片| 国产精品野战在线观看| 99久久精品热视频| 欧美色欧美亚洲另类二区| 熟女电影av网| 有码 亚洲区| 精品久久久久久久末码| 免费观看精品视频网站| 亚洲精品成人久久久久久| 变态另类丝袜制服| 夜夜夜夜夜久久久久| 一个人看视频在线观看www免费| 在线免费观看的www视频| 婷婷色av中文字幕| 一进一出抽搐gif免费好疼| 亚洲aⅴ乱码一区二区在线播放| 免费一级毛片在线播放高清视频| 久久精品国产清高在天天线| 啦啦啦啦在线视频资源| 亚洲精品日韩在线中文字幕 | 国产精品一区二区性色av| 99国产精品一区二区蜜桃av| 18+在线观看网站| 免费电影在线观看免费观看| 国产91av在线免费观看| av免费观看日本| 啦啦啦啦在线视频资源| 国语自产精品视频在线第100页| 激情 狠狠 欧美| 国产成人a区在线观看| 看非洲黑人一级黄片| 精品人妻熟女av久视频| 身体一侧抽搐| 亚洲无线在线观看| 国产精品伦人一区二区| 亚洲中文字幕日韩| 一区二区三区免费毛片| 久久久久久久久久久免费av| 国产黄色小视频在线观看| 久久久久久久亚洲中文字幕| 亚洲av.av天堂| 99久国产av精品| 亚洲精品自拍成人| 国产精品嫩草影院av在线观看| 少妇高潮的动态图| 亚洲av一区综合| 国产成人a区在线观看| 成人国产麻豆网| 国产日本99.免费观看| 久久精品91蜜桃| 一级毛片久久久久久久久女| 亚洲人成网站在线播| 看免费成人av毛片| 国产一区亚洲一区在线观看| 如何舔出高潮| 91av网一区二区| 国产精品美女特级片免费视频播放器| 亚洲经典国产精华液单| av在线播放精品| 亚洲人成网站在线观看播放| 久久人人爽人人爽人人片va| 国产综合懂色| 亚洲欧美精品专区久久| 亚洲成人久久爱视频| 欧美三级亚洲精品| 禁无遮挡网站| 你懂的网址亚洲精品在线观看 | 成人亚洲精品av一区二区| 12—13女人毛片做爰片一| 一本久久精品| 在线天堂最新版资源| 一个人观看的视频www高清免费观看| 成人午夜精彩视频在线观看| 又粗又硬又长又爽又黄的视频 | 一区福利在线观看| 精品99又大又爽又粗少妇毛片| 亚洲精品久久久久久婷婷小说 | 国产一级毛片在线| 91av网一区二区| 青春草国产在线视频 | 久久九九热精品免费| 99久久九九国产精品国产免费| 亚洲av一区综合| 97超碰精品成人国产| 亚洲一级一片aⅴ在线观看| 日本在线视频免费播放| 男女下面进入的视频免费午夜| 欧美丝袜亚洲另类| 91久久精品电影网| 久久九九热精品免费| 国产精品1区2区在线观看.| 亚洲av熟女| 成人永久免费在线观看视频| 亚洲精品国产成人久久av| 精品久久久久久久久久久久久| 深夜a级毛片| 日韩亚洲欧美综合| 岛国毛片在线播放| 亚洲av一区综合| 亚洲av中文字字幕乱码综合| 国产成人福利小说| 精品99又大又爽又粗少妇毛片| 国产精品一区二区性色av| 激情 狠狠 欧美| av专区在线播放| 日韩中字成人| 欧美zozozo另类| 欧美一级a爱片免费观看看| 亚洲最大成人中文| 成人午夜高清在线视频| 白带黄色成豆腐渣| 欧美日韩乱码在线| 久久鲁丝午夜福利片| 青青草视频在线视频观看| 99久久精品一区二区三区| 中文精品一卡2卡3卡4更新| 中文字幕人妻熟人妻熟丝袜美| 国产高清激情床上av| 亚洲av不卡在线观看| 国产精品福利在线免费观看| 国产单亲对白刺激| 神马国产精品三级电影在线观看| 久久久国产成人免费| 成人特级黄色片久久久久久久| 日本爱情动作片www.在线观看| 久久这里有精品视频免费| 久久精品影院6| 国产精品一二三区在线看| 神马国产精品三级电影在线观看| 国产精品一二三区在线看| 亚洲精品久久久久久婷婷小说 | 成熟少妇高潮喷水视频| 看非洲黑人一级黄片| 国产精品蜜桃在线观看 | 岛国在线免费视频观看| 网址你懂的国产日韩在线| 亚洲av第一区精品v没综合| 国产精品99久久久久久久久| 日本av手机在线免费观看| 99热精品在线国产| 亚洲国产欧洲综合997久久,| 美女被艹到高潮喷水动态| 床上黄色一级片| 熟女人妻精品中文字幕| 国产精品久久电影中文字幕| 亚洲第一区二区三区不卡| 欧美成人一区二区免费高清观看| 久久久久久久亚洲中文字幕| 波多野结衣高清无吗| 中文字幕熟女人妻在线| 老师上课跳d突然被开到最大视频| 国产真实伦视频高清在线观看| 欧美色欧美亚洲另类二区| 神马国产精品三级电影在线观看| 亚洲人成网站在线播放欧美日韩| 女人十人毛片免费观看3o分钟| 国产午夜精品论理片| 超碰av人人做人人爽久久| 亚洲欧美成人精品一区二区| 18禁在线播放成人免费| 黄色配什么色好看| 嫩草影院入口| 熟女电影av网| 亚洲国产精品成人久久小说 | 99热这里只有是精品在线观看| 国产av在哪里看| 干丝袜人妻中文字幕| 91aial.com中文字幕在线观看| 国产精品伦人一区二区| 热99在线观看视频| 女的被弄到高潮叫床怎么办| 日韩精品有码人妻一区| 在线播放国产精品三级| 99热只有精品国产| 国产欧美日韩精品一区二区| 国产男人的电影天堂91| 看黄色毛片网站| av在线天堂中文字幕| 久久久久免费精品人妻一区二区| 岛国在线免费视频观看| 在线观看免费视频日本深夜| 久久九九热精品免费| 成人毛片60女人毛片免费| 99在线视频只有这里精品首页| 国产爱豆传媒在线观看| 麻豆乱淫一区二区| 黄片wwwwww| 色尼玛亚洲综合影院| 国产精品久久久久久精品电影| 级片在线观看| 亚洲精品粉嫩美女一区| 欧美日韩国产亚洲二区| 夜夜爽天天搞| 99视频精品全部免费 在线| 男插女下体视频免费在线播放| 色哟哟哟哟哟哟| 成人性生交大片免费视频hd| 午夜福利在线在线| 成人亚洲精品av一区二区| www.色视频.com| 久久人妻av系列| 一本—道久久a久久精品蜜桃钙片 精品乱码久久久久久99久播 | 91aial.com中文字幕在线观看| 极品教师在线视频| 日日撸夜夜添| 免费一级毛片在线播放高清视频| 一级毛片久久久久久久久女| 性色avwww在线观看| 伦理电影大哥的女人| 夜夜爽天天搞| 永久网站在线| 久久久久性生活片| 久久精品人妻少妇| 1024手机看黄色片| 亚洲电影在线观看av| 高清毛片免费观看视频网站| 人人妻人人澡人人爽人人夜夜 | 老熟妇乱子伦视频在线观看| 少妇裸体淫交视频免费看高清| 91精品一卡2卡3卡4卡| 亚洲欧洲日产国产| 午夜精品一区二区三区免费看| 99久久精品热视频| 欧美成人精品欧美一级黄| 国产极品天堂在线| 精品一区二区三区视频在线| 亚洲国产高清在线一区二区三| 国产黄色小视频在线观看| 国产精品人妻久久久久久| 亚洲18禁久久av| 亚洲天堂国产精品一区在线| 日韩精品有码人妻一区| 如何舔出高潮| 国产精品一及| 亚洲国产日韩欧美精品在线观看| 国产精品一区二区性色av| 亚洲内射少妇av| 亚洲美女搞黄在线观看| 国产成人精品久久久久久| 亚洲av第一区精品v没综合| 亚洲在线自拍视频| 男人舔女人下体高潮全视频| 51国产日韩欧美| 赤兔流量卡办理| 色哟哟·www| av在线老鸭窝| eeuss影院久久| 久久久国产成人精品二区| 一个人看的www免费观看视频| 欧美激情在线99| 成人一区二区视频在线观看| 久久鲁丝午夜福利片| 亚洲最大成人av| 久久久精品大字幕| 国产精品久久久久久精品电影| 一卡2卡三卡四卡精品乱码亚洲| 村上凉子中文字幕在线| 国产亚洲精品久久久com| 人妻制服诱惑在线中文字幕| 色哟哟·www| 赤兔流量卡办理| 欧美性猛交黑人性爽| 国产乱人偷精品视频| 欧美bdsm另类| 高清在线视频一区二区三区 | av福利片在线观看| 岛国在线免费视频观看| 国产精品人妻久久久久久| 欧美一区二区国产精品久久精品| 亚洲欧美日韩高清在线视频| 国产成年人精品一区二区| 国产在线精品亚洲第一网站| 美女黄网站色视频| 国产高清三级在线| 亚洲一级一片aⅴ在线观看| 中文在线观看免费www的网站| 男女做爰动态图高潮gif福利片| 一卡2卡三卡四卡精品乱码亚洲| 国产午夜精品久久久久久一区二区三区| 一边摸一边抽搐一进一小说| 日本熟妇午夜| 听说在线观看完整版免费高清| 午夜免费男女啪啪视频观看| 中文精品一卡2卡3卡4更新| 真实男女啪啪啪动态图| 久久久久久久久大av| 观看免费一级毛片| 久久中文看片网| 日韩欧美国产在线观看| 久久久久久久久大av| 中文在线观看免费www的网站| 亚洲av.av天堂| 亚洲在线自拍视频| 国产av不卡久久| 波多野结衣巨乳人妻| 欧美一区二区国产精品久久精品| 成人午夜精彩视频在线观看| 久久精品综合一区二区三区| 亚洲欧美成人综合另类久久久 | 国内久久婷婷六月综合欲色啪| 深夜a级毛片| 少妇人妻精品综合一区二区 | 一夜夜www| 日本熟妇午夜| 久久6这里有精品| kizo精华| 青青草视频在线视频观看| 国产成人freesex在线| 岛国在线免费视频观看| 黄色日韩在线| 一区二区三区免费毛片| 亚洲电影在线观看av| 亚洲国产精品成人久久小说 | 日本黄色片子视频| 最好的美女福利视频网| 欧美xxxx黑人xx丫x性爽| 国产午夜福利久久久久久| 久久精品夜色国产| 成人三级黄色视频| 色尼玛亚洲综合影院| 国产精品人妻久久久影院| 精品久久久噜噜| a级毛片免费高清观看在线播放| 波多野结衣巨乳人妻| 中国国产av一级| a级一级毛片免费在线观看| 中国国产av一级| 久久久久国产网址| 欧美在线一区亚洲| av天堂中文字幕网| 日本欧美国产在线视频| 边亲边吃奶的免费视频| 可以在线观看的亚洲视频| 只有这里有精品99| a级毛色黄片| 亚洲国产精品sss在线观看| 日本在线视频免费播放| 黑人高潮一二区| 99久久精品一区二区三区| 成人漫画全彩无遮挡| 九草在线视频观看| 性色avwww在线观看| av在线蜜桃| 人妻系列 视频| eeuss影院久久| 真实男女啪啪啪动态图| 国产一区二区在线av高清观看| 91精品一卡2卡3卡4卡| 99国产精品一区二区蜜桃av| 在线观看免费视频日本深夜| 亚洲国产色片| 亚洲自拍偷在线| 亚洲国产精品国产精品| 97超视频在线观看视频| 中文字幕av在线有码专区| 亚洲七黄色美女视频| av.在线天堂| 国产美女午夜福利| 日本与韩国留学比较| 中文字幕制服av| 精品午夜福利在线看| 国产精品.久久久| 亚洲人与动物交配视频| 日韩制服骚丝袜av| 国产精品99久久久久久久久| 久99久视频精品免费| 黄片wwwwww| 国产成人午夜福利电影在线观看| 国产精华一区二区三区| 男人舔女人下体高潮全视频| 亚洲成av人片在线播放无| 午夜精品在线福利| 婷婷六月久久综合丁香| 日韩在线高清观看一区二区三区| 久久午夜亚洲精品久久| av免费观看日本| 在线国产一区二区在线| 99久国产av精品国产电影| 国内揄拍国产精品人妻在线| 亚洲丝袜综合中文字幕| 18禁裸乳无遮挡免费网站照片| 婷婷色综合大香蕉| kizo精华| 1024手机看黄色片| www.色视频.com|