• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Multi-Branch Deepfake Detection Algorithm Based on Fine-Grained Features

    2023-12-12 15:50:08WenkaiQinTianliangLuLuZhangShufanPengandDaWan
    Computers Materials&Continua 2023年10期

    Wenkai Qin,Tianliang Lu,?,Lu Zhang,Shufan Peng and Da Wan

    1School of Information Network Security,People’s Public Security University of China,Beijing,100038,China

    2Department of Investigation,Shandong Police College,Jinan,250200,China

    ABSTRACT With the rapid development of deepfake technology,the authenticity of various types of fake synthetic content is increasing rapidly,which brings potential security threats to people’s daily life and social stability.Currently,most algorithms define deepfake detection as a binary classification problem,i.e.,global features are first extracted using a backbone network and then fed into a binary classifier to discriminate true or false.However,the differences between real and fake samples are often subtle and local,and such global feature-based detection algorithms are not optimal in efficiency and accuracy.To this end,to enhance the extraction of forgery details in deep forgery samples,we propose a multi-branch deepfake detection algorithm based on fine-grained features from the perspective of fine-grained classification.First,to address the critical problem in locating discriminative feature regions in fine-grained classification tasks,we investigate a method for locating multiple different discriminative regions and design a lightweight feature localization module to obtain crucial feature representations by augmenting the most significant parts of the feature map.Second,using information complementation,we introduce a correlation-guided fusion module to enhance the discriminative feature information of different branches.Finally,we use the global attention module in the multi-branch model to improve the cross-dimensional interaction of spatial domain and channel domain information and increase the weights of crucial feature regions and feature channels.We conduct sufficient ablation experiments and comparative experiments.The experimental results show that the algorithm outperforms the detection accuracy and effectiveness on the FaceForensics++and Celeb-DF-v2 datasets compared with the representative detection algorithms in recent years,which can achieve better detection results.

    KEYWORDS Deepfake detection;fine-grained classification;multi-branch;global attention

    1 Introduction

    In recent years,with the continuous deep learning research,deepfake techniques have also made rapid development,especially the generative models such as autoencoder and Generative Adversarial Networks(GAN)[1]have been applied to the technology,which makes it easy to achieve high realism generation and replacement of video faces.This technology which uses deep learning methods to generate high-quality forged videos and images is called“deepfake”[2].People can generate realistic fake face images using deepfake that are difficult to detect with traditional techniques.Compared with conventional forgery methods,deepfake techniques enrich the details of face forgery and greatly enhance the realism of forged faces,bringing a new challenge to detecting and identifying forged content.These forged video images are widely disseminated on the Internet through social media,which not only violate personal privacy and reputation but also mislead social opinion,confuse the public,and even cause political disputes,posing a significant threat to national security[3].Therefore,to cope with the security risks brought by deepfake technology,researchers have carried out a series of explorations in the academic community,and the design of a general and efficient deepfake detection algorithm has become one of the research hotspots[4–8].

    The deep generative model is at the heart of deep face video forgery techniques represented by “deepfake”.The deep generative model can be simply described as a model that uses a deep neural network for data generation to generate new data that is not included in the training dataset.Currently,forgery generation algorithms are divided into two categories.One replaces the face to realize the tampering of identity information,and the other replaces only the expression of the face without changing person’s identity information.However,these generation techniques also have obvious drawbacks,such as the presence of noticeable visual artifacts in the generated forged video and the difficulty of maintaining frame-to-frame consistency between consecutive video frames.This is because the characteristics of its probabilistic model,the face generation process often a certain degree of randomness,which makes the synthesized face video has a large jitter,and the final forgery effect still needs to be further improved.Therefore,with the continuous improvement of the generation technology,the new forgery method focuses on solving the various shortcomings existing in the existing techniques,and the quality of the fake face produced in the face forgery process is also continuously improved,making the generated image or video more realistic.

    Various deepfake detection methods have been proposed by researchers from different research perspectives,some of which can achieve high identification accuracy on open-source forgery datasets.Most of them model deepfake detection as a binary classification problem,and they tend to extract the global features of an image using a backbone network and then feed them into a binary classifier to distinguish between true and false.Depending on the type of feature selection,these detection methods can be classified as temporal feature-based and spatial feature-based.The temporal featurebased detection methods mainly use Recurrent Neural Networks(RNN)to learn the temporal features between forgery video frames.However,these methods rely heavily on the video’s preprocessing and the frame sequence’s length,resulting in poor generalization for cross-library detection.At the same time,it can not meet the detection needs in the face of highly compressed videos and complex video backgrounds due to less attention to the local forgery details within frames.Therefore,the detection methods based on spatial features have become the mainstream in deepfake detection tasks due to their better robustness and higher accuracy.These methods transform video detection into an image detection problem by extracting spatial features using Convolutional Neural Networks(CNN)after randomly selecting videos frames.However,in real scenarios,the differences between real and fake samples often exist in subtle ways,such as changes in the curvature of the lips of a face,folds in the facial skin texture,etc.These differences can be easily ignored by these in-frame global featurebased detection algorithms.These detection algorithms based on intra-frame global features can easily overlook these differences,resulting in the model ignoring many discriminative features in the feature extraction phase,which induces the model to make wrong judgments about the authenticity of the samples.In deep forgery detection,due to the extensive presence of forgery traces in different regions of the sample,researchers need to pay more attention to the influence of interactions between various features to avoid monolithic feature learning.In addition,we examined that the commonly used channel or spatial attention mechanisms can only effectively capture local information and cannot establish the functional dependency between space and channel,making it difficult to fully utilize the global information across dimensions.Therefore,this paper makes the detection model focus on critical features more effectively by incorporating global attention mechanism into the backbone network.

    Therefore,to solve the difficulties and problems in the detection mentioned above,this paper investigates the method of introducing localized fine-grained features in the backbone network inspired by the idea of fine-grained classification.And to enhance the cross-dimensional interactions of attention information,we embed the global attention mechanism in the backbone and design a multi-branch deepfake detection algorithm.Specifically,the improvements in this paper are mainly carried out in the following three aspects:

    1.The recent detection algorithms based on intra-frame features directly extract global features as the source of classification.It is often difficult to focus on the subtle differences between authentic and fake samples.Therefore,to mine these fine-grained features hidden in different regions,we design a feature localization module and use it to transform the backbone into a multi-branch detection model,thus locating multiple discriminative regions and obtaining multi-scale fine-grained information.

    2.In the training phase,we introduce the global attention module [9] in the second branch to calculate the model attention weights.The global attention module can enhance the cross-dimensional interaction of spatial and channel information while increasing the weights of crucial feature regions and channels to assign more appropriate attention weights to the feature maps and improve the accuracy of localization and detection.And we validate the effectiveness of embedding the global attention module in the backbone through various ablation experiments.

    3.To explicitly explore the relationship and calculate the complementary information between the features of different branch regions,we employ the correlation-guided fusion module in the feature integration phase.And then,under its guidance,we fuse the information of different branch features to provide a diverse basis for identifying genuine and forged samples.

    4.In the final integrated decision-making phase,to improve the correlation between standard and forged samples,we formulate a multi-branch loss function,which uses cross-entropy loss as an end-to-end loss function in all three branches.

    The structure of our treatise is organized as follows.Section 2 combs through the existing work and relevant references in the field of deepfake detection in recent years and introduces the sources and limitations that inspired us to make algorithmic improvements.Section 3 is the basic methodology and implementation details of our approach,including the principles and model construction for the improvement module.Section 4 provides the details of the parameter settings,environment configurations,dataset processing,and various results of our experiments,which validate the effectiveness of our model.Section 5 concludes with a summary of contributions and proposes future research directions to advance the field of deep forgery detection.

    2 Related Work

    2.1 Deepfake Detection Based on Deep Learning

    To cope with the impact of deepfake techniques on public security,researchers have proposed various deepfake detection methods,which can be divided into detection based on video intraframe features and detection based on video interframe features.

    2.1.1 Detection Based on Interframe Features

    The detection methods based on interframe timing features to focus on mining the timing features between consecutive frames of the video and classifying the video authenticity by detecting the interframe continuity and integrity.Since deepfake techniques can not forge the video as a whole but tamper with specific regions of the face frame by frame,it is difficult to take into account the sequence of previously forged frames when forging the current frame,which leads to discrepancies in the spatiotemporal distribution of successive frames such as differences in the expressions of the characters,inconsistent spatiotemporal states of the tampered regions,visual artifacts and noise inside successive frames.Such discrepancies can be captured by RNN or other methods based on sequence feature extraction and used to detect actual and forged videos.

    Agarwal et al.[10] encoded face expressions,head movements,and other physiological signals,then used a support vector machine(SVM)for detection and classification.Lima et al.[11]employed spatiotemporal convolution to detect temporal and spatial artifacts of forged videos as well as utilized multiple layers of coding units based on the self-attention mechanism to enable each frame feature to aggregate other frame information,thus enabling effective detection of current mainstream forgery datasets.Cozzolino et al.[12]trained an ID-Reveal model using a self-supervised learning paradigm to mine interframe temporal features through ID networks and perform adversarial learning using 3DMM generative networks to identify the authenticity of face videos.Sun et al.[13] proposed a robust framework LRNet to detect forged videos by modeling accurate geometric features.

    2.1.2 Detection Based on Intraframe Features

    The main idea of the detection methods based on intraframe spatial features is to randomly extract image frames from the video stream,input them into CNN for feature extraction after preprocessing,such as face alignment and face cropping,and finally use the feature information for classification.These approaches first transform the detection object from video to image and use classical models such as ResNet[14],Xception[15],and EfficientNet[16]as backbone networks for classification.

    Currently,most deepfake detection is carried out based on interframe spatial features.Afchar et al.[17] found that the artifacts of deepfake images are often present in the middle layer semantic information and proposed using MesoNet combined with the Inception module to extract image middle layer features for detection.Nguyen et al.[18] designed to use the VGG16 model to extract features and then employ a capsule network to learn more detailed features of the face.Nirkin et al.[19]used three encoders based on the Xception structure to encode the whole image,face region,and background region after preprocessing.Finally,they combined the three feature vectors to achieve the effect of detection and localization.Shang et al.[20]constructed a pixel-region relationship network(PRRNet)and used the pixel relationship module and the region relationship module to detect spatial association and inconsistent forgery traces in images.

    However,as the forged content becomes more and more realistic,the above binary classification models based on global features have poor detection performance.The reason is that the highly distinguishable discrepancies between forgery samples and original samples often exist in subtleties of the image,such as lip curvature changes and skin texture differences,which are often not easily captured by the detection algorithm based on global features,resulting in reduced detection accuracy.Such subtle and located differences are similar to the fine-grained classification.Therefore,to improve the detection based on spatial features,researchers have started to analyze the deepfake detection task expressed as a fine-grained classification problem.

    Zhou et al.[21]designed a detection network based on multi-attention and used bilinear pooling to aggregate low-level texture features and high-level semantic features under the guidance of attention graphs.Guarnera et al.[22]utilized an expectation-maximization algorithm to extract convolutional traces from specific local pixels.Chen et al.[23] used a pixelhop++module to extract features and reduce the dimensionality of local regions separately in multiple regions of face images and finally integrated and classified the information of each region.Liu et al.[24] divided the original image into several blocks of the same size and randomly shuffled pixels within blocks and positions between blocks,forcing the model to extract more discriminative forgery traces.Based on the lack of physiological constraints in the current generative models,Hu et al.[25] proposed to explore whether the number,shape,and relative positions of the reflected glare of forged eyes are consistent for detection.Guo et al.[26] first located the human eyes’oval pupil mask and then used the IoU algorithm to calculate the difference between the fake pupil shape and the real pupil shape to determine the authenticity of the input image.

    In recent years,with the rapid development of Transformers[27],Vision Transformer algorithms[28]that can be used in computer vision have been derived.These algorithms can extract the global association relationship of pixels,and the number of operations required to compute the association between two positions does not increase with distance compared with the CNN.Moreover,the internal self-attention mechanism can generate more interpretable models with stronger modeling ability and detection performance,which have been gradually applied in the study of deepfake detection.

    2.2 Fine-Grained Classification

    Fine-grained classification is a challenging research task in computer vision,which distinguishes different fine-grained categories by capturing local discriminative features [29,30].Compared with traditional classification tasks,fine-grained classification provides a more detailed classification of images.Since the discrepancies between categories are minor,the essence of this classification task is to locate the target and local regions and perform feature extraction and processing to complete the training and detection of the classifier.

    The key to fine-grained classification is how to locate the subtle discrepancies that exist in local regions.The main research directions can be divided into strongly supervised and weakly supervised approaches For different ways of extracting fine-grained features.The strongly supervised approach uses additional information,such as annotation frames and part annotation points,in addition to category labels in classification.Therefore,it requires more manual effort and time consumption.The weakly supervised approach only uses category labels to complete the model’s training.Currently,research in this field mainly focuses on locating discriminative regions in a weakly supervised manner[31–34].In this paper,we define deepfake detection as a particular fine-grained classification problem,and both have the same characteristics in extracting subtle and discriminative features.However,deepfake detection only involves two categories,i.e.,real and fake.

    2.3 Attention Mechanism

    In deep neural networks,the attention mechanism can mainly be divided into channel attention mechanism and hybrid attention mechanism.

    The channel attention mechanism focuses on the correlation between different channels of the feature graph and automatically obtains the importance of each feature channel through network learning,and finally assigns different weight coefficients to each channel to reinforce the critical features and suppress the non-important features.The commonly used are the squeeze-and-excitation(SE)module[35]and the efficient channel attention(ECA)module[36].The SE first compresses the spatial dimension of the graph by global average pooling operation and then extracts the channel attention information using a fully connected layer.The ECA,on the other hand,utilizes a 1 × 1 convolution instead of the fully connected layer to achieve information interaction across channels based on the SE.

    The hybrid attention mechanism is a more integrated attention method that combines the advantages of channel dimension attention and spatial dimension attention.The commonly used are the bottleneck attention module(BAM)[37],the convolutional block attention module(CBAM)[38],and the coordinate attention(CA)module[39].The BAM and CBAM can extract the attention information of both channel and spatial dimensions and fuse different attention information.The CA extracts channel attention information from the width and height directions,combining the position information and channel attention information.In 2021,Liu et al.[9]proposed the global attention module(GAM),which still uses the channel-space attention arrangement order.But they redesigned the channel attention submodule and the space attention submodule.When the GAM is embedded in the backbone model,the model cannot only consider the attention weight information extraction in both channel dimension and spatial dimension but also ensure the cross-dimensional information interaction to improve the detection accuracy further.

    3 Methods

    In this paper,we propose a multi-branch deepfake detection algorithm based on fine-grained features,and the algorithm architecture is shown in Fig.1.

    Figure 1:Architecture of multi-branch deepfake detection algorithm based on fine-grained features

    As aforementioned,the discrepancies between real and fake faces are usually subtle and occur in local regions,which are not easily captured by the models based on global features.Therefore,to collect local features from different regions more effectively,we design a multi-branch detection framework.In this framework,three main components are embedded in the backbone network:(1)We use the feature localization module (FLM) to locate the subtle discrepancy in different regions.(2) We employ the correlation-guided fusion module(CGFM)to aggregate complementary information from different branches.(3) We introduce the global attention module (GAM) to increase the weights of critical regions and channels and enhance the cross-dimensional interaction.Besides,to learn effectively the fine-grained features of the sample,we specially design a multi-branch loss.

    3.1 Feature Localization Module

    To locate forgery traces in different regions,we introduce the feature localization module to obtain significant feature representations by mining the most prominent parts of the feature map and then suppressing them to force the model to explore other potential discriminative features in subsequent branches.Moreover,by embedding the feature localization module into different layers of the backbone network,we construct a multi-branch fine-grained feature extraction model,which can obtain feature representations of multiple regions.The specific structure of the feature localization module is shown in Fig.2.

    Figure 2:Structure of feature localization module

    In the feature localization module,the feature map from a particular layer is first defined asF∈RC×W×H,whereC,W,Hdenotes the number of channels,width,and height of the feature map,respectively.In the feature mapF,it is uniformly partitioned intokstriped sections along the width dimension,and each striped section is represented asF(i)∈RC×(W/k)×H,i∈[1,k].We defineA(i)as a high attention activation map for each part as shown in Eq.(1):

    where the nonlinear activation function Reluis employed to remove the negative activation parts.φrepresents the 1×1 convolution,which is used to calculate the importance of each striped section and share the parameters among different parts.In the activation mapA(i),to measure the contribution of each striped section to the detection,we use Eq.(2)to gain its average value:

    whereGAPdenotes the global average pooling.biis used as an importance factor,and its value reflects the contribution of the striped sections.If thebivalue of a part is larger,it indicates that it is more likely to belong to the critical regional features.The normalization ofbiis performed by theSoftmax function,which is expressed in Eq.(3):

    whereis the normalized importance factor,which are combined to obtain the feature vectorB′=The vector has a global sensory field,which can obtain the statistical information of the feature mapFon different striped sections.The most significant region on the feature mapFcan be represented using the feature vector,and the enhanced featureFbis obtained by feature fusion.The feature fusion process is shown in Eq.(4):

    where the magnitude of the weight parameterαreflects the proportion of the most significant region in the feature fusion,and the degree of feature enhancement can be effectively controlled by adjustingα.?denotes element-by-element multiplication.Finally,as shown in Eq.(5),through applying a convolution layerφ′onFb,we can obtain a specific high-concentration featureFpas follows:

    In order for the subsequent modules to continue mining discriminative features in other regions,the suppression factorsican be obtained by Eq.(6),which suppress the high attention regions on the current branch and force the subsequent modules of the model to focus more on other regions of the feature mapF,and then learn different scale features.

    where the magnitude of the weight parameterβreflects the degree of suppression of the most significant region in the feature mapF.The combination of the suppression factorssiyields the feature vectorS=(s1,s2,...,sk)T.In the current branch,the most significant region of the feature mapFis suppressed byS,as shown in Eq.(7),and we can obtain the suppressed potential feature mapFs.

    The feature mapFsstill contains many discriminative features in other regions except the suppressed high-concentration features.Therefore,to locate these discriminative features,Fsis continued to be fed into the feature localization module of the subsequent branches to capture the multi-scale discriminative features more effectively.

    3.2 Correlation-Guided Fusion Module

    Although the feature localization module can locate the highly focused part of the sample,the relationship among the features of each branch can be easily ignored because the branch structures are relatively independent of each other.To extract a more robust feature representation,we introduce a correlation-guided fusion module for the feature information learned from different branches.The correlation-guided fusion module consists of several feature complementary modules(FCM),which first explore the complementary information among the features of each branch to avoid ignoring other discriminative features due to the focus on individual local features,and then enhance the feature representation of the current branch through feature fusion guided by the complementary information.The structure of the feature complementation module is shown in Fig.3.

    In the feature complementation module,the regional feature mapsFp1output from any two branches are defined aswhereCdenotes the number of channels,andW1×H1andW2×H2denote the spatial size of each feature map,respectively.To explore the relationship among the features of different branches,we represent the feature vector of each spatial position along the channel dimension of the feature mapFas a pixel,and characterizes the correlation among the feature maps by the correlation among pixels,as shown in Eq.(8):

    wherepixcel(F,i)denotes the feature vector of theithregion along the channel dimension on the feature mapF,andFj,idenotes theithregion located on thejthchannel of the feature mapF.After obtaining the feature vectors along the channel dimension for each region onFp1andFp2,respectively,we calculate the vector inner product to reflect the similarity among the feature image elements more concisely,as shown in Eqs.(9)and(10):

    where the matrixMrepresents the similarity matrix between feature mapsFp1andFp2,and the elementMi,jrepresents the similarity between theithpixel of feature mapFp1and thejthpixel of feature mapFp2.If theMi,jbetween two pixels is smaller,the less similar and more complementary these two pixels are to each other.Therefore,based on the negative correlation between similarity and complementarity,to reflect the complementarity more intuitively,we define the complementarity matrixCas the negative matrix of the similarity matrixM,and normalize the columns and rows of the complementarity matrix through Eqs.(11)and(12)as follows:

    wheredenotes the complementary information extracted fromFpjagainstFpi,which is used as the basis for enhancingFpiin the subsequent aggregation of branch features.To further explore the complementary relationship between the two feature maps,the mining of complementary information operation is transformed as pixel-specific through Eqs.(15)and(16)as follows:

    where each pixel inis enhanced by all pixels inFp2,and each pixel inis enhanced by all pixels inFp1.If the complementarity among the pixels inFp1andFp2is stronger,it indicates that the pixels in one of the feature maps contribute more to the pixels in the other feature map.Then each pixel in both can mine semantic complementary information from different regions of the other feature.Therefore,the complementary information of regional features in each branch can be obtained according to Eq.(17):

    whereXpjpican be obtained by Eqs.(9),(13)and(14).In addition,as shown in Fig.3,for the regional feature informationFp1andFp2in any two branches,the feature complementation module can calculate both the complementary informationandat the same time.Combining the complementary information from different branches,we can enhance the regional feature information of the current branch through feature fusion to prompt the model focus on different fine-grained features,as Eq.(18)shown in:

    whereYpidenotes the enhanced regional features,which contain discriminative information at different scales.The magnitude of the weight parameterγreflects the proportion of complementary information from different branch features in the feature fusion process.

    Figure 3:Structure of feature complementary module

    3.3 Global Attention Module

    In this paper,we choose the Xception as the backbone model,which has a good effect in image classification,and design a multi-branch structure.We find that the suppressed potential featureFsgenerated in the previous branch still not only contains a large number of potential discriminative features but also includes some background and noise information.These interferences can impact the localization of subsequent potential discriminative features and eventually lead to a drift between feature localization and the target region.

    Therefore,this paper introduces the global attention module in the Middle Flow of Xception to build a more robust network structure by enhancing the model’s ability to extract key features of forged images.The global attention module utilizes spatial and channel information at different scales on the feature map,increases the weights of crucial feature regions and feature channels,and enhances the global cross-dimensional interaction of spatial and channel information so that the network can obtain more effective feature representation capability.The structure of GAM is shown in Fig.4.

    Figure 4:Structure of global attention module

    The GAM comprises two submodules: the channel attention module (CAM) and the spatial attention module (SAM).Similar to the CBAM,the GAM belongs to the hybrid attention module that incorporates channel and spatial attention.

    In the CAM,the input feature map is given asF1∈RC×W×H,whereC,W,Hdenote the number of channels,width,and height of the feature map,respectively.To obtain global cross-dimensional interaction information,the channel attention module uses a 3D permutation to retain information across three dimensions and get the feature map∈RW×H×C.Thenis input to a two-layer multilayer perceptron(MLP),which first compresses the channels asC/rand then reduces them to amplify the cross-dimensional dependence of channel and space.Finally,the feature map dimensions are reduced using another 3D permutation,and the result is element-by-element multiplied by the original featureF1after the activation functionsigmiodto obtain the outputF2∈RC×W×H.The calculation process is shown in Eqs.(19)and(20):

    wheresigmoiddenotes the activation function.The parametersW0∈RC×C/randW1∈RC/r×Cdenote the parameter-sharing multilayer perceptron weights.rdenotes the compression rate of the hidden layer,andMc(F1)denotes the output channel feature vector.The structure of the CAM is shown in Fig.5.

    Figure 5:Structure of channel attention module

    In the SAM,the input feature map is the outputF2∈RC×W×Hof the channel attention module,whereC,W,Hdenote the number of feature map channels,width,and height,respectively.To focus on spatial information more precisely,the spatial attention module employs two convolutional layers of 7×7 for compression of channel dimension and fusion of spatial information.Meanwhile,since max pooling reduces the amount of information,the spatial attention module removes the pooling layer to further preserve the feature information.Besides,the spatial attention module uses grouped convolution on the channel dimension to reduce the negative effect of the significant increase in the number of parameters due to removing of the pooling layer.Finally,the result is element-by-element multiplied with the original featureF2after the activation functionsigmoidto obtain the outputF3∈RC×W×Hof the spatial attention module,i.e.,the final output of the global attention module.The calculation process is shown in Eqs.(21)and(22):

    wheresigmoiddenotes the activation function.Conv7×7denotes the convolutional layer using a convolution kernel of size 7×7.The compression rate of the channel between the two convolution layers is represented asr′.Ms(F2)denotes the output spatial feature vector.The structure of the SAM is shown in Fig.6.

    Figure 6:Structure of spatial attention module

    In this paper,we embed the GAM into the Block of Xception as a kind of data enhancement module,which makes it possible to modify the Block without affecting the branch structure,as shown in Fig.7.The model reconstructs the extracted feature maps by embedding the global attention mechanism.It can improve the attention of subsequent branches on essential features and eliminate the interference of background information as much as possible.In addition,the global attention module achieves cross-dimension interaction without reducing dimensionality,which avoids information loss while reducing the number of parameters and computational cost.

    Figure 7:Block structure embedded in GAM

    3.4 Multi-Branch Loss Function

    To assess the validity of the model classification results,this paper designs a multi-branch loss function that uses cross-entropy loss as an end-to-end loss function in all three branches.It is expressed in Eqs.(23)–(25)as:

    whereLidenotes the loss of theithbranch.yiis the true label of the input sample which is represented by the one-hot vector.clsidenotes the classifier of theithbranch.Pidenotes the classification probability calculated by theithbranch.Ltotaldenotes the final loss.

    In the training phase,the final lossLtotalis obtained through calculating the classification loss of the enhanced regional featuresYpiin each branch.We use the AdamW algorithm to optimize the final loss and accelerate the convergence of the model.The three branches together finally make the model learn the various fine-grained features of the forged sample.In the testing phase,the model takes the average prediction scores of all branches as the final prediction result.

    4 Experimental Results and Analysis

    4.1 Experiment Environment

    The experimental platform in this paper is a 64-bit Linux operating system with Ubuntu 18.04 LTS,an NVIDIA GeForce RTX 2080Ti graphics card with 11 GB of video memory,an Intel(R)Xeon(R)CPU E5-2678 v3@2.50 GHz,and 62 GB of RAM.The experimental code is implemented in the Pytorch deep learning framework with version 1.11.0,Anaconda version 4.12.0,and Python version 3.8.

    4.2 Introductions of Datasets

    In this paper,we select two mainstream deepfake datasets for experiments,i.e.,FaceForensics++[40]and Celeb-DF[41].Some samples of datasets are shown in Fig.8.

    Figure 8:Samples of the three datasets

    The FaceForensics++(FF++) dataset is one of the most significant and widely used datasets.The dataset captures 1000 short videos containing unobstructed faces from YouTube and ensures that the faces are contained in consecutive frames.Then it is manipulated by five forgery techniques(DeepFakes,Face2Face,FaceSwap,Neural Textures,and FaceShifer)to generate a total of 5000 forged videos with three different compression ratios(lossless compression:Raw,high-quality compression:c23,and low-quality compression: c40).Among them,DeepFakes is based on an autoencoder for face replacement;FaceSwap is a graphics-based approach.To verify the effectiveness of the proposed algorithm for the deepfake detection task,we select the DeepFakes and FaceSwap datasets with a compression rate of c23 in this paper.

    The Celeb-DF dataset captures 590 real videos of 59 celebrities from YouTube,which consider people of different genders,ages,and races.It generates 5639 forged videos in MPEG4.0 format with an average length of 13 s using a single DeepFakes approach.Since this dataset can fuse the boundary between the forged and original region by enhancing the face resolution and establishing a color conversion algorithm for the faces in the forged and original videos,it dramatically improves the quality of the forged data.The Celeb-DF dataset is currently recognized as a high-quality deepfake dataset.In this paper,we use the extended Celeb-DF-v2 dataset for our experiments.

    4.3 Experiment Setting

    Since the forgery regions are primarily concentrated in the facial regions,to make the model focus more on the features of the forgery traces,the experiments in this paper firstly intercept 30 frames at equal intervals for each video in the data preprocessing stage.Second,we use the face detection algorithm RetinaFace to locate the critical points of the five facial features in each frame to determine the facial rectangle.After the face alignment,the images are cropped to 320 × 320 input samples.Then,we employ the sharpening operation and Gaussian blur on the samples to refine the images and reduce image noise.We standardize and normalize the images to enhance the model’s generalization ability.Finally,the datasets are divided into train-sets and test-sets in the ratio of 7:3,and the division results are shown in Table 1.

    Table 1:Division results of datasets

    In addition,the specific settings of other hyperparameters are shown in Table 2.

    Table 2:Hyperparameters setting

    4.4 Evaluation Index

    In this paper,we use the Accuracy(Acc)and Area under Receiver Operating Characteristic(ROC)Curve(AUC)to evaluate the model effect comprehensively.

    Acc is used to describe the classification accuracy of the classifier for genuine and fake samples,and a larger value of Acc indicates a higher correct classification rate of the model.AUC is defined as the probability that the predicted probability value of getting real samples is greater than the probability of fake samples,and a larger value of AUC indicates the better performance of the detection model.The calculation formula of Acc and AUC is shown in(26)and(27):

    whereTPdenotes the real face image predicted as true by the model.TNdenotes the fake face image predicted as false by the model.FNdenotes the real face image predicted as false by the model.FPdenotes the fake face image predicted as true by the model.Andndenotes the total number of positive and negative samples.

    4.5 Analysis of Experimental Results

    4.5.1 The Ablation Experiment of Each Improved Strategy in Terms of Performance Gain

    In this experiment,we use the Xception as the baseline model and design a set of ablation experiments on the DeepFakes dataset and FaceSwap dataset to verify the performance gains brought by the feature localization module(FLM),correlation-guided fusion module(CGFM),global attention module(GAM)and their combinations to the model detection.The evaluation metrics include Acc value and AUC value.The experimental results are shown in Table 3 and Fig.9.

    Table 3:The classification effect performance gain generated by each improved strategy

    Figure 9:The performance gains brought by different improvement strategies on two datasets

    Based on Xception,when the FLM is used only,the detection accuracy of the model is improved by 0.94%and 0.67%on the two datasets,respectively,proving FLM’s effectiveness.After embedding the FLM in the baseline model,the Xception is transformed into a multi-branch structure,allowing the model to focus on the subtle discrepancies among different regions of the real and fake images so that more potential feature information can be mined,providing more possible bases for subsequent classification.

    When the FLM and CGFM are used together,the detection accuracy of the model is improved by 0.28%and 0.39%,respectively,compared with using the FLM alone.Furthermore,the AUC and loss are also optimized.It indicates that the model’s performance can be further improved by using the CGFM and FLM together.The reason is that the submodule FCM in CGFM can effectively explore the complementary information between the features of two branches and enhance the regional features in each branch through feature fusion guided by the complementary information to make them more diverse and discriminative.

    After introducing the GAM,compared with the baseline model,using the GAM alone leads to 0.76% and 0.45% improvement in detection accuracy on the two datasets,respectively.Compared with using FLM and CGFM,the simultaneous introduction of the three improvement strategies leads to 0.64% and 0.32% improvement in model detection accuracy,respectively.The experimental comparison results show that the global attention module enables the model to better represent critical information through the global cross-dimensional interaction.At the same time,it further demonstrates that the combination of the three improvement strategies can obtain the best detection results,which verifies the effectiveness of the proposed method in deepfake detection in this paper.

    4.5.2 The Ablation Experiment of Global Attention Module Embedding

    To explore the best embedding position of the global attention module in the multi-branch model,this paper designs seven types of attention module embedding methods and inserts them into the Blocks of different branches,respectively.Acc and AUC are used for the experimental indexes.The specific embedding methods and detection results are shown in Table 4 and Fig.10.

    Table 4:The classification effect performance gain generated by each embedding method

    Figure 10:The classification effect and the number of model parameter change on two datasets

    According to the experimental setting,the results in the table can be classified as unembedded GAM,single-branch embedded GAM,and multi-branch embedded GAM.Method (1) does not embed GAM.Methods (2) (3) (4) embed GAM in different branches,respectively.Methods (5) (6)(7)(8)belong to multi-branch embedded GAM.

    The model achieved the highest Acc and AUC values on both datasets when the GAM is embedded in the second branch only.Compared with method (1),embedding GAM in the second branch increased the Acc values by 0.61%and 0.64%,and embedding GAM in the third branch increased the Acc values by 0.2%and 0.39%.However,there is a slight decrease in the model detection accuracy after embedding the GAM in the first branch.In the first branch,the model still mines the global features of the sample,which contains both forgery information and a lot of background information.Therefore,embedding the GAM in the first branch tends to make the model focus on irrelevant background information,which causes bias in detection.

    At the same time,compared with the case of method(1),the model detection accuracy in the case of embedding GAM in multiple branches does not improve significantly with the increase of attention module numbers.Therefore,considering model detection accuracy and the number of parameters brought by the introduction of the attention modules,in this paper,we choose to insert the GAM in the Block of the second branch.

    4.5.3 The Comparative Experiment of Different Attention Modules

    To further verify the effectiveness of GAM for model performance improvement in the deepfake detection task,we replace the GAM embedded in the backbone network with the SE[29],ECA[30],BAM[31],CBAM[32],and CA[33].Meanwhile,the improved model with the FLM and CGFM is defined as the baseline to compare the detection performance of the model after embedding different attention modules on two kinds of datasets.Experimental metrics use Acc and AUC.The experimental results are shown in Table 5 and Fig.11.

    According to the comparison results in Table 5,it can be found that the model performance does not improve significantly and even decreases after embedding the SE and the ECA in the branches,respectively.The analysis of the reason for this phenomenon may be related to the lack of spatial information.In extracting channel attention weights,the feature map’s spatial dimension is compressed,which makes it easy to ignore the spatial information in the process of feature fusion among channels,leading to a decrease in the model detection accuracy.In addition,although the CA injects the position information into the channel attention,it focuses mainly on the channel information of images and cannot capture spatially long-distance dependencies.

    When using the BAM and CBAM,the final results are still unsatisfactory despite the slight improvement in model performance compared with the baseline model.Due to the lack of global cross-dimensional interaction of spatial and channel information,the network cannot obtain a more accurate feature representation capability.

    Figure 11:The classification effect of using different attention modules on two datasets

    After embedding the GAM,the detection performance of the model is significantly improved,indicating that compared with other mainstream attention modules,the GAM provides more effective attention guidance in deepfake detection tasks,which is conducive to improving model detection accuracy.

    4.5.4 The Comparative Experiment with Other Algorithms

    To improve the effectiveness of the model in this paper for the deepfake detection task,nine mainstream deepfake detection models are selected for comparison in this experiment,and the evaluation metrics include Acc and AUC.The comparison results are shown in Table 6 and Fig.12.

    Table 6:Comparison with other models

    Figure 12:Comparison with other models

    Based on the comparison results,it can be seen that the proposed algorithm in this paper has significant improvements in detection performance compared with the mainstream detection models.On the DeepFakes and FaceSwap datasets,the proposed algorithm achieves the best results regarding Acc and AUC values,with 98.67%and 98.55%for Acc and 99.80%and 99.72%for AUC,respectively.Meanwhile,to test the ability of the algorithm to deal with samples with more complex forgery degree,the Celeb-DF-v2 dataset with higher forgery quality is selected experimentally,and the Acc value of 98.37%and the AUC value of 99.58%are obtained,which have better improvements in the detection accuracy and effect than most current algorithms.

    In summary,the detection algorithm proposed in this paper is more accurate in detecting forged samples,which is mainly due to the multi-branch structure that enables the model to mine the finegrained feature information in different regions of the samples and fuses the features among different branches through the correlation-guided fusion module,to learn more discriminative fine-grained features.

    4.5.5 Visualization of Results

    To represent the gain effect of the improvement module on the multi-branch network more intuitively,we use the Grad-CAM graph[42]to visualize the regions of the detection model in different branches focusing on samples,as shown in Fig.13.

    The Grad-CAM is a deep network visualization method based on gradient localization.Specifically,the activation map is obtained as a cumulative weighted average of activation values across channel dimensions given a feature map.The red part in Fig.8 represents the region of interest for the model.For the forged samples chosen from the three datasets in the experiment,the activation maps in the first column to the third column correspond to the attention maps drawn from different branches of the model,respectively,and the three rows represent the changes in the activation maps in the case of different improvement methods.

    Figure 13:Visualization of the forged region

    By comparing the discrepancies among the proposed different improvement methods in the class activation attention graph,it can be found that: (1) After the introduction of FLM,the model can locate different regions of the sample and starts to mine feature information from these fine-grained regions.(2)After the introduction of CGFM,the feature information learned in all branches is focused on the diverse regional information mined by different branches.(3)After the introduction of GAM,different weights are assigned to different channels and regions,improving the feature map’s multiscale feature representation capability.The visualization experiments demonstrate the contribution of FLM,CGFM,and GAM to the model in the deepfake detection task,which gives the model better detection capability.

    5 Conclusions

    In this paper,we express the deepfake detection task as the fine-grained image classification problem and propose a multi-branch deepfake detection algorithm based on fine-grained features.The algorithm constructs a multi-branch network structure that can focus on the subtleties of different sample regions to learn multi-scale detailed features,which effectively solves the shortage of detection accuracy due to coarse-grained features.We introduce the feature localization module and correlationguided fusion module to complete the forgery traces’localization and detection.In addition,embedding the global attention module in the backbone network enhances the cross-dimensional interaction of spatial and channel information and reduces the influence of irrelevant background regions in the sample.These improved strategies provide a new idea for deepfake detection.

    In terms of detection accuracy,the proposed algorithm in this paper investigates the effectiveness of the improved strategies through a variety of ablation experiments,which has been confirmed to be more comprehensive and accurate in the detection task.

    In future work,we can continue to conduct research from the following two aspects:(1)First,we should explore more efficient fine-grained feature extraction methods.(2) Second,we research how to construct richer deepfake samples to improve the cross-library testing capability of fine-grained features for deepfake detection.

    Acknowledgement:All the face images used in this paper are from publicly available deepfake datasets,including FaceForensics++and Celeb-DF-V2.

    Funding Statement:This work was supported by the 2023 Open Project of Key Laboratory of Ministry of Public Security for Artificial Intelligence Security(RGZNAQ-2304)and the Fundamental Research Funds for the Central Universities of PPSUC(2023JKF01ZK08).

    Author Contributions:Study conception and design,W.K.Qin and T.L.Lu;methodology,W.K.Qin;validation,L.Zhang,S.F.Peng and D.Wan;formal analysis,W.K.Qin and T.L.Lu;investigation,W.K.Qin and S.F.Peng;data collection,W.K.Qin and T.L.Lu;draft manuscript preparation,W.K.Qin;writing review and editing,W.K.Qin and L.Zhang;visualization,W.K.Qin and S.F.Peng;supervision,T.L.Lu and L.Zhang.All authors reviewed the results and approved the final version of the manuscript.

    Availability of Data and Materials:The data used in this paper can be requested from the corresponding author upon request.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    国模一区二区三区四区视频| 宅男免费午夜| 五月玫瑰六月丁香| 欧美zozozo另类| 欧美一级毛片孕妇| 亚洲成人久久爱视频| 国产精品一及| 国产高清激情床上av| 全区人妻精品视频| 欧美最新免费一区二区三区 | 我要搜黄色片| 一个人观看的视频www高清免费观看| 美女cb高潮喷水在线观看| 国产真实乱freesex| 婷婷丁香在线五月| 国产高清三级在线| 成人性生交大片免费视频hd| 少妇人妻精品综合一区二区 | 欧美日韩综合久久久久久 | 我要搜黄色片| 日本一二三区视频观看| 午夜精品久久久久久毛片777| av片东京热男人的天堂| 欧美极品一区二区三区四区| 欧美另类亚洲清纯唯美| 亚洲va日本ⅴa欧美va伊人久久| 午夜免费成人在线视频| 国产精品精品国产色婷婷| 国产精品美女特级片免费视频播放器| 午夜免费男女啪啪视频观看 | xxxwww97欧美| 欧美绝顶高潮抽搐喷水| 国产精品综合久久久久久久免费| 深夜精品福利| 最近最新中文字幕大全免费视频| 无人区码免费观看不卡| 看片在线看免费视频| 在线观看免费视频日本深夜| 18禁国产床啪视频网站| 国产精品 国内视频| 国产69精品久久久久777片| 可以在线观看毛片的网站| 欧美区成人在线视频| 亚洲精品成人久久久久久| 亚洲无线在线观看| 熟女人妻精品中文字幕| 亚洲av中文字字幕乱码综合| 精品国产亚洲在线| 五月玫瑰六月丁香| 成年人黄色毛片网站| 亚洲第一电影网av| 在线观看免费视频日本深夜| 草草在线视频免费看| 69人妻影院| 中出人妻视频一区二区| 欧美区成人在线视频| 他把我摸到了高潮在线观看| 久9热在线精品视频| 一级毛片女人18水好多| 两个人的视频大全免费| 99在线人妻在线中文字幕| 精品欧美国产一区二区三| 久久久久久久亚洲中文字幕 | 岛国视频午夜一区免费看| 俺也久久电影网| 级片在线观看| 久久久久久人人人人人| 久久久国产精品麻豆| 少妇人妻精品综合一区二区 | 国内精品一区二区在线观看| www.色视频.com| 老鸭窝网址在线观看| 亚洲avbb在线观看| 国产一级毛片七仙女欲春2| 国产精品一区二区三区四区久久| 久99久视频精品免费| 国产伦精品一区二区三区四那| 精品电影一区二区在线| 乱人视频在线观看| 亚洲人成网站在线播| 免费无遮挡裸体视频| 久久亚洲精品不卡| 国产欧美日韩一区二区精品| 国产色婷婷99| 欧美日本亚洲视频在线播放| 18禁在线播放成人免费| 亚洲不卡免费看| 黄色日韩在线| 国产av不卡久久| 操出白浆在线播放| 少妇熟女aⅴ在线视频| 久久精品国产自在天天线| 夜夜爽天天搞| 怎么达到女性高潮| 亚洲精品一区av在线观看| 欧美日韩乱码在线| 一本精品99久久精品77| 又粗又爽又猛毛片免费看| 噜噜噜噜噜久久久久久91| 国产亚洲欧美98| 久久精品国产亚洲av涩爱 | 性色av乱码一区二区三区2| 老司机福利观看| av天堂中文字幕网| 国产成年人精品一区二区| 麻豆成人午夜福利视频| 波多野结衣巨乳人妻| 人人妻人人澡欧美一区二区| 中国美女看黄片| 99热这里只有是精品50| ponron亚洲| 一个人免费在线观看的高清视频| 国产成人啪精品午夜网站| 在线a可以看的网站| 在线免费观看的www视频| 99在线视频只有这里精品首页| 啦啦啦观看免费观看视频高清| 香蕉久久夜色| 久久精品国产自在天天线| 欧美成人一区二区免费高清观看| 丁香六月欧美| 3wmmmm亚洲av在线观看| 日日干狠狠操夜夜爽| 内地一区二区视频在线| 国产探花极品一区二区| 欧美一区二区精品小视频在线| 一区福利在线观看| 国产高清视频在线观看网站| 免费大片18禁| 在线观看午夜福利视频| 少妇人妻精品综合一区二区 | 国产高清视频在线观看网站| 男女视频在线观看网站免费| 色视频www国产| 天天一区二区日本电影三级| 深夜精品福利| 亚洲精品456在线播放app | 丁香六月欧美| 亚洲片人在线观看| 操出白浆在线播放| 国产乱人伦免费视频| 欧美黄色片欧美黄色片| 性欧美人与动物交配| 国产乱人伦免费视频| 日本精品一区二区三区蜜桃| 又爽又黄无遮挡网站| 欧美日韩精品网址| 国产欧美日韩一区二区三| 免费观看的影片在线观看| 日日摸夜夜添夜夜添小说| 2021天堂中文幕一二区在线观| 亚洲av熟女| 国产综合懂色| 亚洲乱码一区二区免费版| 亚洲aⅴ乱码一区二区在线播放| 日韩欧美在线二视频| 精品不卡国产一区二区三区| 亚洲精品国产精品久久久不卡| 色视频www国产| 黄色成人免费大全| 欧美性猛交黑人性爽| 欧美一级毛片孕妇| 免费在线观看影片大全网站| 成人午夜高清在线视频| 国产伦精品一区二区三区四那| 2021天堂中文幕一二区在线观| 成人午夜高清在线视频| 美女被艹到高潮喷水动态| 亚洲美女黄片视频| 午夜精品在线福利| 成人国产综合亚洲| 色综合站精品国产| 精品久久久久久成人av| 又粗又爽又猛毛片免费看| 女人十人毛片免费观看3o分钟| 黑人欧美特级aaaaaa片| 午夜影院日韩av| 国产精品 国内视频| 国产探花在线观看一区二区| 欧美av亚洲av综合av国产av| 午夜精品久久久久久毛片777| 国产成+人综合+亚洲专区| 色综合亚洲欧美另类图片| 午夜精品久久久久久毛片777| 久久精品国产亚洲av涩爱 | 一本久久中文字幕| 国产探花极品一区二区| 国产三级黄色录像| 国产av一区在线观看免费| 日韩欧美 国产精品| 午夜亚洲福利在线播放| 怎么达到女性高潮| 久久99热这里只有精品18| 精品一区二区三区视频在线 | 高清毛片免费观看视频网站| 亚洲国产欧美网| 亚洲精品亚洲一区二区| 中文字幕熟女人妻在线| 日韩高清综合在线| 日韩欧美 国产精品| 黄色视频,在线免费观看| 中出人妻视频一区二区| 久久久久免费精品人妻一区二区| 在线观看av片永久免费下载| 最近视频中文字幕2019在线8| 日本黄色片子视频| 国产伦在线观看视频一区| 波多野结衣高清无吗| АⅤ资源中文在线天堂| 午夜福利在线观看吧| 欧洲精品卡2卡3卡4卡5卡区| 18+在线观看网站| 国内少妇人妻偷人精品xxx网站| 深夜精品福利| 嫁个100分男人电影在线观看| 日韩欧美国产在线观看| 动漫黄色视频在线观看| 国产精品久久久久久人妻精品电影| 免费在线观看日本一区| eeuss影院久久| 在线观看午夜福利视频| 两性午夜刺激爽爽歪歪视频在线观看| 久久精品综合一区二区三区| 制服人妻中文乱码| 国产一区二区在线观看日韩 | 国产探花在线观看一区二区| 男人舔女人下体高潮全视频| 久久久久性生活片| 亚洲成人免费电影在线观看| 99国产综合亚洲精品| 中文在线观看免费www的网站| 男人舔奶头视频| 欧美成人a在线观看| 亚洲av中文字字幕乱码综合| 黄片小视频在线播放| 在线看三级毛片| 免费av不卡在线播放| 中文字幕熟女人妻在线| 怎么达到女性高潮| 国产伦精品一区二区三区视频9 | 精品国产超薄肉色丝袜足j| 制服丝袜大香蕉在线| 淫妇啪啪啪对白视频| 搡老岳熟女国产| 国产欧美日韩一区二区精品| 全区人妻精品视频| 亚洲精华国产精华精| 成人精品一区二区免费| 亚洲av免费在线观看| e午夜精品久久久久久久| 国产老妇女一区| 国内毛片毛片毛片毛片毛片| 天堂av国产一区二区熟女人妻| 国产高清激情床上av| 国产日本99.免费观看| 国产黄片美女视频| 亚洲七黄色美女视频| 一个人免费在线观看电影| 国产成人av激情在线播放| 久久精品国产自在天天线| 国产成人aa在线观看| 白带黄色成豆腐渣| 51午夜福利影视在线观看| 国产在线精品亚洲第一网站| 欧美又色又爽又黄视频| 精品久久久久久久毛片微露脸| 51午夜福利影视在线观看| 男人舔奶头视频| 亚洲片人在线观看| 一个人观看的视频www高清免费观看| 亚洲欧美精品综合久久99| 欧美乱码精品一区二区三区| 日本与韩国留学比较| 免费看美女性在线毛片视频| 免费在线观看亚洲国产| 免费看光身美女| 国内少妇人妻偷人精品xxx网站| 国产精品,欧美在线| 欧美av亚洲av综合av国产av| 免费无遮挡裸体视频| 97人妻精品一区二区三区麻豆| 我要搜黄色片| 99热这里只有精品一区| 天天一区二区日本电影三级| 黄色女人牲交| 国产欧美日韩一区二区精品| 精品一区二区三区av网在线观看| 三级国产精品欧美在线观看| 成人特级av手机在线观看| 成人亚洲精品av一区二区| 亚洲专区国产一区二区| 亚洲av免费在线观看| 老司机午夜十八禁免费视频| 国产精品国产高清国产av| 亚洲激情在线av| 毛片女人毛片| 成人精品一区二区免费| 精华霜和精华液先用哪个| 欧美色视频一区免费| 亚洲无线观看免费| 国产精品久久久人人做人人爽| 久久久成人免费电影| 国产精品影院久久| 日本免费a在线| 特大巨黑吊av在线直播| 在线天堂最新版资源| 国产国拍精品亚洲av在线观看 | 观看免费一级毛片| 成人av一区二区三区在线看| 国产精品自产拍在线观看55亚洲| 婷婷丁香在线五月| 亚洲一区二区三区色噜噜| 久久久久久大精品| www国产在线视频色| 亚洲av五月六月丁香网| 中文字幕人妻丝袜一区二区| 97超级碰碰碰精品色视频在线观看| 手机成人av网站| АⅤ资源中文在线天堂| 又爽又黄无遮挡网站| 精品福利观看| 国产成年人精品一区二区| 中文字幕精品亚洲无线码一区| 国产精品乱码一区二三区的特点| 黄片小视频在线播放| 国产亚洲欧美98| 欧美日韩综合久久久久久 | 精品国内亚洲2022精品成人| 一级a爱片免费观看的视频| 国产成人影院久久av| 午夜视频国产福利| 午夜福利欧美成人| 别揉我奶头~嗯~啊~动态视频| 国产精品99久久久久久久久| 宅男免费午夜| 亚洲一区高清亚洲精品| av天堂中文字幕网| 亚洲av日韩精品久久久久久密| 色av中文字幕| 伊人久久大香线蕉亚洲五| 一夜夜www| 男女那种视频在线观看| 两个人视频免费观看高清| 午夜福利18| 嫩草影院入口| av专区在线播放| 啦啦啦免费观看视频1| av女优亚洲男人天堂| 搡老熟女国产l中国老女人| 欧美午夜高清在线| 免费在线观看成人毛片| 麻豆成人av在线观看| 日本精品一区二区三区蜜桃| 色综合欧美亚洲国产小说| 不卡一级毛片| 精品免费久久久久久久清纯| 看免费av毛片| 一a级毛片在线观看| 精品熟女少妇八av免费久了| 国产精品99久久99久久久不卡| 丰满乱子伦码专区| svipshipincom国产片| 女人高潮潮喷娇喘18禁视频| 18禁裸乳无遮挡免费网站照片| 中文字幕高清在线视频| 男女做爰动态图高潮gif福利片| 少妇人妻精品综合一区二区 | 久久久久九九精品影院| 欧美+日韩+精品| 91麻豆精品激情在线观看国产| 一进一出抽搐动态| 日韩国内少妇激情av| 十八禁网站免费在线| 久久草成人影院| 伊人久久大香线蕉亚洲五| 午夜福利在线观看吧| 91九色精品人成在线观看| 丰满的人妻完整版| 在线观看舔阴道视频| 国产蜜桃级精品一区二区三区| 一进一出抽搐动态| 免费av不卡在线播放| av在线天堂中文字幕| 精品一区二区三区av网在线观看| 国产精品香港三级国产av潘金莲| 在线观看免费午夜福利视频| 激情在线观看视频在线高清| 婷婷精品国产亚洲av在线| 亚洲精品在线观看二区| 天天添夜夜摸| 三级毛片av免费| xxxwww97欧美| 在线观看66精品国产| 黄色视频,在线免费观看| 日本一本二区三区精品| 国产精品香港三级国产av潘金莲| 中文字幕精品亚洲无线码一区| 在线a可以看的网站| 男女之事视频高清在线观看| xxxwww97欧美| 两性午夜刺激爽爽歪歪视频在线观看| 国产精品一及| 欧美xxxx黑人xx丫x性爽| 丁香六月欧美| 久久久成人免费电影| 国产成人欧美在线观看| 欧美乱码精品一区二区三区| 色综合婷婷激情| 久久精品夜夜夜夜夜久久蜜豆| 又爽又黄无遮挡网站| 悠悠久久av| 国产高清有码在线观看视频| 免费观看精品视频网站| 国产真人三级小视频在线观看| 亚洲国产色片| 久久国产精品影院| 午夜亚洲福利在线播放| 国产精品电影一区二区三区| 国产私拍福利视频在线观看| 老司机午夜福利在线观看视频| 国内少妇人妻偷人精品xxx网站| 成人av在线播放网站| 此物有八面人人有两片| 婷婷丁香在线五月| 国产精品一及| 国产激情偷乱视频一区二区| av在线天堂中文字幕| 99国产精品一区二区三区| 日韩av在线大香蕉| 亚洲中文字幕日韩| 女警被强在线播放| 亚洲美女黄片视频| 国产久久久一区二区三区| 男女做爰动态图高潮gif福利片| 国产综合懂色| 丰满乱子伦码专区| 天天一区二区日本电影三级| 午夜老司机福利剧场| 身体一侧抽搐| 精品无人区乱码1区二区| 无遮挡黄片免费观看| 亚洲人成网站在线播| 91麻豆精品激情在线观看国产| 夜夜躁狠狠躁天天躁| 久久久国产精品麻豆| 特级一级黄色大片| 中国美女看黄片| 日本黄大片高清| 51国产日韩欧美| 村上凉子中文字幕在线| 日韩欧美在线二视频| 亚洲精品色激情综合| 亚洲国产精品久久男人天堂| 丰满的人妻完整版| 国产欧美日韩精品一区二区| 一个人免费在线观看电影| 一个人看的www免费观看视频| 亚洲五月婷婷丁香| ponron亚洲| 欧美3d第一页| 99久久精品热视频| 99精品在免费线老司机午夜| 乱人视频在线观看| 国产精品1区2区在线观看.| 亚洲av免费高清在线观看| 波多野结衣高清无吗| 婷婷丁香在线五月| 国产成年人精品一区二区| 在线国产一区二区在线| 亚洲黑人精品在线| 国产免费一级a男人的天堂| 亚洲欧美激情综合另类| 男女下面进入的视频免费午夜| 国产高清三级在线| 国产精品自产拍在线观看55亚洲| 亚洲av二区三区四区| 午夜福利视频1000在线观看| 日韩欧美国产一区二区入口| 成熟少妇高潮喷水视频| 亚洲av美国av| 99久久99久久久精品蜜桃| 亚洲aⅴ乱码一区二区在线播放| 亚洲 欧美 日韩 在线 免费| 欧美色欧美亚洲另类二区| 国产主播在线观看一区二区| 成人一区二区视频在线观看| 日韩欧美免费精品| 熟女电影av网| 三级毛片av免费| 一a级毛片在线观看| 国产色爽女视频免费观看| 欧美日韩中文字幕国产精品一区二区三区| 天天一区二区日本电影三级| 亚洲成av人片免费观看| 午夜福利18| 国产在视频线在精品| 久久久久久久午夜电影| 国内精品一区二区在线观看| 禁无遮挡网站| 午夜福利18| 级片在线观看| 久久久精品欧美日韩精品| 白带黄色成豆腐渣| 欧美日韩福利视频一区二区| 一a级毛片在线观看| 欧美一级a爱片免费观看看| 嫩草影院精品99| 亚洲精品一卡2卡三卡4卡5卡| 久久久国产成人免费| 日韩成人在线观看一区二区三区| xxx96com| 18禁在线播放成人免费| 国产视频一区二区在线看| 国产成人av教育| av欧美777| 欧美最黄视频在线播放免费| 香蕉丝袜av| 亚洲成人免费电影在线观看| 国产日本99.免费观看| 色综合亚洲欧美另类图片| 中文字幕高清在线视频| 老鸭窝网址在线观看| ponron亚洲| 亚洲人成网站在线播| 少妇的逼好多水| 欧美绝顶高潮抽搐喷水| 一本一本综合久久| 搡老岳熟女国产| 国内精品久久久久精免费| 国产伦人伦偷精品视频| 伊人久久精品亚洲午夜| 91在线精品国自产拍蜜月 | 天堂影院成人在线观看| 精华霜和精华液先用哪个| 国产视频内射| 婷婷精品国产亚洲av在线| 国内精品久久久久久久电影| 人人妻,人人澡人人爽秒播| 日本撒尿小便嘘嘘汇集6| 国产野战对白在线观看| 欧美极品一区二区三区四区| 男女做爰动态图高潮gif福利片| av黄色大香蕉| 久久精品影院6| 中文字幕av在线有码专区| 国产一区二区三区在线臀色熟女| 18+在线观看网站| 深爱激情五月婷婷| 欧美日本视频| 两个人视频免费观看高清| 亚洲色图av天堂| 久久天躁狠狠躁夜夜2o2o| 国产精品亚洲一级av第二区| 亚洲成av人片免费观看| 亚洲一区二区三区色噜噜| av女优亚洲男人天堂| 亚洲av成人不卡在线观看播放网| 黄色片一级片一级黄色片| 男女午夜视频在线观看| 精品午夜福利视频在线观看一区| 国产主播在线观看一区二区| svipshipincom国产片| 国产一区二区在线观看日韩 | 看黄色毛片网站| 欧美性猛交黑人性爽| 午夜免费男女啪啪视频观看 | 久久久色成人| 亚洲欧美日韩无卡精品| 看免费av毛片| 欧美黑人巨大hd| 亚洲 国产 在线| 国产在视频线在精品| 亚洲av一区综合| 亚洲电影在线观看av| 亚洲精品456在线播放app | 久久欧美精品欧美久久欧美| 99久久综合精品五月天人人| 18禁黄网站禁片免费观看直播| 亚洲久久久久久中文字幕| 一个人看的www免费观看视频| 国产一区二区在线观看日韩 | 黄色丝袜av网址大全| 久久这里只有精品中国| 国产高清videossex| 国产精品一区二区三区四区久久| 成人精品一区二区免费| 免费看美女性在线毛片视频| 网址你懂的国产日韩在线| 禁无遮挡网站| 成人国产一区最新在线观看| 狂野欧美激情性xxxx| 久久精品夜夜夜夜夜久久蜜豆| 中文在线观看免费www的网站| 免费看美女性在线毛片视频| 悠悠久久av| 久久伊人香网站| 乱人视频在线观看| 身体一侧抽搐| 日本三级黄在线观看| 亚洲性夜色夜夜综合| 身体一侧抽搐| 亚洲国产精品久久男人天堂| 听说在线观看完整版免费高清| 午夜福利在线观看免费完整高清在 | 久久午夜亚洲精品久久| 少妇丰满av| 免费av观看视频| 久久午夜亚洲精品久久| 热99在线观看视频| 一本综合久久免费| 99精品久久久久人妻精品| 黄色女人牲交| 内射极品少妇av片p| 成人av在线播放网站| 精品国产美女av久久久久小说|