• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Fake News Classification:Past,Current,and Future

    2023-12-15 03:57:14MuhammadUsmanGhaniKhanAbidMehmoodMouradElhadefandShehzadAshrafChaudhry
    Computers Materials&Continua 2023年11期

    Muhammad Usman Ghani Khan,Abid Mehmood,Mourad Elhadef and Shehzad Ashraf Chaudhry,3,★

    1Department of Computer Science,University of Engineering and Technology,Lahore,54890,Pakistan

    2Department of Computer Science&Information Technology,Abu Dhabi University,Abu Dhabi,59911,United Arab Emirates

    3Department of Software Engineering,Faculty of Engineering and Architecture,Nisantasi University,Istanbul,Turkey

    ABSTRACT The proliferation of deluding data such as fake news and phony audits on news web journals,online publications,and internet business apps has been aided by the availability of the web,cell phones,and social media.Individuals can quickly fabricate comments and news on social media.The most difficult challenge is determining which news is real or fake.Accordingly,tracking down programmed techniques to recognize fake news online is imperative.With an emphasis on false news,this study presents the evolution of artificial intelligence techniques for detecting spurious social media content.This study shows past,current,and possible methods that can be used in the future for fake news classification.Two different publicly available datasets containing political news are utilized for performing experiments.Sixteen supervised learning algorithms are used,and their results show that conventional Machine Learning (ML) algorithms that were used in the past perform better on shorter text classification.In contrast,the currently used Recurrent Neural Network(RNN)and transformer-based algorithms perform better on longer text.Additionally,a brief comparison of all these techniques is provided,and it concluded that transformers have the potential to revolutionize Natural Language Processing(NLP)methods in the near future.

    KEYWORDS Supervised learning algorithms;fake news classification;online disinformation;transformers;recurrent neural network(RNN)

    1 Introduction

    Recent internet advancements have had a considerable impact on social communications and interactions.Social media platforms are being used more and more frequently to obtain information.Additionally,people express themselves through a variety of social media sites.Speedy access to information,low cost,and quick information transmission are just a few of social media’s many advantages.These advantages have led many people to choose social media over more conventional news sources,including television or newspapers,as their preferred method of news consumption.Therefore,social media is replacing traditional news sources quickly.However,social media’s nature can be changed to accomplish different goals[1].One of the reasons that social networks are favored for news access is that it allows for easy commenting and sharing of material with other social media users.Low cost and ease of access are the primary reasons numerous people use social network platforms with rapid access to conventional news sources such as the internet,newsletter,and telecasting.The large volume of internet news data necessitates the development of automated analysis technologies.

    Moreover,recently,during the coronavirus breakdown,the spread of fake news on social networking sites has increased,causing a terrible epidemic worldwide.Fig.1 shows some of the fake news stories circulated on social media during the lockdown1,2,3,4,5https://timesofindia.indiatimes.com/times-fact-check/news/fake-alert-no-russia-does-not-have-lions-roaming-the-streets-to-keep-people-indoors/articleshow/74768135.cms.Emissions from Chinese crematoriums could be visible from space.500 lions are released into the streets of Russia to keep people indoors.In London,doctors are being mugged.The condition can be cured with snake oils or vitamins.How about inhaling a hairdryer’s heated air?Or gargling with garlic water that’s been warmed up?

    Figure 1:Examples of fake news spread on social media

    False information harms people,society,corporations,and governments.The spread of fake news,particularly low-quality news,negatively affects personal and societal beliefs.Spammers or malicious users may distribute false and misleading information that could be very harmful.As a result,identifying fake news has become an essential area of study.Manually identifying and removing fake news or fraudulent reviews from social media takes more effort,money,and time.According to certain prior studies,people perform worse than automated systems when it comes to distinguishing real news from fake news[2].

    ML technologies have been focusing on automatically distinguishing between fake and authentic news for the last few years.Following the 2015 presidential election in the United States,several important social media platforms,including Twitter,Facebook,and Google,focused on developing ML and NLP-based methods to identify and prevent fake news.The extraordinary progress of supervised ML models cleared the path for developing expert systems to detect fake news in English,Portuguese,and other languages [2].Different ML models can have different results on the same classification problem,which is a serious issue [3].Their performance can be affected by corpus features like the size of the corpus and the distribution of instances into classes[3].The performance of the K-Nearby Neighbor (KNN),for example,is determined by the value of (k).Similarly,when handling optimization issues,the Support Vector Machine(SVM)experiences numerical instability[4].

    Various ML algorithms have been utilized in the past to classify fake news.These algorithms are compared against state-of-the-art techniques such as Long Short-Term Memory(LSTM)and Gated Recurrent Unit (GRU),which are currently being used.Transformer models are also experimented with as they are expected to be employed in future fake news classification tasks.This approach enables the evaluation of past techniques.It allows for understanding the current trends in fake news classification and a glimpse into potential future developments in the field.A detection algorithm with two phases has been suggested in this study to detect fake and bogus news on social networking sites.The proposed model is a hybrid of ML algorithms and NLP techniques.Text mining methods are used on the internet news data set in the initial part of this project.Text analysis tools and procedures are designed to extract structured information from raw news data.In the second step,supervised learning algorithms(BayesNet,Logistic Model Tree(LMT),Stochastic Gradient Descent(SGD),decision stump,linear SVM,kernel SVM,Logistic Regression,Decision Tree and Gaussian Discriminant Analysis have been applied to two publicly available Random and Buzzfeed Political News datasets[5].Individual studies employing only a few of these algorithms have been published in the literature.

    Furthermore,they are primarily implemented on a single dataset.In contrast to previous papers,the challenge of detecting fake and fraudulent news has been dealt with and regarded as a classification issue.A wide range of supervised learning algorithms has opted for all two publicly available data sets comprising titles and bodies.The contributions of this research paper are:

    ? We compared the performance of sixteen supervised learning algorithms.

    ? A pipeline for the utilization of transformers on two different datasets.

    ? Analyzed and presented the past,current and future trends of NLP techniques.

    The following is a breakdown of the paper’s structure.The related work is briefly described in Section 2.Details of some of the ML and DL algorithms are described in Section 3.Section 4 contains the details of the methodology and how text preprocessing techniques are applied before utilizing artificial intelligence methods.Section 5 covers datasets and experimental evaluations produced from sixteen supervised artificial intelligence algorithms for two different datasets.Section 5 also describes the results and discussion part.In Section 6,conclusions and future research directions have been examined.

    2 Related Works

    In recent years,detecting rumors and fake news,evaluating web content trustworthiness,and developing fact-checking systems have all been hot subjects.Preprocessing of data can be utilized for the estimation and recovery of various text forms.This includes pre-handling the text utilizing NLP,for example,stemming or lemmatization,standardization,tokenization,and afterward utilization of Term Frequency-Inverse Document Frequency(TF-IDF)[6]for forming words into vectors,Honnibal et al.[7]utilized Spacy for changing words into vectors.Similarly,Mikolov et al.[8]and Pennington et al.[9]used word2vec and Glove for word embeddings.

    Even though the fake news identification problem is only established,it has drawn much attention.Different researchers proposed different methodologies to distinguish fake news in many data types.Reference[10]divided the difficulty of detecting fake news into three categories,i.e.,severe fabrication,comical fake news,and massive scope deception.In [11],Conroy et al.utilized a hybrid technique and proposed a novel detector for fake news.Their proposed methodology[11]incorporates different linguistic cueing and network analysis techniques.In addition to this,they used the vector space model to confirm news [12].In [13] methodology,TF-IDF and SVM were used to categorize news into different groups.In [14],humorous cues were employed to detect false or deceptive news.The authors proposed an SVM-based model and used 360 news articles to evaluate it.To verify the stories,reference[15]found different perspectives on social media.Then,they tested their model against actual data sets.Reference[16]employed ML classifiers such as Decision Tree,K-Nearest Neighbor,Naive Bayes,SVM,and Logistic Regression to classify fake news from online media sources.An ensemble categorization model is suggested in the article[17]for identifying fake news that outperformed the state-of-the-art in terms of accuracy.The recommended approach identified vital characteristics from the datasets.The retrieved characteristics were then categorized using an ensemble model of three well-known ML models:Decision Tree,Random Forest,and Extra Tree Classifier.

    Two categorization models were presented in[18]to address the problem of identifying fake news.One is a Boolean crowd-sourcing approach,while the other is a Logistic Regression model.Aside from this,the preprocessing methods for the problem of false news detection and the creation of assessment measures for data sets have been thoroughly documented [19].Reference [20] employed ML classification techniques and n-gram analysis for classifying spam and fake news.On publicly accessible datasets,the authors assessed their study methods throughout.Gradient boosting,SGD,Random Forests,SVM,and limited Decision Trees were used as classification methods[21].Reference[22]have developed CSI,an algorithm comprises of different characteristics for classifying fake news.Three characteristics were merged in their strategy for a more accurate forecast.Capture,score,and integrate were the three attributes.Reference [23] introduced a tri-relationship false news detection model that considers news attitude,publisher bias,and interactions of users.They evaluated their approach using public datasets for detecting fake news.To classify fake news,the author of [24]suggested a novel hybrid DL model that integrated CNN and RNN.The algorithm was evaluated effectively on two false news datasets,and detection performance was notably superior to previous non-hybrid baseline techniques.

    Reference [25] developed a novel hybrid algorithm based on attention-based LSTM networks for the fake news identification challenge.Evaluation of the method’s performance is conducted on benchmark sets of false news detection data.In early 2017,reference[26]investigated the current state of fake news,provided a remedy for fake news,and described two opposing approaches.Janze et al.[27]developed a detection technique for spotting fake news.The authors of this study evaluated their models on Facebook News during the 2016 presidential election in the United States.Reference[28] developed another automated algorithm.This paper’s authors provided a categorization model based on semantic,syntactic,and lexical information.Reference[29]offered an automated technique for detecting false news in popular Twitter discussions.This approach was tested on three existing data sets.Reference [30] researched the statistical features of misinformation,fraud,and unverified assertions in online social networks.

    Reference[31]developed a competitive model to mitigate the impact of misleading information.The author mainly focused on the interaction between original erroneous and updated information.Reference [32] developed a new algorithm for detecting fake news that considers consumer trust.Reference[33]solved the problem by using a crowded signal.As a result,the authors have presented a novel detective method that uses Bayesian inference and learns the accuracy of users’flagging over time.Reference[34]suggested a content-based false news detection approach.The authors developed a semi-supervised approach for detecting fake news.Reference [35] looked at the different types of social networks and advocated using them to identify and counteract false news on social media.Reference [36] created a model that can identify the truthfulness of Arabic news or claims using a Deep Neural Network (DNN) approach,specifically Convolutional Neural Networks (CNN).The aim was to tackle the fact-checking problem,determining whether a news text claim is authentic or fake.The model achieved an accuracy of 91%,surpassing the performance of previous methods when applied to the same Arabic dataset.Reference[37]discussed the use of Deep Learning(DL)models to detect fake news written in Slovak,using a dataset collected from various local online news sources associated with the COVID-19 epidemic.The DL models were trained and evaluated using this dataset.A bidirectional LSTM network combined with one-dimensional convolutional layers resulted in an average macro F1-score of 94%on an independent test set.

    For accurately identifying misleading information using text sentiment analysis [38] presented“emoratio,”a sentiment scoring algorithm that employs the Linguistic Inquiry Word Count(LIWC)tool’s psychological and linguistic skills.Reference[39]proposed a thorough comparative examination of different DL algorithms,including ensemble methods,CNN,LSTM,and attention mechanisms for fake news identification.The CNN ensembled with bidirectional LSTM using the attention mechanism was found to have the most remarkable accuracy of 88.75%.Another tricky topic of false news classification is the circulation of intentionally generated phony photographs and altered images on social media platforms.The examination was directed by[40]on a dataset of 36,302 picture answers by utilizing both traditional and deep picture forgery techniques for distinguishing fraudulent pictures produced using picture-to-picture transformation based on the Generative Adversarial Networks(GAN) model,a DNN for identifying fake news in its early stages.Reference [41] utilized time and assault for veracity classification[42],style examination of hyperpartisan news[43],are worth focusing on spearheading research in believability investigation on informal organizations.

    Bidirectional Encoder Representations Transformer (BERT) and VGG19,based on a multimodal supervised framework,namely ‘Spotfake’[44],classify the genuine and fictitious articles by utilizing the capacities of encoders and decoders.Moreover,reference[45]used Adversarial training to classify news articles.The purpose of[46]was to develop a model for identifying fake news using the content-based classification approach and focusing on news titles.The model employed a BERT model combined with an LSTM layer.The proposed model was compared to other base classification models,and a standard BERT model was also trained on the same dataset to compare the impact of using an LSTM layer.Results indicated that the proposed model slightly improved accuracy on the datasets compared to the standard BERT model.References [47,48] utilized linear discriminant analysis and KNN for the detection of fake news,even in a vehicular network.The summary of the related work is shown in Table 1.

    Several datasets have been used for fake news detection research.Some of the datasets that have been used in the past are LIAR,FNC-1,and FakeNewsNet datasets.On the other hand,the GossipCop,PolitiFact,and the Fake News Challenge(FNC)datasets are widely used in the current era.It is important to note that many datasets are created to serve a specific research problem;therefore,they might not be generalizable to other scenarios and might not have the same size,type of data,quality,and time coverage.Thus,it is essential to consider these factors when selecting the dataset for a specific task.

    This work uses transformers,RNN,and conventional ML algorithms to classify fake news and provide an in-depth comparison of all these models.Results depict that ML algorithms perform better than complex DL-based models on shorter text.While for longer text,transformers outperform other algorithms.

    3 Machine Learning and Deep Learning

    This section briefly describes the algorithms used in this study’s experiments.Moreover,it is further divided into ML and DL methods.

    3.1 Supervised ML Algorithms

    3.1.1 Linear SVM

    One of the most well-known supervised learning methods,SVM,is used to tackle classification and regression problems.“Linear separable data” refers to information that can be split into two groups by a single straight line.Linear SVM is used to classify such information,and the classifier employed is referred to as a linear SVM classifier.

    3.1.2 Kernel SVM

    When the collection of samples cannot be divided linearly,SVM can be expanded to address nonlinear classification challenges.The data are mapped onto a high-dimensional feature space by applying kernel functions,where linear grouping is conceivable.

    3.1.3 Logistic Regression

    In contrast to Linear Regression,Logistic Regression is used as a classification technique.Logistic regression predicts the outcome by utilizing values of different independent variables.It is undoubtedly one of the most utilized ML techniques.Rather than giving a constant value,it provides the result as a binary,i.e.,valid or invalid,fake and real,yes or no,etc.Its probabilistic value ranges between 0 and 1.

    3.1.4 Naive Bayes

    It is a supervised ML algorithm based on the Bayesian theorem for classification tasks.This classifier posits that features in a class are independent of each other.This type of classifier is relatively easy to construct and is especially good for massive datasets.Naive Bayes outperforms even the most advanced classification systems due to its simplicity.

    3.1.5 Decision Tree

    One of the supervised algorithms based on rules is the Decision Tree.The Decision Tree is used for classification and regression and is a non-parametric method.In the Decision Tree,every node has one of the rules and gives output that is passed to another node,and then another rule-based testing is applied.

    3.1.6 Random Forest

    Random Forest is a supervised algorithm that combines Decision Trees for the different samples and gives results by giving an average from each Decision Tree.It is one of the flexible algorithms that can produce a good result for classification even without tuning.

    3.1.7 Gaussian Discriminant Analysis-Linear

    ML algorithms that directly predict the class from the training set are known as discriminant algorithms.One of the discriminative algorithms applied in our study is Gaussian Discriminant Analysis.Gaussian Discriminant Analysis fits a Gaussian distribution to each class of data separately to capture the distribution of each class.The probability will be high if the predicted value lies at the center of the contour of one of the classes in the training dataset.Linear Discriminant Analysis is a particular type of Quadratic Discriminant Analysis with linear boundaries.

    3.1.8 Gaussian Discriminant Analysis-Quadratic

    ML algorithms that directly predict the class from the training set are known as discriminant algorithms.One of the discriminative algorithms applied in our study is Gaussian Discriminant Analysis.Gaussian Discriminant Analysis fits a Gaussian distribution to each class of data separately to capture the distribution of each class.The probability will be high if the predicted value lies at the center of the contour of one of the classes in the training dataset.

    3.1.9 KNN

    KNN is one of the most well-known and widely utilized supervised learning methods.It works by finding the distance between new data points and comparing it with the number of K points provided as input.The data point is allocated to the class where the distance is minimum.Euclidean distance is one of the distance functions used in KNN.

    3.1.10 Weighted KNN

    It is a specially modified version of KNN.In contrast to traditional KNN,it assigns the highest weight to the points which are near and less weight to those which are far away.Its performance varies with the change in the hyperparameter K.Weighted KNN may produce outliers if the value of K is too small.

    3.2 RNN-Based Algorithms

    3.2.1 Gated Recurrent Units(GRU)

    GRU comprises two gates,i.e.,the update gate and the reset gate.An update gate combines the features of an input gate and a forget gate.A forget gate makes decisions about which data will be discarded and which will be stored.On the other hand,reset gates prevent gradient explosions by erasing previous information.It regulates how much past data must be discarded.

    3.2.2 Long Short-Term Memory(LSTM)

    Each LSTM network has three gates that govern data flow and cells storing data.The data is transmitted from the beginning to the end of the time step by the Cell States.In LSTM,the forget gate determines whether data must be pushed forward or omitted.While in the input gate,upon deciding on the relevant data,the data is sent to the input gate,which carries the data onto the cell states,causing them to be updated.It is as simple as storing and changing the weight.An output gate is triggered when the information has been transferred via the input gate.The output gate produces the hidden phases,and the current condition of the cells is carried forward to the next step.

    3.3 Transformers

    3.3.1 Bidirectional Encoder Representations from Transformers(BERT)

    BERT is an excellent addition to the Transformers community,especially for dealing with longer text.It is a bidirectional encoder-based transformer proposed by Google.BERT currently consists of two versions BERT base and BERT large.For the input,BERT took 512 tokens sequence at once.BERT can take three input embedding types:position embeddings,segment embeddings,and token embeddings.

    3.3.2 ALBERT

    ALBERT is a particularly lite variant of BERT with efficient training speed and fewer parameters as compared to BERT.Because ALBERT uses absolute position embeddings,it is best to pad the right side of the inputs rather than the left.Moreover,the computation cost remains the same as BERT because it has the same number of hidden layers as BERT.

    3.3.3 DeBERTa

    A neural language model based on Transformer called Decoding enhanced BERT with disentangled attention(DeBERTa)trains on enormous amounts of raw text corpora using self-supervised learning.To do numerous NLP jobs in the future,DeBERTa is built to accumulate universal language representations.By utilizing three unique strategies,DeBERTa trumps the previous state-of-the-art BERT.The strategies are as follows:

    ? A precise attention mechanism.

    ? Mask decoder improvement.

    ? A technique for virtual adversarial training that can be fine-tuned.

    3.3.4 RoBERTa

    The architecture of RoBERTa is similar to that of BERT,but it employs a different pre-training strategy and a byte-level BPE as a tokenizer.It extends BERT by changing crucial hyperparameters,such as deleting the following sentence pre-training goal and training with considerably bigger minibatches and learning rates.

    4 Methodology

    This section provides the detail of our methodology for fake news classification.Each step is discussed in sequence.First,duplicated words and unwanted characters,such as numbers,stopwords,dates,time,etc.,are removed from the dataset.Then,feature extraction was performed on the fake news dataset to reduce the feature space.Each word frequency is calculated for the construction of a document term matrix.Sixteen supervised algorithms are applied to the two political news datasets in the final step.Fig.2 shows the whole methodology,and Table 2 shows the specifications of the dataset being utilized in it.

    Table 2: Stats for the dataset

    4.1 Preprocessing for ML Algorithms

    ? Tokenization

    From the word tokenization,it is clear that it is used to make tokens of text by dividing the text into smaller chunks.Punctuation marks are also eradicated from the corpus.Moreover,a number filter has been used to remove those words that contain numeric values,followed by a case converter filter for converting all text from upper to lower case.Lastly,in this step,a filter is used to remove DateTime from the textual data.

    Figure 2:The overall process flow of methodology

    ? Stopwords and line removal

    Stopwords,usually little,are used to join phrases and finish sentences.These are regional language terms that do not transmit knowledge.Pronouns,prepositions,and conjunctions are all examples of stop words.The number of stopwords in the English language is between 400 and 500 [49].Stop words include that,does,a,an,so on,where on,too,above,I until,but,again,what,all,and when.

    ? Stemming

    Stemming is a technique to identify the fundamental forms of words with similar semantics but diverse word forms.This process converts grammatical structures such as a word’s verb,adjective,noun,adverb,and so on into their root form.The words“collect,”“collections,”and“collecting,”for example,all come from the word“collect.”

    The specifics of the preprocessing processes are displayed in Table 3.

    Table 3: Steps for preprocessing data

    4.2 Feature Engineering

    Managing high-dimensional information is the most challenging part of text mining.To increase performance,unrelated and inconsequential qualities should be disposed of from the model.The means of information preprocessing include extracting features from high-layered unstructured data.In this work,stem phrases in informational collections with a recurrence over the edge are tracked down utilizing a feature selection method.Following this technique,each record was changed into a vector of term loads by weighing the terms in its informational index.The Vector Space Model(VSM)is the most direct essential portrayal.VSM assigns a value to each word that indicates the word’s weight in the text.Term frequency is one approach for calculating these weigh.Inverse Document Frequency(IDF)and Term Frequency Inverse Document Frequency(TF-IDF)are the two most well-known of these methods.In this paper,the TF-IDF approach is applied.The TF-IDF approach is used to weigh a phrase in any piece of information and assign value based on the number of times it appears in the document.It also looks at the keyword’s significance across the internet,a technique known as corpus.

    4.3 Evaluation Measures

    The performance of our model is evaluated using precision,accuracy,F1-score,and recall,represented in Eqs.(1)-(4),respectively.

    whereas TPN stands for True Positive News—news that is real and anticipated by the model to be real.TNN stands for True Negative News—fake news projected to be fake by the model.FPN stands for False Positive News or the real fake news that the model incorrectly anticipated to be true.FNN stands for False Negative News;the actual real news projected to be fake by the model.

    5 Results and Discussion

    In this section,dataset and training details are provided.Moreover,the comparison results of RNN,transformers,and ML-based algorithms are also discussed.

    5.1 Experimental Settings

    In this work,two publicly available datasets from the political domain[5]are used.As discussed above,sixteen (RNN,transformers,and conventional ML) algorithms are applied to the title and body text of the dataset.Before applying the algorithms,the dataset is split with a ratio of 70% to 30%,respectively,for training and testing.The TF-IDF is used to form the word-weighted matrix for feature extraction for the conventional ML algorithms.While for the RNN-based algorithm,GloVe vectors are utilized.

    5.1.1 Dataset

    The dataset[5]described in Table 2 is used for the tests.“Buzzfeed News Data and Random News Data”are just two news datasets that are included.48 examples of false news and 53 instances of actual news are included in”Buzzfeed News Data.”O(jiān)n the other hand,the“Random News Data”collection contains 75 instances of satire,true news,and false news.Real news and false news data are both used in this study.Both datasets include the headline and the story’s content,which are utilized separately to classify the dataset.A few examples of these datasets are shown in Table 4.

    Table 4: Instances from Buzzfeed and political news dataset

    Table 5: Results on title(Buzzfeed political news dataset)

    Table 6: Results on body(Buzzfeed political news dataset)

    Table 8: Results on title(Random political news dataset)

    5.1.2 Hyperparameters

    For the DL-based method,i.e.,RNN and GRU,a glove matrix of 300 embedding dimensions and 60 epochs with a batch size of 16 are used.The hidden units are set to 256,which is the number of neurons in the hidden layer.The number of hidden units is chosen based on the task’s complexity and the dataset’s size.A dropout rate of 0.3 is used during the training of the model.This rate is chosen to strike a balance between preventing overfitting and maintaining the model’s ability to capture the relevant information from the data.The optimization algorithm used for training the model is Stochastic Gradient Descent(SGD),a widely used optimization algorithm for training neural networks.To further prevent overfitting,an early stopping strategy is implemented.Moreover,the learning rate is set to 0.0001,which determines the optimization algorithm’s step size in finding the model’s optimal parameters.

    For both datasets,experiments are run 10 times for conventional ML algorithms because there is a massive distinction between the outcomes due to random data selection.After running each traditional algorithm of ML 10 times,the mean value of evaluation measures,i.e.,accuracy,precision,recall,and F1-score,is taken.

    These hyperparameters were chosen through a combination of literature review and experimental tuning,demonstrating that they provided optimal performance for the task.Finally,in addition to the RNNs,the transformers model is trained using BERT embeddings with a dropout rate of 0.2.The dropout rate of 0.2 is used on the BERT embeddings during the fine-tuning process to prevent overfitting.

    5.2 Dataset 1:Buzzfeed Political News Dataset

    The Buzzfeed Political News dataset has been subjected to the recently mentioned supervised ML,RNN,and transformer-based algorithms to determine whether the news is accurate.The features are disengaged from the dataset using TF-IDF.On the dataset for Buzzfeed Political News,Tables 5 and 6 compare the effectiveness of various supervised ML algorithms on the title and body of the Buzzfeed Political News dataset,respectively.Tables 5 and 6 show that in terms of precision,kernel SVM and quadratic Gradient Discriminant Analysis(GDA)perform worst on the title and body of the Buzzfeed Political News dataset,respectively.On the other hand,linear GDA and Random Forest perform best in terms of precision on the title and body text of the Buzzfeed News dataset.Tables 5 and 6 depict that kernel SVM has the worst performance regarding the recall and F1-score linear GDA,Logistic Regression,and Random Forest on title and body text,respectively.It seems that kernel SVM and BERT perform best in terms of recall and F1-score on the title,while kernel SVM and RoBERTa perform best on body text.Regarding accuracy,kernel SVM performs worst on both title and body,while BERT and RoBERTa perform best on the title and body text.Figs.3 and 4 depict a graphical illustration of the algorithm’s performance in terms of accuracy,precision,recall,and F-measure metrics.While Fig.5 shows the comparison of loss on the title and body of the Buzzfeed Political News dataset.

    Figure 3:Comparison of RNN,transformers,and ML-based algorithms on the title text of Buzzfeed political news dataset

    Figure 4:Comparison of RNN,transformers,and ML-based algorithms on the body text of Buzzfeed political news dataset

    Figure 5:Comparison of loss on title and body of Buzzfeed political news dataset

    5.3 Dataset 2:Random Political News Dataset

    This section provides the results of the applied artificial intelligence algorithms with respect to their evaluation measures on both datasets.On the title and body of the Random Political News dataset,Tables 7 and 8 show the outcomes of the various supervised ML algorithms.Figs.6 and 7 visually represent a comparison of sixteen supervised learning algorithms’outputs.In addition to this,Fig.8 shows the comparison of loss on the title and body of the Random Political News dataset.

    In Tables 7 and 8 for the Random Political News dataset,kernel SVM performs worst in terms of precision on both title and body text.While on precision,Decision Tree and linear SVM performance are best of all others.For recall,Decision Tree performance is worst on both title and body.On the other hand,kernel SVM performs best in terms of recall.

    Figure 6:Comparison of RNN,transformers,and ML-based algorithms on the title text of the random political news dataset

    Figure 7: Comparison of RNN,transformers,and ML-based algorithms on the body text of the random political news dataset

    Figure 8:Comparison of loss on title and body of random political news dataset

    For F1-score and accuracy,kernel SVM remains on the lowest performance for title and body text,while DeBERTa performs best on title and body,respectively.

    From the above results and analysis,it is depicted that in the title text of both datasets,the performance of conventional ML algorithms is better than RNN and transformer-based algorithms in terms of computation and evaluation measures.For the longer text,i.e.,transformers outperform the remaining applied algorithms for the body of both datasets.

    Other than this,Table 9 shows the comparison of different algorithms used for the detection of fake news in recent surveys.

    Table 9: Comparison of the different algorithms used in recent studies for fake news detection

    6 Conclusion

    This paper compares supervised learning models for detecting fake news on social media based on NLP techniques and supervised RNN,transformers,and conventional ML algorithms.The accuracy,recall,precision,and F1-measure values for supervised artificial intelligence algorithms are examined.Two datasets are used to determine the average performance of all supervised AI algorithms.From our obtained results,it is clear that ML algorithms perform better on short text classification.It depicts that it is better to use an ML algorithm when the text is one or two lines,and also ML algorithms are efficient in computation.In contrast,longer text transformers outperform the other algorithms.

    In the future,this work could be improved with the advancement in transformers,existing hybridizing techniques,and intelligent optimization algorithms.In addition,we will be looking for multi-modal data(images,videos,audio)to detect fake news.The experiments will be undertaken on a multi-modal dataset to understand better the aspects of fake news identification and how to employ ML algorithms better.

    Acknowledgement:ADU authors acknowledge financial support from Abu Dhabi University’s Office of Research and Grant programs.

    Funding Statement:Abu Dhabi University’s Office of sponsored programs in the United Arab Emirates(Grant Number:19300752)funded this endeavor.

    Author Contributions:All the authors contributed equally.

    Availability of Data and Materials:https://github.com/rpitrust/fakenewsdata1.

    Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

    久久久久免费精品人妻一区二区| 精品免费久久久久久久清纯| 精品福利观看| 欧美3d第一页| 久久九九热精品免费| 久久久久久久久大av| 搞女人的毛片| 精品无人区乱码1区二区| 亚洲综合色惰| 国产极品精品免费视频能看的| 日本一本二区三区精品| 久久久久久久精品吃奶| 一a级毛片在线观看| 91在线观看av| 免费看日本二区| 国产精品精品国产色婷婷| 亚洲精品日韩av片在线观看| 一个人看的www免费观看视频| 国产在视频线在精品| 久久亚洲真实| 国产主播在线观看一区二区| 久久国产精品影院| 乱码一卡2卡4卡精品| 热99re8久久精品国产| 亚洲精品粉嫩美女一区| 麻豆久久精品国产亚洲av| 久久久精品大字幕| 精品人妻熟女av久视频| 嫩草影院新地址| 美女免费视频网站| 日韩精品中文字幕看吧| 亚洲第一电影网av| 午夜福利视频1000在线观看| 国产一区二区亚洲精品在线观看| 成人毛片a级毛片在线播放| 亚洲av中文字字幕乱码综合| 久久99热6这里只有精品| 啪啪无遮挡十八禁网站| 国产高清视频在线观看网站| 亚洲成人久久性| 免费av观看视频| 女人被狂操c到高潮| 99国产综合亚洲精品| av欧美777| 亚洲第一区二区三区不卡| 成人欧美大片| 国产精品伦人一区二区| 狠狠狠狠99中文字幕| 国产91精品成人一区二区三区| 一区二区三区高清视频在线| 国产欧美日韩精品一区二区| 精品无人区乱码1区二区| 69人妻影院| 日韩欧美一区二区三区在线观看| 99久久九九国产精品国产免费| 久久人妻av系列| 精品一区二区三区视频在线| 久久久久久久午夜电影| 免费看a级黄色片| 好看av亚洲va欧美ⅴa在| 老师上课跳d突然被开到最大视频 久久午夜综合久久蜜桃 | 免费人成视频x8x8入口观看| 国产在线男女| 嫩草影院入口| 亚洲18禁久久av| 12—13女人毛片做爰片一| 中文字幕久久专区| 亚洲aⅴ乱码一区二区在线播放| 亚洲成人中文字幕在线播放| 性色av乱码一区二区三区2| 久久久久久九九精品二区国产| 色尼玛亚洲综合影院| av女优亚洲男人天堂| 精品久久久久久久久av| 久久精品国产自在天天线| 成人特级黄色片久久久久久久| 天堂网av新在线| 亚洲欧美日韩高清专用| 男插女下体视频免费在线播放| 久久久精品欧美日韩精品| 国产欧美日韩一区二区三| 亚洲电影在线观看av| 亚洲av中文字字幕乱码综合| 国产欧美日韩一区二区精品| 最近视频中文字幕2019在线8| 国产精品亚洲美女久久久| 热99re8久久精品国产| 国产高清视频在线观看网站| 2021天堂中文幕一二区在线观| 黄色一级大片看看| 淫妇啪啪啪对白视频| 桃色一区二区三区在线观看| 在线观看午夜福利视频| 露出奶头的视频| 亚洲美女视频黄频| 好男人电影高清在线观看| 一个人看的www免费观看视频| 国产精品野战在线观看| 精品日产1卡2卡| 一级作爱视频免费观看| а√天堂www在线а√下载| 少妇的逼好多水| 一a级毛片在线观看| 国内精品美女久久久久久| 国产在线男女| 国产精品久久久久久久久免 | 一级作爱视频免费观看| 偷拍熟女少妇极品色| 免费电影在线观看免费观看| 免费看a级黄色片| 人妻制服诱惑在线中文字幕| 99热这里只有是精品50| 国产精品美女特级片免费视频播放器| 天美传媒精品一区二区| 五月玫瑰六月丁香| 亚洲精品成人久久久久久| 日韩免费av在线播放| 亚洲经典国产精华液单 | 亚洲国产精品999在线| 在线a可以看的网站| 国产精品99久久久久久久久| 欧美极品一区二区三区四区| 亚洲成av人片免费观看| 毛片一级片免费看久久久久 | 亚洲色图av天堂| 国产精品野战在线观看| 天堂影院成人在线观看| 亚洲男人的天堂狠狠| 精品午夜福利视频在线观看一区| 久久99热6这里只有精品| 美女cb高潮喷水在线观看| 性插视频无遮挡在线免费观看| 欧美日韩乱码在线| 久久久精品欧美日韩精品| www.999成人在线观看| 黄片小视频在线播放| 最新中文字幕久久久久| 精品久久久久久,| 麻豆成人午夜福利视频| 久久天躁狠狠躁夜夜2o2o| 51午夜福利影视在线观看| 美女xxoo啪啪120秒动态图 | 久久久精品大字幕| 国产久久久一区二区三区| 精品午夜福利视频在线观看一区| 99热6这里只有精品| 日韩欧美精品免费久久 | 国产 一区 欧美 日韩| 国产精品自产拍在线观看55亚洲| 亚洲一区二区三区不卡视频| 欧美高清性xxxxhd video| 国产精品伦人一区二区| 日韩av在线大香蕉| 日韩欧美三级三区| 女同久久另类99精品国产91| 淫妇啪啪啪对白视频| 美女 人体艺术 gogo| 成人特级av手机在线观看| 国产伦精品一区二区三区视频9| 特级一级黄色大片| 18禁黄网站禁片午夜丰满| 男女之事视频高清在线观看| 俄罗斯特黄特色一大片| av中文乱码字幕在线| 日韩 亚洲 欧美在线| 看黄色毛片网站| 精品一区二区三区视频在线| 亚洲av.av天堂| 国产精品亚洲美女久久久| 精品午夜福利视频在线观看一区| 成人特级av手机在线观看| 神马国产精品三级电影在线观看| 亚洲精品乱码久久久v下载方式| 久久久久久大精品| 看片在线看免费视频| 国产精品伦人一区二区| 男女下面进入的视频免费午夜| 久久久精品大字幕| 午夜精品久久久久久毛片777| 一个人免费在线观看电影| 三级国产精品欧美在线观看| 久久国产精品人妻蜜桃| 国产精品嫩草影院av在线观看 | 国产精品一区二区性色av| 欧美潮喷喷水| 毛片女人毛片| 十八禁网站免费在线| 国产av在哪里看| 欧美最黄视频在线播放免费| 99热6这里只有精品| 一区二区三区四区激情视频 | 日韩有码中文字幕| 国产三级黄色录像| 日本与韩国留学比较| 国产精品不卡视频一区二区 | 国产成人av教育| 91在线精品国自产拍蜜月| 国产91精品成人一区二区三区| 在线观看av片永久免费下载| 亚洲色图av天堂| 久久精品国产亚洲av天美| 亚洲精品久久国产高清桃花| 亚洲国产欧美人成| 国内揄拍国产精品人妻在线| 亚洲中文字幕一区二区三区有码在线看| 欧美激情国产日韩精品一区| 直男gayav资源| 亚洲人与动物交配视频| 欧美xxxx黑人xx丫x性爽| 国产成人福利小说| 精华霜和精华液先用哪个| 亚洲电影在线观看av| 欧美激情国产日韩精品一区| 久久久精品大字幕| 黄色女人牲交| 国产亚洲精品久久久久久毛片| 精品日产1卡2卡| 午夜老司机福利剧场| 国产精品爽爽va在线观看网站| av欧美777| 成人av一区二区三区在线看| 欧美最新免费一区二区三区 | 老司机福利观看| 国产伦人伦偷精品视频| 日韩欧美三级三区| 国产av在哪里看| 在线观看66精品国产| 色综合婷婷激情| 亚洲久久久久久中文字幕| 91在线精品国自产拍蜜月| 精品午夜福利在线看| 最好的美女福利视频网| 欧美三级亚洲精品| 亚洲午夜理论影院| 在线播放国产精品三级| 亚洲欧美日韩高清专用| 免费人成视频x8x8入口观看| or卡值多少钱| 日韩欧美三级三区| 女生性感内裤真人,穿戴方法视频| 亚洲自拍偷在线| 1000部很黄的大片| 99国产精品一区二区三区| 欧美xxxx性猛交bbbb| 91在线精品国自产拍蜜月| 哪里可以看免费的av片| 色在线成人网| 国内精品美女久久久久久| 波多野结衣巨乳人妻| 国产 一区 欧美 日韩| 很黄的视频免费| 最后的刺客免费高清国语| 深夜精品福利| 午夜福利在线在线| 中文亚洲av片在线观看爽| 欧美激情久久久久久爽电影| 日韩大尺度精品在线看网址| 亚洲第一欧美日韩一区二区三区| 禁无遮挡网站| 亚洲av熟女| 久久久久久久久久黄片| 亚洲国产精品sss在线观看| 亚洲人成网站在线播| 亚洲av电影不卡..在线观看| 亚洲国产色片| 99riav亚洲国产免费| av视频在线观看入口| 午夜a级毛片| 日韩精品中文字幕看吧| 天堂网av新在线| 国产v大片淫在线免费观看| 欧美成人a在线观看| 一二三四社区在线视频社区8| 18禁黄网站禁片免费观看直播| 悠悠久久av| 色av中文字幕| 听说在线观看完整版免费高清| 亚洲精品成人久久久久久| 波野结衣二区三区在线| 午夜日韩欧美国产| 最近最新中文字幕大全电影3| bbb黄色大片| 成人无遮挡网站| 色哟哟·www| 色吧在线观看| 午夜福利欧美成人| 深夜a级毛片| 999久久久精品免费观看国产| 又黄又爽又免费观看的视频| 国产精品不卡视频一区二区 | 一级黄片播放器| 中国美女看黄片| 亚洲熟妇中文字幕五十中出| 99国产综合亚洲精品| 有码 亚洲区| 琪琪午夜伦伦电影理论片6080| 一二三四社区在线视频社区8| 免费观看人在逋| 听说在线观看完整版免费高清| 久久久久久久午夜电影| 国产不卡一卡二| 成人精品一区二区免费| 成人一区二区视频在线观看| 男人狂女人下面高潮的视频| 日韩欧美国产在线观看| 国产 一区 欧美 日韩| 直男gayav资源| 午夜福利成人在线免费观看| 亚洲av成人精品一区久久| 欧美国产日韩亚洲一区| 欧美日韩福利视频一区二区| 日本 欧美在线| 亚洲aⅴ乱码一区二区在线播放| 亚洲aⅴ乱码一区二区在线播放| 草草在线视频免费看| 日本精品一区二区三区蜜桃| 久久久成人免费电影| 97碰自拍视频| 少妇被粗大猛烈的视频| 久久久久久大精品| 丰满乱子伦码专区| 亚洲国产欧美人成| 深爱激情五月婷婷| 中亚洲国语对白在线视频| 日本免费a在线| 欧美性猛交╳xxx乱大交人| 黄片小视频在线播放| 欧美性感艳星| 国产精品爽爽va在线观看网站| 久久久精品欧美日韩精品| 90打野战视频偷拍视频| 国产精品一区二区三区四区久久| 91在线精品国自产拍蜜月| 99国产精品一区二区蜜桃av| 天天躁日日操中文字幕| 国产一区二区激情短视频| 成人永久免费在线观看视频| 欧美黄色片欧美黄色片| 日韩有码中文字幕| 精品人妻偷拍中文字幕| 在线观看66精品国产| 午夜两性在线视频| 成人特级黄色片久久久久久久| 亚洲午夜理论影院| 国产精品亚洲一级av第二区| 亚洲 欧美 日韩 在线 免费| 午夜福利在线观看免费完整高清在 | 国产精品永久免费网站| 久久99热6这里只有精品| 一级av片app| 欧美中文日本在线观看视频| 欧美高清性xxxxhd video| 亚洲avbb在线观看| 午夜福利成人在线免费观看| 蜜桃久久精品国产亚洲av| 国产精品野战在线观看| 18+在线观看网站| 亚洲欧美日韩高清在线视频| 亚洲在线自拍视频| 无人区码免费观看不卡| 尤物成人国产欧美一区二区三区| 日日夜夜操网爽| 国产av麻豆久久久久久久| 精品久久久久久久久久免费视频| 国产精品一区二区三区四区免费观看 | 男女做爰动态图高潮gif福利片| 哪里可以看免费的av片| 欧美色视频一区免费| 无人区码免费观看不卡| 日本撒尿小便嘘嘘汇集6| 真实男女啪啪啪动态图| 国产老妇女一区| 亚洲 欧美 日韩 在线 免费| 两个人的视频大全免费| 美女大奶头视频| 无遮挡黄片免费观看| 日韩欧美 国产精品| 国产精品一区二区三区四区免费观看 | 在线免费观看的www视频| 欧美精品国产亚洲| 亚洲五月天丁香| 亚洲欧美日韩卡通动漫| 亚洲五月婷婷丁香| 免费在线观看亚洲国产| 蜜桃亚洲精品一区二区三区| 亚洲第一电影网av| 99热精品在线国产| 国产成+人综合+亚洲专区| 亚洲精品在线美女| 成人永久免费在线观看视频| 国产精品一区二区三区四区免费观看 | 中文字幕精品亚洲无线码一区| x7x7x7水蜜桃| 亚洲欧美日韩高清在线视频| 国产精品人妻久久久久久| 少妇的逼水好多| 国产精品亚洲一级av第二区| 欧美日本亚洲视频在线播放| 97人妻精品一区二区三区麻豆| 91麻豆精品激情在线观看国产| 美女免费视频网站| 亚洲av.av天堂| 久久中文看片网| 久久久久久久久久黄片| 噜噜噜噜噜久久久久久91| 嫩草影院入口| 别揉我奶头~嗯~啊~动态视频| 中文字幕人妻熟人妻熟丝袜美| 成人精品一区二区免费| 精品一区二区三区av网在线观看| 精品久久久久久成人av| 亚洲在线观看片| 国产精品国产高清国产av| 国产老妇女一区| 精品福利观看| 麻豆久久精品国产亚洲av| 日本成人三级电影网站| 1000部很黄的大片| 久久精品国产亚洲av涩爱 | 精品一区二区三区视频在线| 老司机午夜十八禁免费视频| 国产精华一区二区三区| 久久九九热精品免费| 欧美日韩福利视频一区二区| x7x7x7水蜜桃| 中文字幕免费在线视频6| 最近最新中文字幕大全电影3| 免费黄网站久久成人精品 | 一本一本综合久久| 一本一本综合久久| 热99re8久久精品国产| 久久精品国产亚洲av天美| 内地一区二区视频在线| 免费观看精品视频网站| 成人一区二区视频在线观看| 亚洲 欧美 日韩 在线 免费| 亚洲专区中文字幕在线| 99久久精品热视频| 18禁在线播放成人免费| 禁无遮挡网站| 中文字幕免费在线视频6| 九色国产91popny在线| 韩国av一区二区三区四区| 亚洲熟妇熟女久久| www日本黄色视频网| 亚洲无线在线观看| 亚洲无线在线观看| 国内毛片毛片毛片毛片毛片| 真人一进一出gif抽搐免费| 久久久国产成人精品二区| 99久久成人亚洲精品观看| 欧美高清成人免费视频www| 国产午夜福利久久久久久| 久久精品夜夜夜夜夜久久蜜豆| www.999成人在线观看| 超碰av人人做人人爽久久| 久久久色成人| 一进一出抽搐动态| 久久久精品大字幕| 三级男女做爰猛烈吃奶摸视频| 国产男靠女视频免费网站| 国产国拍精品亚洲av在线观看| 免费在线观看影片大全网站| h日本视频在线播放| 深夜a级毛片| 亚洲av日韩精品久久久久久密| 国产成人影院久久av| 国产成人欧美在线观看| 午夜福利成人在线免费观看| 黄色一级大片看看| 久久久久久九九精品二区国产| 男女做爰动态图高潮gif福利片| 十八禁人妻一区二区| 久久久久精品国产欧美久久久| 国产黄片美女视频| 欧美在线黄色| 久久性视频一级片| 99精品在免费线老司机午夜| 男人舔奶头视频| 精品久久久久久久久av| 成人鲁丝片一二三区免费| 亚洲va日本ⅴa欧美va伊人久久| 色综合站精品国产| 久久久色成人| 欧美最黄视频在线播放免费| 免费看a级黄色片| 成人午夜高清在线视频| 中文字幕人妻熟人妻熟丝袜美| 国产毛片a区久久久久| 长腿黑丝高跟| 成人无遮挡网站| 小说图片视频综合网站| 毛片一级片免费看久久久久 | 欧美极品一区二区三区四区| 日韩精品青青久久久久久| 99在线视频只有这里精品首页| 在线天堂最新版资源| 免费人成视频x8x8入口观看| 成人永久免费在线观看视频| 极品教师在线免费播放| 怎么达到女性高潮| 国产精品三级大全| 中文字幕久久专区| 欧美午夜高清在线| 国产精品久久电影中文字幕| 亚洲精品日韩av片在线观看| 久久久国产成人精品二区| 天天一区二区日本电影三级| 极品教师在线免费播放| 国产色婷婷99| 99久久精品热视频| 亚洲人成电影免费在线| 国产黄片美女视频| 亚洲av第一区精品v没综合| 小蜜桃在线观看免费完整版高清| 丰满的人妻完整版| 中文字幕免费在线视频6| 18禁黄网站禁片免费观看直播| xxxwww97欧美| 亚洲真实伦在线观看| 午夜亚洲福利在线播放| 国产伦一二天堂av在线观看| 国产成人影院久久av| 一个人看的www免费观看视频| 69av精品久久久久久| 特级一级黄色大片| 听说在线观看完整版免费高清| 午夜免费男女啪啪视频观看 | 国产探花极品一区二区| 露出奶头的视频| 精品国内亚洲2022精品成人| 亚洲自偷自拍三级| 又爽又黄无遮挡网站| 美女黄网站色视频| 99久久成人亚洲精品观看| 精品国产三级普通话版| 亚洲av电影不卡..在线观看| 欧美日本亚洲视频在线播放| 久久久色成人| 免费无遮挡裸体视频| 99久久无色码亚洲精品果冻| 亚洲国产精品合色在线| 欧美成人一区二区免费高清观看| 在线观看66精品国产| 老鸭窝网址在线观看| 日本熟妇午夜| 女人十人毛片免费观看3o分钟| 搡老熟女国产l中国老女人| 他把我摸到了高潮在线观看| 婷婷丁香在线五月| 美女大奶头视频| 极品教师在线视频| 久久久久亚洲av毛片大全| 亚洲久久久久久中文字幕| 国产精品三级大全| h日本视频在线播放| 国产美女午夜福利| 久久精品国产亚洲av香蕉五月| 真人做人爱边吃奶动态| 色精品久久人妻99蜜桃| 久久久久亚洲av毛片大全| 久久久久精品国产欧美久久久| 亚洲最大成人av| 在线观看免费视频日本深夜| 久久99热6这里只有精品| 国产精品免费一区二区三区在线| 欧美日韩国产亚洲二区| 亚洲最大成人手机在线| 一级a爱片免费观看的视频| 亚洲三级黄色毛片| 色播亚洲综合网| 好男人电影高清在线观看| 直男gayav资源| 嫩草影院精品99| 日韩中文字幕欧美一区二区| 性欧美人与动物交配| 国产精品一区二区三区四区久久| 欧美精品啪啪一区二区三区| 日韩欧美免费精品| 亚洲欧美日韩东京热| 亚洲国产精品成人综合色| 亚洲va日本ⅴa欧美va伊人久久| 麻豆国产av国片精品| 99久久久亚洲精品蜜臀av| 亚洲人与动物交配视频| 中文字幕av成人在线电影| 欧美在线一区亚洲| 日韩免费av在线播放| 免费在线观看成人毛片| 日本与韩国留学比较| 中出人妻视频一区二区| 五月伊人婷婷丁香| 看片在线看免费视频| 我要看日韩黄色一级片| 国产伦在线观看视频一区| 男女之事视频高清在线观看| 亚洲午夜理论影院| 亚洲最大成人手机在线| 两个人的视频大全免费| 在线免费观看不下载黄p国产 | 日本熟妇午夜| 毛片女人毛片| 国产精品综合久久久久久久免费| 国产精品美女特级片免费视频播放器| 国产高潮美女av| 国产精品三级大全| 亚洲,欧美,日韩| 99国产精品一区二区蜜桃av| 99久久九九国产精品国产免费| 少妇的逼好多水| 又爽又黄a免费视频| 天美传媒精品一区二区| 久久久久国产精品人妻aⅴ院| 一个人观看的视频www高清免费观看| 在线观看免费视频日本深夜|