• 
    

    
    

      99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

      DNEF:A New Ensemble Framework Based on Deep Network Structure

      2024-01-12 03:48:44SiyuYangGeSongYuqiaoDengChangyuLiuandZhuoyuOu
      Computers Materials&Continua 2023年12期

      Siyu Yang ,Ge Song,? ,Yuqiao Deng ,Changyu Liu and Zhuoyu Ou

      1College of Mathematics and Informatics,South China Agricultural University,Guangzhou,510642,China

      2School of Statistics and Mathematics,Guangdong University of Finance and Economics,Guangzhou,510120,China

      ABSTRACT Deep neural networks have achieved tremendous success in various fields,and the structure of these networks is a key factor in their success.In this paper,we focus on the research of ensemble learning based on deep network structure and propose a new deep network ensemble framework(DNEF).Unlike other ensemble learning models,DNEF is an ensemble learning architecture of network structures,with serial iteration between the hidden layers,while base classifiers are trained in parallel within these hidden layers.Specifically,DNEF uses randomly sampled data as input and implements serial iteration based on the weighting strategy between hidden layers.In the hidden layers,each node represents a base classifier,and multiple nodes generate training data for the next hidden layer according to the transfer strategy.The DNEF operates based on two strategies:(1)The weighting strategy calculates the training instance weights of the nodes according to their weaknesses in the previous layer.(2) The transfer strategy adaptively selects each node’s instances with weights as transfer instances and transfer weights,which are combined with the training data of nodes as input for the next hidden layer.These two strategies improve the accuracy and generalization of DNEF.This research integrates the ensemble of all nodes as the final output of DNEF.The experimental results reveal that the DNEF framework surpasses the traditional ensemble models and functions with high accuracy and innovative deep ensemble methods.

      KEYWORDS Machine learning;ensemble learning;deep ensemble;deep network structure;classification

      1 Introduction

      Nowadays,machine learning and artificial intelligence have been applied to various fields with remarkable success.Ensemble learning is of great interest among machine learning and Artificial intelligence methods,because of its ability to combine numerous learners to achieve significantly improving the accuracy and generalization performance of the model.Dietterich demonstrated the performance advantages of ensemble learning over a single classifier in statistical,computational,and representational ways [1].In addition,base classifiers in the ensemble model are typically weak classifiers,which are more trainable than a single strong classifier.

      The main challenges in designing ensemble models are the diversity of sub-learners and the ensemble strategies.(1) The diversity of sub-learners: diversity prompts the provision of complementary information among sub-learners[2] and is more advantageous in improving the performance of the ensemble model [3].To maintain the diversity of sub-learners,most ensemble algorithms construct different types of sub-learners on the same instance space or the same type of sub-learners on different instance space (e.g.,ensemble methods based on bagging,including Random Forest (RF)[4],Bagging of Extrapolation Borderline-SMOTE SVM(BEBS)[5],etc...)(2)Ensemble strategies for base classifiers.Typical ensemble strategies include boosting[6]and the base classifier weight-based strategy.Boosting methods improve the classification ability of misclassified instances by adjusting the instance weights to obtain a higher performance of the ensemble model.The sub-classifier weightbased strategy improves the performance of the ensemble model by adjusting the weights of the base classifiers.

      Traditional algorithms often adopt a single strategy,ignoring the complex representation of the ensemble model.In this paper,we propose the Deep Network Ensemble Framework (DNEF),which draws on the concept of deep learning and the structure of the neural network.The DNEF framework adopts a hierarchical network structure that combines various ensemble strategies to improve both the diversity of base classifiers and the generalization ability of the ensemble framework.The novelty of the proposed DNEF includes: (1) Construction of different types of base classifiers on various instance subspaces to maximize the diversity of base classifiers in the hidden layer.(2) Combining the boost ensemble strategy and the metric-based ensemble strategy to optimize instance distribution,improve generalization capability,determine base classifier weights,and transfer misclassified instances between the hidden layers.(3) Adoption of an intra-layer parallel inter-layer serial training model to enhance training efficiency.(4) Comparative experiments show that our proposed DNEF achieves significant improvement in classification accuracy compared to popular ensemble algorithms.

      2 Related Works

      Our approach is related to ensemble learning.We briefly describe the related work including the present ensemble learning and deep ensemble learning in the following sub-sections.

      2.1 Traditional Ensemble Learning

      Ensemble learning is one of the highly effective learning algorithms in machine learning.The procedure of ensemble learning involves combining multiple weak learners according to a certain strategy into a strong learner with higher predictive performance[7].Compared to individual algorithms,it has been found that ensemble learning could combine the advantages of each weak learner to represent complex models more accurately,reducing overfitting and providing higher learning performance.Thus,ensemble learning has a wide range of applications for complex data models such as loan approval[8],speech recognition[9],image recognition[10],and industrial process simulation[11].

      As mentioned above,the main challenges in designing ensemble models are the diversity of base classifiers and the ensemble strategy.Several typical Ensemble Learning models are described below.

      2.1.1 Random Forest

      Random Forest is a model for classification and regression that combines the“bagging”method with random feature selection.Random Forest employs the bagging method to select instances to train different random decision trees,which are then combined based on an ensemble strategy.

      Current improvements to Random Forest have focused on increasing the diversity of the trees by adding random factors (Menze [12],Zhang et al.[13–15]) and on designing ensemble strategies.For example,Utkin et al.proposed optimizing weights in terms of forest accuracy [16,17].Random Forest has been widely used in different classification scenarios in recent years.In the industrial scenario,Paul et al.used the improved Random Forest for the classification of dual-phase(DP)steel microstructures[18].In Intrusion Detection Systems,Resende et al.proposed that the Random Forest has the advantage of speed and efficiency over common machine learning methods[19].

      2.1.2 AdaBoost

      AdaBoost[20]is a highly popular boosting learning algorithm for improving classification accuracy.It iteratively constructs multiple weak classifiers,with each classifier improving its performance by adjusting weights.During each iteration,AdaBoost increases the weights of the instances that were misclassified in the previous iteration,directing the classifier’s attention towards these instances in subsequent rounds,ultimately leading to higher accuracy.In the realm of chemical detection,Chen et al.found that AdaBoost can be used simultaneously to determine the trace copper and cobalt in high-concentration zinc solution[21].In the field of electrocardiogram(ECG)classification,Barstu?gan et al.employed AdaBoost as a classifier based on a dictionary,using it to classify ECG signals[22].

      2.1.3 Stacking

      Stacking[23–25]is a hierarchical ensemble learning framework.Stacking methods learn some base classifiers(usually,these learners are of different types)using the initial training data and employ the predictions generated by these base classifiers as a new training set to train a new classifier.Stacking can improve accuracy by reducing the bias in data.Stacking technology has made significant progress in a large number of areas of application.In mining engineering,Koopialipoor et al.predicted rock deformation using the structure of a stacked tree,Random Forest(RF),K-Nearest-Neighbors(KNN),and Multilayer Perceptron(MLP)[26].For diabetes identification,Kalagotla et al.developed a novel stacking technique with a multi-layer perceptron,support vector machine,and logistic regression to predict diabetes[27].

      2.1.4 Weighted-Based Voting

      The voting strategy,as a decision rule,is important in ensemble learning.Typically,the predictions are made through majority voting.The voting strategy can be generally divided into weighted and unweighted voting.Among them,weighted voting is a commonly employed approach.Weighted voting can be further divided into dynamic weighted voting [28] and static weighted voting.Voting technology has become well-established in a variety of classification scenarios.In the domain of tuberculosis prediction,Okezie et al.used the weighted voting ensemble technique to improve the accuracy of tuberculosis diagnosis [29].For Business failure prediction,Kim et al.used a majority voting ensemble method with a decision tree to predict business failure[30].

      2.2 Deep Ensemble Learning

      After the significant development of deep learning,there has been a widespread focus on how to combine deep learning and ensemble learning to leverage their advantages.Deep ensemble learning encompasses the design of ensemble models and fusion mechanisms.By aggregating multiple deep learning models,the hierarchical relationships between these models are used to achieve more powerful feature extraction and representation capabilities.In this section,we begin by introducing ensemble learning based on neural network modules,followed by an introduction to deep forest.

      Sivakumar et al.proposed a deep learning-based graph neural network (DL-GNN) based on the ensemble of recurrent neural networks (RNN) and feedforward neural networks (FNN) [31].Devarajan et al.introduced a deep learning model integrated with natural language processing (NDCBL),which combines a convolutional neural network (CNN),Bidirectional Long Short-Term Memory (Bi-LSTM),and Attention [32].José et al.presented an automated deep learning-based breast cancer diagnosis model (ADL-BCD) for breast cancer diagnosis using digital mammograms.ADL-BCD is an ensemble model of Gaussian filter(GF)-based preprocessing,Tsallis entropy-based segmentation,ResNet34 based feature extraction,chimp optimization algorithm(COA)based parameter tuning,and the wavelet neural network(WNN)based classification[33].Hussain et al.developed an Ensemble Deep-Learning-Enabled Clinical Decision Support System for Breast Cancer Diagnosis(EDLCDS-BCD).The EDLCDS-BCDC ensemble of VGG-16,VGG-19,and SqueezeNet for feature extraction[34].

      Zhou et al.[35]proposed the Deep Forest(DF)model,which is a multi-grained cascade structure.DF typically comprises two steps: the multi-grained scanner and the cascade forest.The former extracts information from raw data while the latter constructs an adaptive-depth ensemble model.DF is a model that combines ensemble learning and deep learning.On the one hand,it is an ensemble learning method based on the decision tree,inheriting most of the advantages of general ensemble learning methods.On the other hand,DF also has the advantages of deep neural networks to increase the diversity of ensemble learning and improve model performance.Subsequently,the deep forest methods have been improved,such as siamese deep forest[36],imprecise deep forest for classification[37],and applied to various domains,including software defect prediction,price prediction,prediction of protein interactions,anti-cancer drug response prediction[38–41].

      2.3 Characteristics of DNEF

      Our deep ensemble framework,DNEF,combines the structure of deep networks with ensemble learning.Specifically,we use a weighting strategy to facilitate the iteration of DNEF and a transfer strategy to ensure the architectural complexity of the DNEF.Popular ensemble models often employ a single strategy,such as RF,AdaBoost,etc.,while DNEF employs two ensemble strategies:the weighting strategy and the transfer strategy.These two strategies take into account both the relationship between classifiers and the complexity of the ensemble structure.Compared to the deep ensemble models based on the neural network,DNEF does not incorporate deep learning modules and does not use neural networks and backpropagation.The weighting strategy in DNEF supports the iteration of the hidden layer,while the transfer strategy enhances the connectivity of the hidden layer.These key differences represent the biggest differentiating factor between DNEF and deep ensemble learning.

      3 The Proposed Method

      In this section,we introduce our proposed DNEF framework.We first describe the DNEF in Section 3.1.We then explain the weighting strategy and transfer strategy of DNEF in Section 3.2.

      3.1 Introduction of DNEF

      Inspired by deep neural networks,we propose the DNEF framework,which employs a deep network architecture.From Fig.1,the DNEF framework includes an input layer,hidden layers,and an output layer.The data in the input layer consists of randomly sampled instances X0.The DNEF framework iterates between hidden layers based on a weighting strategy.Each hidden layer containsmnodes,and each node represents a base classifierh(in this paper,we use the decision tree as the base classifier).The inputs to the node are instances with their weights.In the hidden layer interior,each node generates transfer instances and corresponding weights based on the transfer strategy.These transfer instances and transfer instance weights of all nodes are aggregated with the input data to produce the output data of the next layer.The output of DNEF is obtained through a weighted ensemble of classifiers across all layers.

      Figure 1:The architecture of DNEF

      It is worth noting that the“ensemble”in this DNEF is based on a deep network structure.Two factors contribute to the accuracy and robustness of DNEF:Firstly,the hidden layer of DNEF builds more accurate base classifiers by utilizing the results of the previous layer.Secondly,independent classifiers are trained in parallel in the hidden layer to reduce variance.

      It is also worth mentioning that the key issues are the calculation of instance weights and the transfer strategy that directly affect the performance of the DNEF.As one of the main contributions of this paper,we propose the weighting strategy and the transfer strategy.

      The procedures for the weighting strategy and transfer strategy are described below:

      1.The procedure of weighting strategy:

      (1) Calculation of the base classifier weightsWc: We compute the weight of the base classifier based on its current class errore.

      (2)Calculation of instance weightsWs:The instance weightsWsare calculated based on theWc.

      2.The procedure of transfer strategy:

      (1)Selecting transfer instancesX′and transfer instance weightsWs′:We select transfer instances and their weights based on the size of training instances and the accuracy of the base classifiers.

      (2)TransferringX′with transfer instance weightsWs′to the next layer:We Combine the transfer instancesX′of each base classifier with their weightsWs′into the input data of each base classifier to form the output data.It should be noted that the input data of the base classifier does not include its own transfer instances andtransfer instance weights.

      After describing the weighting strategy and the transfer strategy,we show the process of the DNEF framework.Each node in the input layer receives randomly sampled dataD0jas its input:

      whereX∈Rn,X0jrepresents the training instances of j-th node in the input layer,is the instance weights ofX0jandyrepresents the labels of instancesy∈Y={1,-1}.mrepresents the total number of nodes in the hidden layer.D0represents the input data of the input layer.

      In the i-th hidden layer,Dijrepresents the training data of the j-th node.The weightsof base classifiersare calculated using the weighting strategy.The training dataDi+1of the next hidden layer are generated according to the transfer strategy.

      whereXijrepresents the training instances of j-th node in the i-th layer,is the instance weights ofXij.

      The output of DNEFfis to ensemble the base classifiersHwith weightsWcthat were trained by the deep network structure.

      Algorithm 1 gives the process of DNEF,which combines the weighting strategy and the transfer strategy in hidden layers.

      3.2 The Strategies of DNEF

      In this section,we illustrate the two main strategies of the proposed DNEF framework.In Section 3.2.1,we explain the weighting strategy for the DNEF,followed by a description of the transfer strategy in Section 3.2.2.

      3.2.1 DNEF’s Weighting Strategy

      In this subsection,we describe the weighting strategy used in the DNEF framework.Firstly,given the training data for the j-th node in the i-th hidden layerDij,The base classifierhjis trained using the instancesXij,with each instance being assigned the weightsNext,the instance weightsare calculated based on the classifier errore.The computing procedure of these weights is as follows:1.Base classifier weight2.Instance weightsAlgorithm 2 illustrates the weighting strategy employed by the DNEF in the i-th hidden layer.

      We first normalize the instance weights:

      Then,the erroreof the base classifier is calculated based on the classification result of the classifier

      The base classifier weight ofhjis expressed as

      whereγis the threshold value we set.αγis a constant needed so that the equation is continuous[42].

      When considering the weight ofhj,the higher-performing base classifier should have a larger weight.Therefore,we use the square root function to scale the weights of oversized classifiers:

      After calculating the classifier weightswe update the instance weightsas follows:

      3.2.2 The Adaptive Transfer Strategy in DNEF

      We design a rule for the base classifier to adaptively choose transfer instances X′and transfer instance weights Ws′.We then add transfer instances X′with their weights Ws′to the input as output data.

      The new training dataD(i+1)jis obtained:

      We obtainDi+1by executing the transfer strategy on all nodes of the i-th hidden layer.

      The transfer strategy dynamically selects the misclassified instances with weights into the next layer of training data.Algorithm 3 describes the transfer strategy.

      4 Experiments

      The experimental environment used in this study consisted of an Intel(R)Xeon(R)CPU E5-2640 v3 and Tesla K80 GPU 11 GB memory.All the experiments were conducted in a Python environment.The required Python libraries for carrying out the experiments were Sklearn,Numpy,Pandas,and Pytorch.The selected methods for providing comparison results included DNEF,Random Forest(RF),AdaBoost,Linear Support Vector Machine(SVM),MLP,and Voting(Decision Tree).

      4.1 Datasets

      The performance of the DNEF was validated using four datasets of varying dimensions.The Adult dataset has 14 features,the REJAFADA dataset with 6824 features,the Segmentation dataset has 19 features,and the IMDB dataset has 74849 features.We downloaded the Adult dataset,REJAFADA,and Segmentation datasets from the UCI machine learning repository and the IMDB dataset from the Stanford repository.

      Table 1 gives the class,instance size,and dimensional information of these datasets.

      Table 1:Overview of the datasets

      The ADULT dataset comprises 14 features and 48842 instances.The problem is to classify whether the income exceeds$50K/year.

      The REJAFADA dataset contains 6824 features and 1996 instances.The problem it addresses is classifying files between benign and malware.

      The IMDB dataset with 74849 features and 50,000 instances.The IMDB dataset is represented by tf-idf features with positive and negative labels.

      The Segmentation dataset consists of 19 features and 2310 instances.Its objective is to classify seven distinct scenarios.

      4.2 Experimental Settings

      All results for each model were calculated using the same data set and random seeds for the metrics.The DNEF was set up as follows: (1) Each hidden layer consisted of three nodes.(2) The thresholdsγ,ε,μandρof DNEF were set at 0.15%,0.9%,150%,and 10%.Here,γrepresents the threshold for base classifier error,εrepresents the threshold for accuracy,μrepresents the threshold for instance size and μ represents the percentage of the transfer weights.The base classifier in DNEF can be any classification model with instance weights.To ensure fairness in our evaluation,we employed the same decision tree as RF,AdaBoost,and Voting for the base classifier of DNEF.The parameters for SVM and Voting remained consistent across all datasets.Specifically,for Hard Voting,we employed a decision tree with the same number and depth as DNEF,while SVM used default parameters for all datasets.

      4.2.1 The Adult Dataset

      We used 20 hidden layers DNEF whose base classifier is a 6-depth decision tree.We compared with an MLP with structure input-30-20-output and a“Sigmoid layer”is added in the last.Both AdaBoost and Random Forest adopt default parameters from the Sklearn library,employing 50 classifiers and 100 classifiers,respectively.

      4.2.2 The REJAFADA Dataset

      We used a 20 hidden layers DNEF whose base classifier is a 5-depth decision tree.In addition,we compared it with an MLP having two hidden layers,with 512 and 256 units.For both AdaBoost and Random Forest,we employed 50 and 100 classifiers,respectively.

      4.2.3 The IMDB Dataset

      We use a 50 hidden layers DNEF whose base classifier is a 10-depth decision tree.We increased the depth of the decision tree to deal with high-dimensional data while improving DNEF performance.We compared it with an MLP with structure input-1024-512-output.AdaBoost and Random Forest had 150 and 200 classifiers,respectively.

      4.2.4 The Segmentation Dataset

      We used a 20 hidden layers DNEF whose base classifier is a 5-depth decision tree.In comparison,we employed an MLP with a structure of input-256-128-output.We used 150 classifiers and 200 classifiers respectively for AdaBoost and Random Forest.

      4.3 Results and Analysis

      In this section we analyze the performance of DNEF based on the experimental results then we change the number of base-classifiers in the hidden layer to analyze the DNEF.

      4.3.1 Overall Performance

      We describe the performance of DNEF in general according to the figures.

      Fig.2 displays the performance of DNEF and the baseline model on each metric on the Adult dataset.As shown in Fig.1,DNEF performed the best in the Adult dataset,with the highest scores in Accuracy,AUC,F1,and Recall,although with slightly lower Precision metrics.

      Fig.3 demonstrates that DNEF and the baseline model performed on each metric on the REJAFADA dataset.DNEF achieved the best scores in the REJAFADA dataset,and the scores of Accuracy,AUC,and F1 indicators exceeded 0.98.

      Fig.4 shows the scores of DNEF and the baseline model on the IMDB dataset.In the IMDB highdimensional sparse dataset,all ensemble models with the decision tree as base classifier performed worse than MLP and SVM.In this case,DNEF performed better than the three ensemble models of RF,AdaBoost,and Voting.

      Fig.5 illustrates the scores of DNEF and the baseline model for the Segmentation dataset.DNEF performed well in multi-class classification.DNEF outperformed RF,AdaBoost,SVM,and MLP in terms of Accuracy,F1 score,Precision,and Recall.Next,we will analyze in detail the specific performance of DNEF on each dataset.

      Figure 2:DNEF performance in the adult dataset

      Figure 3:DNEF performance in the REJAFADA dataset

      Figure 4:DNEF performance in the IMDB dataset

      Figure 5:DNEF performance in the segment dataset

      4.3.2 The Adult Dataset

      Table 2 provides a detailed performance comparison of DNEF and other models.DNEF achieved first place with an accuracy rate of 86.18%,slightly higher than the second AdaBoost by 0.09%.Moreover,DNEF achieved the highest Recall among all models,with a 1.98% higher than RF.In terms of the F1 score,DNEF was 1.01%higher than the second-place AdaBoost.It can be seen from the AUC that DNEF was least affected by the number of positive and negative instances,which was 0.94% higher than AdaBoost.SVM and MLP were prone to overfitting on the Adult dataset with small feature dimensions;therefore,they were not as effective as DNEF.Voting (decision tree) had the worst ability to recognize negative instances,so the recall score was significantly lower than other models.

      Table 2:Classification performance in the adult dataset

      In the Adult dataset,the weighting strategy made DNEF more accurate than RF,SVM,and MLP.DNEF has achieved the highest recall score,possibly due to the transfer strategy that improves the recognition ability of positive instances and also favors the prediction of instances as positive classes,resulting in a decrease in Precision.DNEF not only surpassed AdaBoost in terms of accuracy but also led in F1 and AUC.We can conclude that the transfer strategy effectively enhanced DNEF’s ability to learn from data,making it an excellent classification framework.

      4.3.3 The REJAFADA Dataset

      Table 3 shows a detailed analysis of DNEF’s performance on the REJAFADA data.Compared to the other baseline models,DNEF achieved first place(98.09%)in experimental results in Accuracy,F1,and AUC.When all baselines achieved an accuracy of 98%or less,DNEF broke through to rank first with 98%,improving by 0.35%over the second-place RF and 0.65%over the third-place SVM.Although SVM has the highest Recall,its Precision was 2.55%lower than DNEF.

      Table 3:Classification performance in the REJAFADA dataset

      Since the baseline models had high accuracy rates of over 97%,the challenge of classification was concentrated in a small number of indistinguishable instances.DNEF effectively handled these tricky instances,benefiting from its transfer strategy.DNEF used the square root to scale down the weights of base classifiers with high accuracy,effectively solving the frequent overfitting problem.Among the three ensemble learning models,AdaBoost and Voting,DNEF outperformed them in terms of Accuracy,AUC,and F1 metrics.This superior performance was attributed to DNEF’s effective weighting strategy,which ensures accurate classification,and its transfer strategy,which enhances the recognition ability for samples from various classes.Consequently,DNEF exhibited exceptional performance in F1 and AUC.SVM and MLP had a significant bias in their ability to recognize positive and negative instances,leading to lower F1 scores.

      4.3.4 The IMDB Dataset

      In this subsection,we evaluated the performance of DNEF on the IMDB dataset based on Table 4 Regarding Accuracy,DNEF improved by 0.21% and 0.98% compared to RF and AdaBoost.The Precision of DNEF was slightly better than RF by 0.17%.Although the recall score of DNEF was 1.12% lower than that of AdaBoost,DENF scored higher than AdaBoost on all other metrics.In terms of F1 metrics,DNEF’s Precision and Recall combined exceed all ensemble models.Additionally,DNEF achieves the highest AUC score among all ensemble models,surpassing RF by 0.21%.

      Table 4:Classification performance in the IMDB dataset

      IMDB datasets were represented by tf-idf features with high-dimensional sparsity.According to Table 4,DNEF was still better than the ensemble learning models with decision trees such as RF,AdaBoost,and Voting in processing tf-idf feature data.The inferior performance of Voting suggested that the decision tree was inadequate for handling the IMDB dataset.Comparing the boosting ensemble strategy of AdaBoost with the bagging ensemble strategy of RF,we can obtain that DNEF surpasses these two representative ensemble models in terms of Accuracy,AUC,F1,and so on.SVM and MLP performed well on the IMDB dataset,which can be attributed to their ability to handle high-dimensional sparse data.

      4.3.5 The Segmentation Dataset

      In this subsection,we present an analysis of DNEF’s performance based on the results summarized in Table 5.Notably,DNEF demonstrated a significant advantage in multi-class classification.It achieved the highest Accuracy,F1,Precision,and Recall among the evaluated models.Specifically,in terms of Accuracy,DNEF surpassed the second-place RF by a margin of 0.97%,reaching 97.32%.Moreover,DNEF exhibited exceptional performance in F1,outperforming RF by 0.97%.Regarding Precision and Recall,DNEF consistently delivered the best results.The AUC of DNEF was only 0.04%lower than that of RF,underscoring its competitive performance across all measured metrics.

      Table 5:Classification performance in the segmentation dataset

      DNEF demonstrated strong performance in multi-class classification tasks,likely due to the effective utilization of its weighting strategy in conjunction with its transfer strategy.DNEF effectively addressed the complexity of multi-class classification through its weighting strategy.By adapting the classifier and instance weights for each class,DNEF further enhanced its accuracy.This tailored adjustment ensured that DNEF could effectively handle the intricacies associated with multi-class classification.Notably,DNEF maintained high classification accuracy while excelling in the F1 metric.This achievement was attributed to its transfer strategy,which intelligently guided indistinguishable instances from each class to the next hidden layer during operation.Consequently,DNEF maintained excellent Precision and Recall,further solidifying its suitability for multi-class classification tasks.

      4.3.6 Parameter Analysis

      The number of nodes in the hidden layer is an important parameter of DNEF.This subsection analyzes the impact of the number of nodes in the hidden layer on the performance of DNEF.The number of nodes in each hidden layer of DNEF was changed to observe the different effects during training.Fig.6 displays the training loss of DNEF with three different numbers(1,2,3)of nodes per hidden layer on the Adult dataset.As the number of nodes in the hidden layer increased,the loss line became smoother,and the training loss decreased in Fig.6.We analyzed the details of the three lines in Fig.6 as follows:

      Figure 6:The training loss of DNEF with three different numbers of nodes

      The gray line represents only one classifier in each layer of DNEF.In this case,the transfer strategy did not work.The training loss of the gray line decreased from 4.92 to 4.71.The loss decreased in an oscillating manner as the number of layers increased.Compared to the other lines,the gray line was the most oscillating and had a higher loss in the same number of hidden layer cases.The orange lines represented two nodes in each hidden layer in DNEF,in which case the transfer strategy worked.The training loss dropped from 4.81 to 4.66.Compared to the gray line,we could see that the orange line dropped more smoothly,and the loss was smaller with the same number of hidden layers.The blue line represents three nodes in each layer in DNEF.The blue line dropped from 4.77 to 4.51,which was the smallest of all the lines.From Fig.6,the loss drop line was smoother than the other lines,and the training loss was consistently smaller.It showed that the weighting strategy worked better with the transfer strategy as the number of nodes in the hidden layer increased.This effectively reduced the training loss and made the training smoother.

      4.4 Experiment Conclusion

      We compare DNEF with the baseline algorithm on four datasets.Our experimental results demonstrate that DNEF outperforms the traditional ensemble model.DNEF significantly improved accuracy on the Adult,REJAFADA,and Segmentation datasets with F1 metrics compared to the baseline model.In the case of the high-dimensional dataset IMDB,DNEF outperforms traditional ensemble models in Accuracy,F1,AUC,and Precision.DNEF shows excellent performance,again proving its strength in multi-class classification.At the same time,we can infer the limitations of DNEF,as it is not directly comparable to MLP and SVM in high-dimensional data when using the decision tree as the base classifier.This is also the challenge of DNEF at present.It can be concluded from the parameter analysis that increasing the number of nodes benefits the effective implementation of DNEF’s transfer and weighting strategy,ultimately enhancing training effectiveness and stability.

      DENF framework uses the weighting strategy between hidden layers and the transfer strategy in hidden layers,which is our main difference from other ensemble models.From the experimental results and parameter analysis,we conclude that it is the network structure of DNEF that contributes to its superior performance compared to other traditional ensemble models.

      5 Conclusion

      In this paper,we propose DNEF,a new ensemble learning architecture,DNEF incorporates a deep network structure that iterates between hidden layers and trains classifiers in parallel within the hidden layer.The weighting strategy trains the classifier based on the instance weights generated in the previous layer and further adjusts the instance weights;the transfer strategy operates within the hidden layer,selecting instances with weights for each node in the layer and combining them with the training data for the next layer.Compared to popular ensemble learning approaches,DNEF accounts for the relationship between base classifiers using the weighting strategy and enhances model complexity with the transfer strategy.The DNEF demonstrated promising results on four real datasets.Specifically,on the multi-class dataset Segmentation,DNEF achieved 0.9732 in Accuracy and F1,respectively.These experimental findings validate that DNEF,as a novel and exceptional deep ensemble architecture,outperforms traditional ensemble models.The primary contribution of this paper lies in exploring the ensemble structure under the network structure and expanding the ideas for ensemble learning.In our future work,we will extend DNEF in two ways:(1)We will explore various base classifiers for highdimensional datasets.(2)We will also consider applying DNEF to different data such as images,sound signals,etc.,and other learning tasks such as semi-supervised learning and incremental learning.

      Acknowledgement:We thank the School of Mathematics and Information of South China Agricultural University for supporting this study.

      Funding Statement:This work is supported by the National Natural Science Foundation of China under Grant 62002122,Guangzhou Municipal Science and Technology Bureau under Grant 202102080492,and Key Scientific and Technological Research and Department of Education of Guangdong Province under Grant 2019KTSCX014.

      Author Contributions:Study conception and design:Ge Song;data processing:Changyu Liu,Zhuoyu Ou;analysis and interpretation of results: Ge Song,Yuqiao Deng,Siyu Yang;draft manuscript preparation: Siyu Yang,Ge Song.All authors reviewed the results and approved the final version of the manuscript.

      Availability of Data and Materials:Four publicly datasets were used for analyzing our model.They can be found at https://archive.ics.uci.edu and https://ai.stanford.edu/~amaas/data/sentiment.

      Conflicts of Interest:The authors declare that they have no conflicts of interest to report regarding the present study.

      筠连县| 湖口县| 绵竹市| 嵊州市| 永仁县| 湘乡市| 成武县| 梁平县| 湘潭县| 北安市| 闸北区| 凤阳县| 马鞍山市| 丽江市| 广宁县| 石首市| 五原县| 大渡口区| 通海县| 泸定县| 邛崃市| 新郑市| 石泉县| 马关县| 湖南省| 铁力市| 乌鲁木齐县| 云安县| 明水县| 基隆市| 綦江县| 武山县| 乳山市| 酉阳| 宣恩县| 泽州县| 河源市| 锡林郭勒盟| 宣威市| 平湖市| 阳春市|