• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Relative attribute based incremental learning for image recognition

    2017-05-16 10:26:43EmrahErgul

    Emrah Ergul

    Electronics&Communication Eng.Dept.of Kocaeli University,Umuttepe Campus,41380 Kocaeli,Turkey

    Original article

    Relative attribute based incremental learning for image recognition

    Emrah Ergul

    Electronics&Communication Eng.Dept.of Kocaeli University,Umuttepe Campus,41380 Kocaeli,Turkey

    A R T I C L E I N F O

    Article history:

    Received 6 November 2016

    Received in revised form

    15 January 2017

    Accepted 19 January 2017

    Available online 30 January 2017

    Image classi fication

    In this study,we propose an incremental learning approach based on a machine-machine interaction via relative attribute feedbacks that exploit comparative relationships among top level image categories.One machine acts as‘Student(S)’with initially limited information and it endeavors to capture the task domain gradually by questioning its mentor on a pool of unlabeled data.The other machine is‘Teacher (T)’with the implicit knowledge for helping S on learning the class models.T initiates relative attributes as a communication channel by randomly sorting the classes on attribute space in an unsupervised manner.S starts modeling the categories in this intermediate level by using only a limited number of labeled data.Thereafter,it first selects an entropy-based sample from the pool of unlabeled data and triggers the conversation by propagating the selected image with its belief class in a query.Since T already knows the ground truth labels,it not only decides whether the belief is true or false,but it also provides an attribute-based feedback to S in each case without revealing the true label of the query sample if the belief is false.So the number of training data is increased virtually by dropping the falsely predicted sample back into the unlabeled pool.Next,S updates the attribute space which,in fact,has an impact on T's future responses,and then the category models are updated concurrently for the next run. We experience the weakly supervised algorithm on the real world datasets of faces and natural scenes in comparison with direct attribute prediction and semi-supervised learning approaches,and a noteworthy performance increase is achieved.

    ?2017 Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http://creativecommons.org/licenses/by-nc-nd/4.0/).

    1.Introduction

    Instead of mapping low level features(color/edge histograms, bag of words,Fourier,Wavelet transform features etc.)directly to the class labels,attribute-based approaches are nowadays popular instruments in the computer vision applications[1-8].Attributes are mainly human interpretable and visually detectable entities on the visual data.They are used for intermediate level data representations between machine-alone understandable low level features and the human speci fic top level class labels.Attributes are introduced mostly because of sparse category labeled training data and large variations among low level features caused by both viewing conditions and high intra-class variance[1,9].

    Since many top level categories share attributes as the common platforms,they can be learned with much more data samples than class labels.In addition to traditional problems like classi fication and retrieval[2,6],this leads to transferring the previously learned visual attributes to learn/detect a novel class which has not been seen during the training stage but we still know its attribute values, called zero-shot learning[10,11].Besides,we can even infer an unfamiliar class which we have not known its attribute groundtruths[12,13]at all.We may also produce more semantic queries by using the visual attributes to perceive scenes/objects in the image search engines[14].

    In this work,we propose a new approach which tries to recover from aforementioned insuf ficiencies of the supervision caused by the human factor,and the lack of connections in the attribute-based class modeling.It is based on the relative attributes for incremental visual recognition.We generate the relative attributes in an unsupervised manner by randomly ordering the target classes.On the other hand,we model the visual classes in a weakly supervised environment where a very limited number of training data is used. Meanwhile,we update both the attribute space and class models concurrently in an iterative algorithm.To do so,we establish a machine-machine interaction protocol that projects a Student(S)-Teacher(T)relationship that is available in the real world.

    1.1.Contributions

    We present three contributions in this paper.First,the attribute space and the class models are learned simultaneously in an iterative manner.This is very intuitive since the class models are constructed on the attribute space,and there must be a connection in a combined learning procedure.The second contribution is the bilateral learning paradigm where S updates the attribute space which,in fact,in fluences T's future attribute-based explanations. We can say that T is also learning or it should generate its responses with respect to S'inferences on the attributes.Also note that attributes are initialized randomly and T's attribute-based explanations are prepared online in an unsupervised manner.The last contribution is about the virtual growth of the training set.Unlike semi-supervised methods where each sample is used only once for the learning procedure,T does not reveal the true label of the selected image if the belief is false.So the number of training data is increased by dropping the falsely predicted sample back into the unlabeled pool.In the experiments,we see that the number of iterations is almost doubled,and this leads to a better performance in visual recognition.

    The general structure of the rest of the paper is as follows: Section 2 summarizes the related work about attribute-based applications with their important aspects.Then,we explain the proposed work in details in section 3,and the experimental evaluations are presented with comparative results on two popular datasets in section 4.We conclude the paper with our clear inferences in section 5.

    2.Related work

    The literature of attribute-based computer vision problems can be generally summarized in the types and extraction methods of attributes,the applications and datasets on which they are implemented,and evaluation criteria in the experiments,as shown in Fig.1.

    2.1.Attribute types

    The importance of an attribute for each class varies for recognition purposes.Thus,we may produce an image/categoryattribute co-occurrence matrix such that attribute membership for each class/image and the corresponding correlations with respect to the attributes may be captured.These relations can be produced in a supervised manner beforehand;or in a semisupervised/unsupervised fashion.Some researchers suggest that binary correlations would be suf ficient while others claim real valued ranking scores are essential to measure the attribute strength among categories/samples.

    Relative attributes are first introduced by Parikh and Grauman in Ref.[6]with the assumption that semantically more enriched data descriptions and discrimination will be achieved if we use relative class memberships on attributes,instead of binary relations.Biswas and Parikh propose an interactive learning scheme for both object and face recognition with relative attributes in Ref.[15].They improve this semi-supervised work in Ref.[10]by weighting negative candidate samples,selecting the query images and updating the attributes.

    Kovashka and Grauman[16]open a new perspective to learn binary attributes by modeling them with many visual de finitions instead of training only one discriminative model for each which is called shades of an attribute.Jayaraman and Grauman[17]study zero-shot classi fication paradigm with the assumption that traditional attribute scores are not suitable enough to predict them perfectly in unseen images.Chao-Yeh and Grauman[18]convey the zero-shot discussion on a new idea that class-sensitive attribute classi fiers can transfer the learning to infer the new ones unobserved during training.Liang and Grauman[19]explore active learning strategies to learn relative attributes better without exploiting all pairwise comparisons.

    Fig.1.Summary of the literature work on attribute-based computer vision problems.

    2.2.Attribute extraction

    In supervised settings,the attribute labeling can be constructed on whether the instances individually or the visual categories generally.Akata et al.[8]try to solve the attribute-based classi f ication problem by constructing a composite framework which embeds mapping images from both sides of low-level features and class labels into a common attribute space at once.Jing et al.[20] generate a deep Convolutional Neural Network(CNN)architecture to learn appearance and motion based binary attributes.Kiapour et al.[21]experiment a challenging task to extract deep learning features that are based on retrieval and similarity learning methods for this task.Xinlei and Abhinav[22]propose an approach to utilize web data for learning visual attributes through CNN.

    Unlike supervised learning where the attributes are named speci fically,one may not name the attributes but only use them to discriminate the data classes.Farhadi et al.[5]embed the object recognition problem into describing objects and mainly focus on attribute learning.Ma et al.[11]implement an algorithm to learn class-level relative attributes in an unsupervised manner where they produce relative attributes as many as the number of pairwise class combinations in a greedy fashion.Karayel and Arica[23] follow the similar way of[11,24]but they produce the attributes completely randomly and attribute selection is implemented afterwards.Yu et al.[25]formalize a category-attribute co-occurrence matrix for cross-category generalization.

    One may relate the attributes with each other such that hierarchical structure is established based on spatial layout,heritage and encapsulation aspects.Ran et al.[26]combine categoryspeci fic attributes and category-level information for generic instance search in the web.Victor et al.[27]extract relative attributes in different layers of the CNNs architecture by encoding the location,sparsity and the relationship between visual attributes. Local attributes,besides global ones,are also considered when the spatial ordering is important for recognition.But in this case,extra segmentation methods are required such as bounding boxes[4,7] or intensive labeling approaches[28]lead to additional cost to learn local attributes.

    3.Materials and methods

    3.1.Overview of the proposed method

    In this work,we introduce a new approach for incremental visual recognition to modelvisual categories simultaneously with the relative attributes in an incremental learning procedure.To do so, we establish a machine-machine interaction protocol that projects a Student(S)-Teacher(T)relationship available in the real world. Relative attributes are first initialized by T in a way that sorts the image classes randomly in an unsupervised manner.Because we do not aim to locate the attributes speci fically on spatial regions of the visual data,they are learned from global low-level features(i.e. GIST).The classes are then modeled by S with a limited number of category labeled data in the mid-level attribute space.Although it is T's initiative to generate random class orderings for the attribute learning,S then takes the responsibility to update both the attribute space and class models concurrently in an iterative algorithm. While S selects a sample,x(q),from a pool of unlabeled data each time and propagates it with the belief class,T paves the way for S' corrective actions on class modeling with its attribute-based, comparative and af firmative/rejective responses.

    Unlike[10,15]where the attribute-based explanations are prepared by the Amazon Mechanical Turk(AMT)workers before the learning procedure,the T machine decides online what explanation should be given to S for mentoring.While doing updates on class models,S also appends some candidate samples to its labeled dataset.The candidates are selected out of the unlabeled data,Uset, temporarily with respect to the conditions on the current attributes by constructing a weighting cube.This alternate routine which describes the attribute-class model updates continues until S runs out of the unlabeled data that leads to incremental learning for recognition.The flow chart of the algorithm is given in Fig.2.

    3.2.Preliminaries of initiative setup

    Relative attributes are used as a communication channel in bidirectional interactions between Student(S)and Teacher(T) machines.To do so,we first divide the training set randomly, apart from test set,randomly into three discrete subsets:Student(Sset),Teacher(Uset)and Validation (Vset)sets.Sset consists of a very limited number of class-labeled images per category,and S uses this subset to achieve initial class models.On the other hand,S does not know the class labels of Uset and it basically represents T's teaching materials about the task domain.S utilizes Uset for both picking a sample in its belief propagation through questioning and for selecting some candidates in a weighting scheme to update its current category models. On the contrary,T handles this subset for preparing the best attribute-based response to indicate why the belief is correct or not. Vset is used for optimizing the free parameters of the algorithm and the initial setup is summarized in Fig.3.

    Another important detail of the proposed work is to construct an attribute space which constitutes the communication channel between S and T machines.Given a class based ordering,which relates every category to each other with a less/more and similarity conditions,we use the Newton's method in Ref.[6]for a relative attribute as:

    where rmis the real valued ranking score for the training instance,xi,on the attribute basis,is the parameter vector of the relative attribute model.Omis the set which includes pairs of data instances holding for the more/less conditions whileSmrefers to the set of similar samples.So the optimum solution would then order the classes on the weight vector,wm,by minimizing the cost function in(4);where Z is the constant that regulates the balance between weight decreasing and the non-negative slack variables,εijandγij.This results in maximizing the margin between classes in the order de finition ofam.

    T initializes the relative attribute space on Sset by randomly generated class ordering[23]for each attribute in an unsupervised manner,and shares this information with S.After preparation of the subsets for training and random initialization of the attribute space by T,it is now time for S to initialize category models by using Sset.S uses RBF kernel-based one-vs-all SVM[29]for class discriminants and the free parameters are optimized in a grid search method with 10-fold cross validation on Vset.To summarize,low level features, x(i),are first projected into attribute space,A={am}∈?M,and classi fication is implemented in this mid-level with Sset at the beginning.

    Fig.2.Flow chart of the proposed algorithm.

    3.3.Student-teacher interactions for visual recognition

    Incremental learning corresponds to gradually updating the attribute space and category models alternately in S-T interactions until S labels all of the supplementary data set,Uset,correctly with the help of T's attribute-based explanations.In detail,S picks a sample,x(q),out of Uset which will be as much informative as possible for its class learning and this is achieved by maximum entropy,H,[10]:

    wherep(x(i))yis the probability output of SVM for the sample,x(i), for each class,y.Hence,the class label c(j)with the highest probability,y*,is attached to the query sample,x(q),and it is questioned for an explanation like‘Does x(q)belong to class c(j)?’As aforementioned,the falsely predicted sample is dropped back into Uset which helps increasing the training set virtually.

    When S machine often picks the same set of query samples,x(q), out of Uset,T machine would generate similar attribute-based explanations for the S'beliefs.Because S uses these explanations for the updates of attribute space,over fitting may occur respectively. To avoid over fitting at this point,we propose three enhancements for S.First,S machine also propagates its best prediction once in a while(i.e.one in ten iterations)on the contrary to the selection method which is based on the maximum entropy.Secondly,S restricts one sample image to be selected often by deploying a counter for each.For instance,if S sends the x(q)back to the Uset its counter will be initialized(for example set to three).The counters are then decreased by one through iterations,and the previously selected images are not allowed to be chosen again until their counters reach zero,i.e.not for the three next runs in this case. Finally,S keeps track of the selection history of samples in the current Uset.Hence,the image is not propagated again with a class label which has been predicted falsely in the previous iterations.

    After S propagates its belief in a query it is now T's turn to reply, and this process is displayed in Fig.4.The attributes are computed relatively with respect to the class labels and T already knows the class labels of both Sset and Uset.So one may append an attributebased explanation to the decision of’accepted’or’rejected’by selecting an attribute and its comparative term.Unlike semisupervised learning[30,31],S is in fluenced in the complementary attribute space for category modeling in a way that the query sample,x(q),is sent back to Uset if the prediction is false,or it is added to Sset,otherwise.T is then responsible to select the best attribute-condition pair for the explanation and this is achieved by first projecting x(q)and the datasets,Sset and Uset,on each attribute axes withra(x(i))=(x(i))Twa,wherera(x)is the ordering function computed as in Ref.[6]and wais the weight vector of attribute am.Thereafter,the samples are sorted increasingly withtheir ranks,ra,as in Ref.[15]and the voting method is implemented as:

    Fig.3.Preliminaries of the proposed incremental learning algorithm.

    where indnum(·)refers to the index number of samples on the attribute axes in the ordered list.The intuition is that if x(q)does not belong to the falsely predicted class,Cpredicted,but it belongs to another class,Ctrue,where Ctrue>Cpredictedon am,thenwe expect its vote to be positively high for that attribute.On the contrary,the vote should be as low as possible in its absolute form when the prediction is true.So T has 2 M options(i.e.more or less conditions for M attributes)to choose for false and M options for true beliefs since the condition is already selected as’similar’.

    Finally,T may mislead S by just using the probabilities computed in(8).For example,giving an explanation like‘No,x(q)is not as natural(am)as c(1)’contradicts the reality when it is the case‘Ctrue>c(1)on naturalness'.Because T knows the category orderings on attributes in addition to Ctrueof x(q),the response may be reestablished with the next highest probability.Also note that there is no need of such considerations when the belief is true,and the attribute with the lowest vote(i.e.unsigned)is selected with the condition of similarity.

    3.4.Updating parameters of the learning algorithm

    3.4.1.Attribute spacef ine-tuning

    Basically,the student-teacher interaction produces a pair of query-response through the iterations which triggers updating the parameters of attribute space and category models alternately. Assuming that T accepts S belief,x(q)is appended to Sset and subtracted from Uset since the answer reveals the ground truth category label of x(q).Thus,all attributes can be updated by creating new pairs with x(q)in its class orderings.The situation becomes interesting when T rejects the prediction because the true class label of x(q)is not given in the answer and it is not added to Sset.But S still finds additional pairs to update only the selected attribute by perceptual induction.For instance in‘No,x(q)is more natural than class c(1)’,it is known that x(q)is not in the predicted class with’more’condition on attribute’naturalness'.So one may simply produce new ordering pairssuch that query image visualizes attribute amstronger thanon Sset.

    Additionally,S may infer even more pairs by using its prior knowledge.When the class-based ordering on amisand the previous answer is given for false prediction,then x(q)is actually more natural thanThe same idea holds for the‘less’condition except reverse ordering rule.Hence,this would increase the number of pairing samples for attribute learning which leads to better discrimination.Also note that newly added pairs are deleted for the next iteration since S has not discovered the class label of x(q)yet,unlike in case of true prediction.Besides,previously updated attribute weight vectors, wa,are carried through iterations for incremental learning insteadof random initialization,and only the paring samples which violate the constraints of ranking margin are used in the attribute update procedure.It accounts for avoidance of the over fitting when using the same ordering pairs through iterations.

    Fig.4.Student-Teacher interaction scenario.

    3.4.2.Selecting candidates for updating the class models

    Like in attribute updating,S makes inferences out of T's answers but now it is for picking the candidate samples from Uset temporarily to update class models.To do so,S produces a Uset’Weighting Cube’on the Uset in which the weights are summed cumulatively in iterations for each candidate with respect to positive and negative sides,in addition to target categories.The weighting scheme is visualized in Fig.5.In detail,the candidate samples,Xcandidates={x(p)}p=1,2,…,U;x(p)∈Uset,are chosen on the attribute base,am,which has been given in T's explanation by first computing the ranks of all samples,ra(x(i))=(x(i))Twa; x(i)∈(Uset+Sset).Afterwards,the samples are sorted and positions of the Uset samples in the sorted list are checked with x(q)as it is the benchmark on the condition embedded in T's response.If S predicts x(q)correctly to be c(1),for instance,the explanatory condition will be’similarity’,and the candidates should be even more similar to c(1)than x(q)on the am.This is achieved by first finding the median image of c(1)(i.e.μ(x(j)),see Fig.5a.)and then marking the symmetric area which is centered on the median point and bounded by x(q).The Uset samples projected on this region are selected as the’positive’candidates for c(1)in this iteration only if they are not pinned as’negative’for c(1)previously.

    On the other hand,the‘negative’candidates for the predicted class can be achieved in false predictions easily in a similar manner except relative conditioning which determines the side moving away from x(q)(i.e.to the left for‘less’and to the right for’more’). The intuition for‘less’condition is that if x(q)is not as amas c(3),the samples which are even less amthan x(q)would not belong to c(3), either.The same idea holds for the’more’condition in the opposite direction(see Fig.5b and c).After selecting the candidate positive/ negative samples,x(p),for the predicted class,Cpredicted,S computes their cumulative weights as:

    where UB is the upper bound of the iterations andσis the scale factor.Summation yields to cumulative votes over iterations which is independent on the selected attribute,am.So the weights happen to be the distance that refers to the number of images between the candidate and x(q).Also note that the exponential factor in(9) ensures that weights are added with an increasing ratio.Eventually, S selects some candidate samples out of Uset with their weights in each iteration,and these candidates will be appended to Sset for updating the category models.

    3.4.3.Updating the class models

    So far,S fine-tunes the attribute space and picks some candidate samples with their weights from Uset which will be appended to Sset as a temporary extension.Now it is time to update the category models,and this is achieved by one-vs-allL2RBF kernel based SVM[29].In detail,the samples in the Sset are known with their labels by S,so their weights are naturally 1 like in Ref.[15].Hence,the candidate weights are first normalized to[0,1]such that they are summed to 1 for true predictions while the maximumvalue is set to 1 for false predictions because absolutely negative samples exist in the candidate list.

    Fig.5.Candidate selection with a weighting scheme.{x(1),x(2),x(3),x(4)}∈Uset are the selected candidates for an explanation of:a)x(q)is quite as amas c(1),b)x(q)is not as amas c(3),c)x(q)is more amthan c(4).

    Once S predicts x(q)truly the candidates are then labeled as Cpredictedwith their normalized weights.The situation becomes confusing when T rejects the belief because S simply assumes that the candidates are not be in Cpredicted,either.For example,if T answers‘x(q)is not as amas c(3)’and the class-ordering is c(2)>c(3)>c(4)>c(1)on am,S may expect that the candidates x(p),in addition to x(q),may come from c(4)or c(1).The decision is trivial when there is only one candidate category(i.e.c(1)if c(3)is interchanged with c(4))and the negative side of the Cpredictedvote is taken directly as the candidate weight.But S should make its best guess for x(p)if there are more possible classes.In detail,S picks the best one among candidate classes by using its current category models if there is no weighting for these classes yet.Otherwise,the one with the highest vote,computed by the difference between positive and negative weights of x(p)on the candidate class,is chosen directly.

    So long as S determines the weights and class labels for the candidate samples,x(p),which are appended to Sset temporarily, SVM algorithm is implemented on the new training set by initializing the model parameters with the previous ones in each iteration.The free parameters are optimized in a grid search method with 10-fold cross validation on Vset as mentioned in the’preliminaries(see 3.2)’subsection.Also note that class models and attribute space are not updated each time when S achieves worse accuracy performance than the previous setups on the Vset.The intuition is that T's explanations sometimes may not help S improve its visual recognition because of over fitting,and it can downgrade the performance in iterations.The alternate update procedure continues iteratively until S empties Uset with its correct predictions.

    4.Performance evaluation

    4.1.Experimental setup

    We use two real world datasets to evaluate the proposed algorithm for comparative experiments with the other attribute based methods:Outdoor Scene Recognition(OSR)[6,10,11]that contains 2688 images in 8 scene categories,and Public Figure Face(PUBFIG) [6,15,23]which has 800 images from 8 face classes.The global low level features(i.e.512D GIST and additional 30D LAB color histograms for PUBFIG)are also provided with the datasets.We repeat the experiments 30 times with different training/test splits,and the average results are noted for generalization.In each run,there exist 15 samples in the Sset,25 images in the Uset through random selection,and the rest is reserved for the test.Vset is generated from Sset+Uset by injecting Gaussian noise.Furthermore,the supervised attributes which are given with the datasets are also used(6 for OSR and 11 for PUBFIG[6])for the performance evaluation.In the implementation,LibSVM[29]algorithm is utilized for category modeling while Newton's method in Ref.[6]is used for the extraction of relative attributes.Also note that RBF kernel outputs are whitened as in Ref.[32]for normalization before SVM.

    Basically,the proposed work is compared with two framework methods of attribute-based recognition:Direct Attribute Prediction (DAP)and Semi-supervised Attribute Classi fication(SSAC).DAP offers a two-phase approach where the attribute space and class models are learned independently.Hence,Sset and Uset are now combined in a joined set.We first learn the relative attributes; thereafter categories are modeled on this mid-level via the joined set.SSAC,on the other side,represents’in-between’methodology between DAP and the proposed work.The attribute space is generated with the joined set,Sset+Uset,and the attribute weight vectors,w,are fixed during the class modeling like in DAP.S initializes the classi fiers on the attribute space by using Sset at the beginning,and it picks a sample out of Uset with its current best guess in an iterative manner.So,one sample is transferred fromUset to Sset with the belief class label and S updates the class models with the new Sset in each run.

    4.2.Multi-class classification

    Our main concerns in multi-class classi fication are whether:a. Attributes constitute a fruitful intermediate space for data representation,b.Unsupervised attributes are more discriminative than the crowd sourced attribute labeling in supervision,c.Learning attributes and category models incrementally in a machinemachine interaction increases the recognition accuracy when compared with DAP and SSAC frameworks.

    The plot of mean accuracy vs.varying number of attributes for OSR and PUBFIG datasets are displayed in Fig.6.With regards to the supervised attributes alone,all three algorithms seem to achieve the similar performance despite the fact that the proposed and SSAC algorithms maintain a better start.We believe that it is due to the limited number of attributes and the class orderings on attributes are akin to each other.We learn only 28 unsupervised attributes in order to compare the proposed work with the other studies in literature.For these,all methods achieve much lover accuracy at the beginning but they make up the difference quickly when compared with the supervised attributes.As expected,the performance increases little about 1-2%for DAP and 3-4%for SSAC overall,whereas an obvious rise is experienced with the proposed algorithm,that is up to 6-8%for the 28 unsupervised attributes. The same evaluation is still true if we combine both.When the unsupervised attributes are appended,the accuracy rises up again about 3%with respect to the single usage of unsupervised attributes.

    As SSAC and the proposed method are common in learning the class models iteratively,we display their plots of accuracy vs.iteration step in Fig.7.Remembering that the relative attributes are learned with Sset+Uset initially,SSAC keeps them fixed during the class model updates whereas S machine updates the attribute space and classi fiers concurrently with respect to T's attribute-based explanations in the proposed work.Also note that the query image, x(q),is transferred with its belief class label directly into the Sset each time in the SSAC algorithm.So the iteration bound is restricted simply to the number of samples in the Uset,i.e.200 for 25 images per category.On the contrary,the bound varies in the proposed method since the query image is dropped back into the Uset when it is predicted falsely.

    In Fig.7,it is observed that the SSAC makes a better startof about 2-4%than the proposed method.Because SSAC uses the joined set, Sset+Uset,for the attribute space initialization,we assume that SSAC has started to learn the attributes more consistently with a larger dataset,as expected.But the accuracy trend of SSAC is not as steep as the one of the proposed work during further iterations.For example in the PUBFIG experiments,the best accuracy level of the SSAC is already captured at about iteration#230 and 4%final increase is achieved in the proposed algorithm.We conclude that theperformance is increased by expanding the training set virtually while some Uset samples are used multiple times in iterations.

    Fig.6.Accuracy performance analysis with varying number of attributes on:a)OSR,b)PUBFIG.

    Fig.7.Iterative performance analysis of SSAC and the proposed work on OSR and PUBFIG datasets.

    Receiver Operating Characteristic(ROC)curve is frequently used in literature to evaluate the performance of classi fiers.Basically,the ratio of false and true positive samples is plotted by changing thresholds in a step-wise manner.The classi fier is regarded as more successful when its plot rises up earlier and sharper than the others.We compare the performances of the DAP,SSAC and proposed work with supervised and combined(i.e. supervised+unsupervised)attributes on ROC curves for both datasets in Fig.8.As seen,the accuracy is increased obviously when all attributes are used together,and this con firms that theunsupervised attributes add discriminative power in dimensionality.Additionally,the proposed algorithm outperforms the others clearly while SSAC is slightly better than the DAP method.

    Fig.8.ROC analysis of the DAP,SSAC and the proposed algorithms on:a)OSR,b)PUBFIG.

    Table 1 Performance(%)comparison of the algorithms.

    Finally,the proposed method is compared with the similar approaches in literature on the same experimental setup,and the mean±std accuracy results of 30 repeats are listed in Table 1.BINs, PCA and FLD algorithms are actually used for dimension reduction. Nevertheless,the basis vectors which are extracted during their implementations help representing the data in a new feature space, so they are included as baselines for this reason.The other methods generate supervised/unsupervised attributes in the mid-level for visual recognition,like the proposed work.The results in Table 1 show that the proposed method outperforms the other approaches for about minimum of 2%although the attributes are produced by random class orderings unlike in Ref.[11],and there is no speci fic attribute selection,differently from Ref.[23].The intuition behind such a notable result is that we shape the attribute space simultaneously with the class models and the number of training samples is increased during attribute-based feedbacks.We also assume that the expanded number of unsupervised attributes with distinct class orderings establish a better representation without human laboring,leading to more effective classi fiers.

    5.Conclusion

    In this study,we use attributes as the mid-level representation, instead of low-level features,in an incremental learning procedure for visual recognition.Attributes constitute a communication channelthrough S-T machine interactions.The conversation mainly indicates a question-answer protocol where S selects an informative sample from the unlabeled pool of data,Uset,each time and T gives attribute-based explanations for mentoring.Also note that the training set is expanded virtually since the query image is dropped back into the Uset without exposing its ground-truth category label when S predicts it falsely.So we introduce a student centered learning method where S is actually responsible for updating the attribute space and class models simultaneously in an iterative manner.Although T initiates the relative attributes by sorting the image categories randomly in an unsupervised fashion, S updates the attributes with the selected query images from then on.This means that T also learns in time how the student interprets the representative space on which the class discriminants are built. Additionally,S does not use only its labeled dataset,Sset,for updating the class models,but it also implements a candidate selection method to append some samples from Uset to Sset in a weighting scheme.

    We experiment the proposed work comparatively with the frameworks DAP and SSAC,and other similar approaches in literature on two popular mid-scale datasets,OSR and PUBFIG.Because classi fication is the focal point,the mean accuracy vs.iteration step/ number of attributes and the ROC curves are the main themes for analysis during multiple experiments.To summarize,the proposed method outperforms other works signi ficantly,and it is concluded that the unsupervised attributes exploit the recognition performance when learned incrementally and simultaneously with the class models even though they are achieved by random class orderings without human laboring.

    Acknowledgements

    This work was supported in part by the Scienti fic and Technological Research Council of Turkey(TUBITAK)2214B.14.2 TBT.0.06.01.214.83.

    Appendix A.Supplementary data

    Supplementary data related to this article can be found at http:// dx.doi.org/10.1016/j.trit.2017.01.001.

    [1]V.Ferrari and A.Zisserman,Learning visual attributes,Neural Information Processing Systems Conference,3-8 December 2007,Vancouver,Canada,pp. 433-440.

    [2]C.H.Lampert,H.Nickisch,S.Harmeling,Attribute-based classi fication for zeroshot visual object categorization,J.IEEE Trans.Pat.Anal.Mach.Intel.36(2014) 453-465.

    [3]C.H.Lampert,H.Nickisch and Harmeling S,Learning to detect unseen object classes by between-class attribute transfer,IEEE CVPR Conference,20-25 June 2009,Miami,FL,USA,pp.951-958.

    [4]A.Farhadi,I.Endres and D.Hoiem,Attribute-centric recognition for crosscategory generalization,IEEE CVPR Conference,13-18 June 2010,San Francisco,CA,USA,pp.2352-2359.

    [5]A.Farhadi,I.Endres,D.Hoiem and D.Forsyth,Describing objects by their attributes,IEEE CVPR Conference;20-25 June 2009,Miami,FL,USA,pp. 1778-1785.

    [6]D.Parikh and K.Grauman,Relative attributes,IEEE ICCV Conference,6-13 November 2011,Barcelona,Spain,pp.503-510.

    [7]G.Sharma,F.Jurie and C.Schmid,Expanded parts model for human attribute and action recognition in still images,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp.652-659.

    [8]Z.Akata,F.Perronnin,Z.Harchaoui and C.Schmid,Label embedding for attribute based classi fication,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp.819-826.

    [9]A.Bosch,M.Xavier,R.Marti,A review:which is the best way to organize/ classify images by content,J.Image Vis.Comput.25(2007)778-791.

    [10]A.Parkash and D.Parikh,Attributes for classi fier feedback,Springer ECCV Conference,7-13 October 2012,Florence,Italy,pp.354-368.

    [11]S.Ma,S.Sclaroff and N.I.Cinbis,Unsupervised learning of discriminative relative visual attributes,Springer ECCV Conference,7-13 October 2012, Florence,Italy,pp.61-70.

    [12]Y.Wang and G.Mori,A discriminative latent model of object classes and attributes,Springer ECCV Conference,5-11 September 2010,Crete,Greece,pp. 155-168.

    [13]C.Wah and S.Belongie,Attribute-based detection of unfamiliar classes with humans in the loop,IEEE CVPR Conference,23-28 June 2013,Portland,OR, USA,pp.779-786.

    [14]M.Rastegari,D.Parikh and A.Farhadi,Multi-attribute queries:to merge or not to merge,IEEE CVPR Conference,23-28 June 2013,Portland,OR,USA,pp. 3310-3317.

    [15]A.Biswas and D.Parikh,Simultaneous active learning of classi fiers and attributes via relative feedback,IEEE CVPR Conference,23-28 June 2013, Portland,OR,USA,pp.644-651.

    [16]A.Kovashka,K.Grauman,Discovering attribute shades of meaning with the crowd,Int.J.Comput.Vis.114(2015)56-73.

    [17]D.Jayaraman and K.Grauman,Zero-shot recognition with unreliable attributes,Neural Information Processing Systems Conference,8-13 December 2014,Montreal,Canada,pp.3464-3472.

    [18]C.Chao-Yeh and K.Grauman,Inferring analogous attributes,IEEE CVPR Conference,23-28 June 2014,Columbus,OH,USA,pp.200-207.

    [19]L.Liang and K.Grauman,Beyond comparing image pairs:setwise active learning for relative attributes,IEEE CVPR Conference,23-28 June 2014,Columbus,OH,USA,pp.208-215.

    [20]S.Jing,K.Kai,C.L.Chen,and W.Xiaogang,Deeply learned attributes for crowded scene understanding,IEEE CVPR Conference,7-12 June 2015,Boston,MA,USA,pp.4657-4666.

    [21]M.H.Kiapour,H.Xufeng,L.Svetlana,C.B.Alexander and L.B.Tamara,Where tobuy it:matching street clothing photos in online Shops,IEEE ICCV Conference, 7-13 December 2015,Santiago,CA,USA,pp.3343-3351.

    [22]C.Xinlei and G.Abhinav,Webly supervised learning of convolutional networks,IEEE ICCV Conference,7-13 December 2015,Santiago,CA,USA,pp. 1431-1439.

    [23]M.Karayel and N.Arica,Random attributes for image classi fication,IEEE SIU Conference,24-26 April 2013,Girne,TRNC,pp.1-4.

    [24]N.Kumar,A.C.Berg,P.N.Belhumeur and S.K.Nayar,Attribute and simile classi fiers for face veri fication,IEEE ICCV Conference,29 September-2 October 2009,Kyoto,Japan,pp.365-372.

    [25]F.X.Yu,L.Cao,R.S.Feris,J.R.Smith and S.Chang,Designing category-level attributes for discriminative visual recognition,IEEE CVPR Conference, 23-28 June 2013,Portland,OR,USA,pp.771-778..

    [26]T.Ran,W.M.S.Arnold and C.Shih-Fu,Attributes and categories for generic instance search from one example,IEEE CVPR Conference,7-12 June 2015, Boston,MA,USA,pp.177-186.

    [27]E.Victor,C.N.Juan and G.Bernard,On the relationship between visual attributes and convolutional networks,IEEE CVPR Conference,7-12 June 2015, Boston,MA,USA,pp.1256-1264.

    [28]S.Zhiyuan,Y.Yongxin,M.H.Timothy and X.Tao,Weakly supervised learning of objects,attributes and their associations,Springer ECCV Conference,6-12 September 2014,Zurich,Switzerland,pp.472-487.

    [29]C.C.Chang,C.J.Lin,LIBSVM:a library for support vector machines,ACM T Intel.Syst.Tech.27(2011)1-27.

    [30]J.Choi,M.Rastegari,A.Farhadi and L.S.Davis,Adding unlabeled samples to categories by learned attributes,IEEE CVPR Conference,23-28 June 2013, Portland,OR,USA,pp.875-882.

    [31]A.Shrivastava,S.Singh and A.Gupta,Constrained semi-supervised learning using attributes and comparative attributes,Springer ECCV Conference,7-13 October 2012,Florence,Italy,pp.369-383..

    [32]A.Coates,Y.N.G.Andrew,H.Lee,An analysis of single-layer networks in unsupervised feature learning,J.Mach.Learn Res.15(2011)215-223.

    [33]E.Ergul,S.Erturk and N.Arica,Unsupervised relative attribute extraction,IEEE SIU Conference,24-26 April 2013,Girne,TRNC,pp.1-4.

    [34]E.Ergul,M.Karayel,O.Timus,E.Kiyak,Unsupervised feature learning for midlevel data representation,J.Nav.Sci.Eng.12(2016)51-79.

    Emrah Ergulreceived the B.S.degree in Electrical and Electronics Engineering from Turkish Naval Academy, Istanbul,Turkey,in 2001.From 2001 to 2007,he served in the Navy Fleet Command,Kocaeli,Turkey as a weapon systems and electronics of ficer.He received the M.S.degree in Computer Engineering from Naval Science and Engineering Institute,Istanbul,Turkey,in 2009.He also studied at the Computer Science Dept.of University of Shef field,UK for one semester with respect to Erasmus Student Exchange Program while pursuing the M.S.degree.He studied in the Computer Vision and Robotics Laboratory at the Beckman Institute of the University of Illinois,Urbana-Champaign IL,USA,between 2013 and 2014 with a scholarship granted by The Scienti fic and Technological Research Council of Turkey.He received the Ph.D.degree in Electronics and Communication Engineering from Kocaeli University,Kocaeli,Turkey in 2016.He has been working for Navy Inventory Control Center,Kocaeli,Turkey since 2009 for stock estimation and process control.His research interest includes computer vision for image retrieval and classi fication,logistics decision support systems and data exchange standards for inventory planning,pattern recognition analysis,data mining and machine learning applications for the recommender systems.

    E-mail addresses:106103002@kocaeli.edu.tr,emergul13@yahoo.com.

    Peer review under responsibility of Chongqing University of Technology.

    http://dx.doi.org/10.1016/j.trit.2017.01.001

    2468-2322/?2017 Chongqing University of Technology.Production and hosting by Elsevier B.V.This is an open access article under the CC BY-NC-ND license(http:// creativecommons.org/licenses/by-nc-nd/4.0/).

    Incremental learning

    Relative attribute

    Visual recognition

    春色校园在线视频观看| 波多野结衣高清无吗| 嫩草影视91久久| 亚洲自拍偷在线| 久久这里只有精品中国| 国产高清有码在线观看视频| 在线看三级毛片| 又黄又爽又刺激的免费视频.| h日本视频在线播放| 国产三级中文精品| 亚洲av免费在线观看| 精品久久久久久久末码| 国产大屁股一区二区在线视频| 99久久九九国产精品国产免费| 最好的美女福利视频网| 69人妻影院| 三级毛片av免费| 日本一本二区三区精品| 国产女主播在线喷水免费视频网站 | 国产午夜精品久久久久久一区二区三区 | 国内精品宾馆在线| 国产色婷婷99| 一进一出抽搐gif免费好疼| 亚洲av免费在线观看| 99riav亚洲国产免费| 亚洲人成网站在线播| 亚洲欧美成人精品一区二区| 2021天堂中文幕一二区在线观| 成人特级av手机在线观看| 亚洲欧美日韩高清专用| av专区在线播放| 寂寞人妻少妇视频99o| 亚洲综合色惰| 成人无遮挡网站| 欧美另类亚洲清纯唯美| 在线a可以看的网站| 亚洲欧美成人综合另类久久久 | 色综合站精品国产| 麻豆乱淫一区二区| 不卡视频在线观看欧美| 搡女人真爽免费视频火全软件 | 中文资源天堂在线| 亚洲精品日韩av片在线观看| 少妇猛男粗大的猛烈进出视频 | 看十八女毛片水多多多| 久久这里只有精品中国| 久久国内精品自在自线图片| 国产av麻豆久久久久久久| 我的老师免费观看完整版| 国产精品人妻久久久影院| 给我免费播放毛片高清在线观看| 日日摸夜夜添夜夜添小说| 1000部很黄的大片| 哪里可以看免费的av片| 少妇熟女欧美另类| 久久久久久久久久成人| 午夜a级毛片| 69av精品久久久久久| 国产单亲对白刺激| 国产欧美日韩精品一区二区| 可以在线观看的亚洲视频| 一个人观看的视频www高清免费观看| 一级毛片我不卡| 狂野欧美白嫩少妇大欣赏| 欧美极品一区二区三区四区| 亚洲中文日韩欧美视频| 国语自产精品视频在线第100页| 日韩人妻高清精品专区| 少妇人妻一区二区三区视频| 国产真实伦视频高清在线观看| 波多野结衣高清作品| 欧美激情在线99| 日韩 亚洲 欧美在线| 亚洲天堂国产精品一区在线| 亚洲av中文字字幕乱码综合| 久久精品91蜜桃| 亚洲人成网站在线播放欧美日韩| 18禁黄网站禁片免费观看直播| 变态另类丝袜制服| 91久久精品国产一区二区成人| 国产黄色视频一区二区在线观看 | 欧美人与善性xxx| 日韩,欧美,国产一区二区三区 | 亚洲av不卡在线观看| 男女那种视频在线观看| 欧美日韩国产亚洲二区| 亚洲欧美精品自产自拍| 波多野结衣高清无吗| 蜜臀久久99精品久久宅男| 国内揄拍国产精品人妻在线| 长腿黑丝高跟| 国产黄色小视频在线观看| 男女边吃奶边做爰视频| 欧美激情在线99| 日韩高清综合在线| 在线看三级毛片| 国产又黄又爽又无遮挡在线| 国产伦在线观看视频一区| 亚洲国产高清在线一区二区三| 少妇人妻一区二区三区视频| 久久久久国内视频| 九九在线视频观看精品| 亚洲av中文av极速乱| 在线免费十八禁| 别揉我奶头~嗯~啊~动态视频| 美女内射精品一级片tv| 国产精品福利在线免费观看| 精品熟女少妇av免费看| 在线观看免费视频日本深夜| 又黄又爽又刺激的免费视频.| 青春草视频在线免费观看| 国产毛片a区久久久久| 婷婷精品国产亚洲av在线| 国产一区二区亚洲精品在线观看| 麻豆精品久久久久久蜜桃| 成人综合一区亚洲| 99国产极品粉嫩在线观看| 久久久久久大精品| 成人亚洲精品av一区二区| 搞女人的毛片| 久久婷婷人人爽人人干人人爱| 一区二区三区免费毛片| 可以在线观看毛片的网站| 你懂的网址亚洲精品在线观看 | 一进一出抽搐gif免费好疼| 久久久久久伊人网av| 国产精品,欧美在线| 日本爱情动作片www.在线观看 | 亚洲综合色惰| av女优亚洲男人天堂| 观看免费一级毛片| 日韩人妻高清精品专区| 观看美女的网站| 欧美成人免费av一区二区三区| 少妇高潮的动态图| 亚洲av免费在线观看| 亚洲欧美日韩高清专用| 少妇的逼水好多| 日韩,欧美,国产一区二区三区 | 亚洲18禁久久av| 99热只有精品国产| 亚洲18禁久久av| 久久国内精品自在自线图片| 久久人人爽人人片av| 国产单亲对白刺激| 国产精品乱码一区二三区的特点| 亚洲专区国产一区二区| 成人特级黄色片久久久久久久| .国产精品久久| 亚洲精品亚洲一区二区| 欧美一区二区精品小视频在线| 波野结衣二区三区在线| 国产精品一区二区性色av| 国产高潮美女av| 黄色配什么色好看| 欧美日韩国产亚洲二区| 国产免费一级a男人的天堂| 亚洲熟妇熟女久久| 一级av片app| 国语自产精品视频在线第100页| 丰满人妻一区二区三区视频av| 国产精品嫩草影院av在线观看| 别揉我奶头 嗯啊视频| 麻豆国产av国片精品| 国语自产精品视频在线第100页| 免费黄网站久久成人精品| 色播亚洲综合网| 狠狠狠狠99中文字幕| 久久韩国三级中文字幕| 日本一二三区视频观看| av在线亚洲专区| 国内揄拍国产精品人妻在线| 亚洲成人久久爱视频| 国产亚洲精品av在线| 五月伊人婷婷丁香| 成人av在线播放网站| 日韩成人伦理影院| 一进一出好大好爽视频| 热99re8久久精品国产| 一个人看视频在线观看www免费| 亚洲美女视频黄频| 国产美女午夜福利| .国产精品久久| 99在线人妻在线中文字幕| 日本撒尿小便嘘嘘汇集6| 亚洲欧美日韩东京热| 日韩精品青青久久久久久| 日韩一本色道免费dvd| 久久久午夜欧美精品| 国产精品三级大全| 欧美高清性xxxxhd video| 日本黄色视频三级网站网址| 观看美女的网站| 午夜久久久久精精品| 久久久久久九九精品二区国产| 一个人免费在线观看电影| 欧美另类亚洲清纯唯美| 国产乱人视频| 一区二区三区免费毛片| 国产精品亚洲美女久久久| 男女做爰动态图高潮gif福利片| 亚洲av免费高清在线观看| 亚洲成av人片在线播放无| 午夜a级毛片| 美女cb高潮喷水在线观看| 午夜免费男女啪啪视频观看 | 国产乱人视频| 日产精品乱码卡一卡2卡三| 少妇被粗大猛烈的视频| 天堂动漫精品| 在线看三级毛片| 三级经典国产精品| 哪里可以看免费的av片| 高清毛片免费看| 久久久久久九九精品二区国产| 日韩制服骚丝袜av| 久久久精品欧美日韩精品| 精品99又大又爽又粗少妇毛片| 搞女人的毛片| 日本成人三级电影网站| 91午夜精品亚洲一区二区三区| 天天躁夜夜躁狠狠久久av| 日韩人妻高清精品专区| 久久久久久国产a免费观看| 成人特级av手机在线观看| 成人特级av手机在线观看| 国产精品美女特级片免费视频播放器| 少妇高潮的动态图| 国产一区二区三区在线臀色熟女| 亚洲欧美日韩高清在线视频| 久久精品夜夜夜夜夜久久蜜豆| 国产真实乱freesex| 国产探花极品一区二区| 狂野欧美激情性xxxx在线观看| 99久久成人亚洲精品观看| 此物有八面人人有两片| 国内精品宾馆在线| 免费av不卡在线播放| or卡值多少钱| 国产欧美日韩精品一区二区| 国产精品免费一区二区三区在线| 一卡2卡三卡四卡精品乱码亚洲| 久久精品91蜜桃| 免费av毛片视频| 六月丁香七月| 校园春色视频在线观看| 大香蕉久久网| 久久精品久久久久久噜噜老黄 | av女优亚洲男人天堂| 最近最新中文字幕大全电影3| 99视频精品全部免费 在线| 国产蜜桃级精品一区二区三区| 卡戴珊不雅视频在线播放| 欧美高清性xxxxhd video| 国产精品1区2区在线观看.| 成人国产麻豆网| 99riav亚洲国产免费| 国产午夜精品论理片| 黑人高潮一二区| 欧美成人精品欧美一级黄| 一进一出抽搐gif免费好疼| 一区二区三区免费毛片| 色综合色国产| 观看美女的网站| 国产欧美日韩精品一区二区| 十八禁网站免费在线| 亚洲人成网站在线播| 婷婷精品国产亚洲av在线| 99久国产av精品国产电影| 老熟妇乱子伦视频在线观看| 一夜夜www| 日日摸夜夜添夜夜添小说| 丰满的人妻完整版| 亚洲人成网站在线播| av中文乱码字幕在线| 一本一本综合久久| 欧美区成人在线视频| 我要搜黄色片| 午夜日韩欧美国产| 亚洲内射少妇av| 国产综合懂色| 日韩欧美一区二区三区在线观看| 蜜桃亚洲精品一区二区三区| 此物有八面人人有两片| 亚洲人与动物交配视频| 欧美一区二区精品小视频在线| 国产成人影院久久av| 黄色一级大片看看| 91在线精品国自产拍蜜月| 国产精品野战在线观看| 国产精品,欧美在线| 国产精品久久电影中文字幕| 一级a爱片免费观看的视频| 久久精品夜色国产| 国产精品综合久久久久久久免费| 好男人在线观看高清免费视频| 国产黄片美女视频| 女人十人毛片免费观看3o分钟| 国产精品不卡视频一区二区| 国产高清视频在线观看网站| av在线老鸭窝| 亚洲一级一片aⅴ在线观看| 午夜精品国产一区二区电影 | 一个人看视频在线观看www免费| 1000部很黄的大片| 国产精品国产三级国产av玫瑰| 一区二区三区免费毛片| 免费无遮挡裸体视频| 国产精品女同一区二区软件| 91在线观看av| 国产一区二区亚洲精品在线观看| 日韩一区二区视频免费看| 色吧在线观看| 国产在视频线在精品| 色噜噜av男人的天堂激情| 麻豆精品久久久久久蜜桃| 欧美一区二区精品小视频在线| 国产亚洲av嫩草精品影院| 亚洲av二区三区四区| 69人妻影院| 91狼人影院| a级毛片a级免费在线| 精品国内亚洲2022精品成人| 亚洲18禁久久av| 乱码一卡2卡4卡精品| 热99在线观看视频| 内地一区二区视频在线| 国产伦精品一区二区三区视频9| 国产精品伦人一区二区| 自拍偷自拍亚洲精品老妇| 久久精品夜色国产| 欧美又色又爽又黄视频| 少妇熟女欧美另类| 2021天堂中文幕一二区在线观| 久久精品夜色国产| 久久久午夜欧美精品| 精品人妻熟女av久视频| 又粗又爽又猛毛片免费看| 最近在线观看免费完整版| 插逼视频在线观看| 18禁在线无遮挡免费观看视频 | 春色校园在线视频观看| 亚洲电影在线观看av| 成人无遮挡网站| 亚洲精品一卡2卡三卡4卡5卡| 亚洲av熟女| 久久久久久久久久成人| 欧美性感艳星| 国产午夜精品论理片| 中国国产av一级| 成人精品一区二区免费| 舔av片在线| 中文字幕av在线有码专区| 亚洲精品粉嫩美女一区| 国产欧美日韩精品一区二区| 91久久精品国产一区二区三区| 亚洲丝袜综合中文字幕| 色哟哟·www| 啦啦啦韩国在线观看视频| 国产精品人妻久久久影院| 国产 一区 欧美 日韩| 久久久久久伊人网av| 久久精品国产鲁丝片午夜精品| 国产不卡一卡二| 老司机福利观看| 少妇被粗大猛烈的视频| 日本三级黄在线观看| 婷婷六月久久综合丁香| 久久精品国产鲁丝片午夜精品| 久久婷婷人人爽人人干人人爱| 亚洲美女搞黄在线观看 | 青春草视频在线免费观看| 国产欧美日韩精品亚洲av| 国产成人a区在线观看| 我的女老师完整版在线观看| 男女那种视频在线观看| 欧美精品国产亚洲| 99久久精品一区二区三区| 亚洲精品粉嫩美女一区| 日韩在线高清观看一区二区三区| 欧美日韩乱码在线| 欧美绝顶高潮抽搐喷水| 少妇猛男粗大的猛烈进出视频 | 九九久久精品国产亚洲av麻豆| 亚洲人成网站在线播| 日本 av在线| 蜜桃久久精品国产亚洲av| 精品国产三级普通话版| 中文字幕人妻熟人妻熟丝袜美| av免费在线看不卡| 久久久久久久亚洲中文字幕| 国产欧美日韩精品亚洲av| 可以在线观看毛片的网站| 色综合站精品国产| av在线亚洲专区| 综合色丁香网| 色综合色国产| 一本久久中文字幕| 12—13女人毛片做爰片一| 国产黄色视频一区二区在线观看 | 两个人的视频大全免费| 国产黄色视频一区二区在线观看 | 99久久久亚洲精品蜜臀av| 国产成人aa在线观看| 久久精品人妻少妇| 美女高潮的动态| av国产免费在线观看| 亚洲久久久久久中文字幕| 久久久成人免费电影| 久久久a久久爽久久v久久| 国产亚洲91精品色在线| 国产亚洲精品综合一区在线观看| 日本免费a在线| 国产成人freesex在线 | 日本免费一区二区三区高清不卡| 在线免费十八禁| 欧美+亚洲+日韩+国产| 99热这里只有是精品在线观看| 亚洲国产精品sss在线观看| 成熟少妇高潮喷水视频| 99热网站在线观看| 国产精品电影一区二区三区| 舔av片在线| 亚洲电影在线观看av| 最近在线观看免费完整版| 91午夜精品亚洲一区二区三区| 欧美色欧美亚洲另类二区| 午夜影院日韩av| 91精品国产九色| 日韩精品有码人妻一区| 观看免费一级毛片| 一级毛片电影观看 | 国产一区二区三区在线臀色熟女| 亚洲不卡免费看| 国产乱人视频| 亚洲欧美清纯卡通| 不卡视频在线观看欧美| 欧美性猛交黑人性爽| 天堂网av新在线| 国产精品久久久久久精品电影| 在线免费观看不下载黄p国产| 亚洲第一区二区三区不卡| 卡戴珊不雅视频在线播放| 国产精品亚洲美女久久久| 乱人视频在线观看| 久久久精品大字幕| 久久婷婷人人爽人人干人人爱| 亚洲av免费高清在线观看| 美女被艹到高潮喷水动态| 精品午夜福利在线看| 尤物成人国产欧美一区二区三区| 国产伦精品一区二区三区视频9| 国产在线精品亚洲第一网站| 欧美性猛交黑人性爽| 内射极品少妇av片p| 99久久无色码亚洲精品果冻| 寂寞人妻少妇视频99o| 熟女人妻精品中文字幕| 亚洲乱码一区二区免费版| 午夜福利成人在线免费观看| 亚州av有码| 精品久久久久久久人妻蜜臀av| 日本免费a在线| .国产精品久久| 1024手机看黄色片| 亚洲激情五月婷婷啪啪| 成熟少妇高潮喷水视频| 国产精品人妻久久久影院| 18+在线观看网站| 午夜亚洲福利在线播放| 熟女电影av网| 精品无人区乱码1区二区| 国产一区二区在线观看日韩| 国产色爽女视频免费观看| 国产精品日韩av在线免费观看| 亚洲,欧美,日韩| videossex国产| 身体一侧抽搐| 久久99热6这里只有精品| 国产不卡一卡二| 精品久久久久久久久久免费视频| 亚洲性夜色夜夜综合| 网址你懂的国产日韩在线| 国产亚洲av嫩草精品影院| 亚洲精品日韩在线中文字幕 | 国产精品亚洲美女久久久| 精品久久久久久久末码| 高清午夜精品一区二区三区 | aaaaa片日本免费| 日韩av不卡免费在线播放| 亚洲国产色片| av在线蜜桃| 好男人在线观看高清免费视频| 午夜福利在线观看免费完整高清在 | 白带黄色成豆腐渣| 国产黄色小视频在线观看| 日韩成人伦理影院| 最近视频中文字幕2019在线8| 久久久欧美国产精品| 亚洲无线在线观看| 少妇的逼水好多| 一级毛片电影观看 | 哪里可以看免费的av片| 亚洲三级黄色毛片| 精品久久国产蜜桃| 日日摸夜夜添夜夜添小说| 97在线视频观看| 日韩在线高清观看一区二区三区| 欧美一区二区国产精品久久精品| 大又大粗又爽又黄少妇毛片口| 精品乱码久久久久久99久播| 国产久久久一区二区三区| 搡女人真爽免费视频火全软件 | 精品不卡国产一区二区三区| 欧美成人a在线观看| 亚洲最大成人中文| 一个人观看的视频www高清免费观看| 亚洲av二区三区四区| 在线观看午夜福利视频| 国产黄片美女视频| 精品久久久久久久末码| 亚洲成人av在线免费| videossex国产| 欧美日韩综合久久久久久| 高清毛片免费观看视频网站| 精品不卡国产一区二区三区| 午夜老司机福利剧场| 三级男女做爰猛烈吃奶摸视频| 插阴视频在线观看视频| 全区人妻精品视频| 在线观看午夜福利视频| 国产亚洲精品久久久久久毛片| 最后的刺客免费高清国语| 人人妻人人澡人人爽人人夜夜 | 一区二区三区四区激情视频 | .国产精品久久| 国产成人影院久久av| 熟女电影av网| 给我免费播放毛片高清在线观看| 国产极品精品免费视频能看的| 国产精品美女特级片免费视频播放器| av国产免费在线观看| 啦啦啦观看免费观看视频高清| 天美传媒精品一区二区| 三级国产精品欧美在线观看| 精品日产1卡2卡| 国产黄a三级三级三级人| 亚洲国产日韩欧美精品在线观看| 久久综合国产亚洲精品| 三级经典国产精品| 最近手机中文字幕大全| 国内少妇人妻偷人精品xxx网站| 精品久久久久久成人av| 久久久久久大精品| 国产精品亚洲一级av第二区| 97超级碰碰碰精品色视频在线观看| 国产精品免费一区二区三区在线| 国产免费男女视频| av女优亚洲男人天堂| 久久亚洲精品不卡| 看黄色毛片网站| 日韩高清综合在线| 蜜臀久久99精品久久宅男| 乱码一卡2卡4卡精品| 少妇裸体淫交视频免费看高清| 国产午夜福利久久久久久| 极品教师在线视频| 男女边吃奶边做爰视频| 日本成人三级电影网站| 免费看av在线观看网站| 99在线人妻在线中文字幕| 网址你懂的国产日韩在线| 少妇熟女欧美另类| 国产精品免费一区二区三区在线| 日韩国内少妇激情av| av女优亚洲男人天堂| 国产伦精品一区二区三区视频9| 91在线观看av| 久久午夜亚洲精品久久| 国内揄拍国产精品人妻在线| 人人妻人人看人人澡| 日本一二三区视频观看| 日韩中字成人| 成人精品一区二区免费| 亚洲精品国产av成人精品 | 午夜久久久久精精品| 午夜亚洲福利在线播放| 国产爱豆传媒在线观看| 在线播放国产精品三级| 一级毛片久久久久久久久女| 中文亚洲av片在线观看爽| 最近2019中文字幕mv第一页| 国产一区二区三区av在线 | 成人一区二区视频在线观看| 欧美一区二区国产精品久久精品| 特级一级黄色大片| 成人无遮挡网站| 久久热精品热| 国产v大片淫在线免费观看| 国产av一区在线观看免费| av免费在线看不卡| 亚洲丝袜综合中文字幕| 少妇猛男粗大的猛烈进出视频 | 校园人妻丝袜中文字幕| 悠悠久久av| 久久久久久久久久久丰满| 99九九线精品视频在线观看视频| 三级国产精品欧美在线观看| 直男gayav资源| 日韩成人av中文字幕在线观看 | 午夜福利成人在线免费观看| 日韩强制内射视频| 日本-黄色视频高清免费观看| 成熟少妇高潮喷水视频| 噜噜噜噜噜久久久久久91|