• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Two-Order Approximate and Large Stepsize Numerical Direction Based on the Quadratic Hypothesis and Fitting Method

    2020-05-21 05:45:32XiaoliYinChunmingLiandYuanZhang
    IEEE/CAA Journal of Automatica Sinica 2020年3期

    Xiaoli Yin, Chunming Li, and Yuan Zhang

    Abstract—Many effective optimization algorithms require partial derivatives of objective functions, while some optimization problems’ objective functions have no derivatives. According to former research studies, some search directions are obtained using the quadratic hypothesis of objective functions. Based on derivatives, quadratic function assumptions, and directional derivatives, the computational formulas of numerical first-order partial derivatives, second-order partial derivatives, and numerical second-order mixed partial derivatives were constructed. Based on the coordinate transformation relation, a set of orthogonal vectors in the fixed coordinate system was established according to the optimization direction. A numerical algorithm was proposed, taking the second order approximation direction as an example. A large stepsize numerical algorithm based on coordinate transformation was proposed. Several algorithms were validated by an unconstrained optimization of the two-dimensional Rosenbrock objective function. The numerical second order approximation direction with the numerical mixed partial derivatives showed good results. Its calculated amount is 0.2843% of that of without second-order mixed partial derivative. In the process of rotating the local coordinate system 360°, because the objective function is more complex than the quadratic function, if the numerical direction derivative is used instead of the analytic partial derivative, the optimization direction varies with a range of 103.05°. Because theoretical error is in the numerical negative gradient direction, the calculation with the coordinate transformation is 94.71% less than the calculation without coordinate transformation. If there is no theoretical error in the numerical negative gradient direction or in the large-stepsize numerical optimization algorithm based on the coordinate transformation, the sawtooth phenomenon occurs. When each numerical mixed partial derivative takes more than one point, the optimization results cannot be improved. The numerical direction based on the quadratic hypothesis only requires the objective function to be obtained, but does not require derivability and does not take into account truncation error and rounding error. Thus, the application scopes of many optimization methods are extended.

    I. Introduction

    OPTIMIZATION methods are widely used in various research fields and play an important role in scientific and technological problems [1]–[7]. At present, the research of optimization methods is mainly divided into multi-objective[8]–[10], intelligence, genetic algorithm [10], [11], clustering[12] methods, and so on. Classical multidimensional optimization methods are also studied [13].

    Many efficient optimization methods need derivative and partial derivative information of the objective function, such as the linear fitting derivative method, high-order second approximate fixed-point method, tangent intersection point method, negative gradient direction method, zigzag line negative gradient direction method (blind- walking method),second approximate direction method [14], linear fitting gradient method [15], various conjugate direction methods[16]–[18], the optimal available direction method (feasible direction method), half step method [19], and so on. These methods are not useful in optimization problems where the objective function is not differentiable. Numerical differentiation algorithms [20] may allow these methods to work.

    Comparing the above algorithms, the second approximation direction is the best, and the direction of the negative gradient is most commonly used. The negative gradient direction is the fastest local descent direction, and usually deviates from the extreme value point in the whole design space. The conjugate directions constructed on the basis of the negative gradient direction have better optimization efficiency [21], [22].

    The numerical differential algorithm can not only determine the numerical partial derivative, but also obtain the numerical directional derivative in the general direction.

    Looking at the teaching contents of the course “mechanical optimum design” or “optimization methods” and the research literature of optimization methods, and at classical optimization methods, it can be determined that they are all proposed under the quadratic objective function hypothesis and then applied to the general objective function. Although there is an optimization direction based on finite differences[23], the truncation error and rounding error are emphasized while the quadratic hypothesis is neglected due to the fitting effect of the objective function. By representing each derivative in the system of differential equations by its finite difference approximation [24], the quadratic hypothesis is ignored. The research based on the quadratic hypothesis and fitting method is more general.

    Classical optimization methods are not proposed to deal with large-scale matrices. However, some large-scale sparse matrix problems borrow optimization ideas, such as the graph coloring problem [25] and nonlinear equations solving method[26]. Although document [25] deals with the optimization direction based on finite differences, it is far from the research in this paper.

    According to the characteristics of the optimization method,the numerical partial derivative and the numerical directional derivative of the objective function are constructed by using the numerical differentiation algorithm under the quadratic function assumption. On this basis, the optimization effect of the numerical second approximate direction and the local coordinate negative gradient direction is studied.

    II. Construction Method of Numerical First and Two Order Partial Derivatives Based on the Quadratic Hypothesis

    The optimization methods with a clear direction of optimization are put forward under the unimodal assumption.The optimization method based on the quadratic function is universal. There is no universal significance in the study of higher order objective functions. In the design space, the objective function is assumed to be quadratic in any direction,so there are three undetermined coefficients. Take the current pointx(k)as the center and taking pointx(k?1)and pointx(k+1)by a microstep ?sat two sides. Assume that the objective function values of the three points aref(k?1),f(k)andf(k+1),respectively. ?sshall be reasonably determined according to the computer’s expressive ability and termination condition value. Three equations can be obtained from the above three points. Thus the above three parameters are obtained so that the quadratic function of the hypothesis is known. The partial derivatives at pointsx(k?1),x(k)andx(k+1)are shown as follows:

    The above three formulas are consistent with the results obtained by the Lagrange interpolation function method. The second partial derivatives at the three points are equal.

    Equation (4) is consistent with the Newton center difference formula. The numerical gradient and the numerical second partial derivative at the pointx(k)can be obtained from the above 4 formulas, but the second mixed partial derivative cannot be obtained.

    III. Construction Method of Numerical Directional Derivative

    The directional derivative is the rate of change of a function along a certain direction, which is a generalization of the concept of a partial derivative. Letsbe the direction fromx(k)tox(k+1), wherescan be expressed as follows:

    The gradient function of the objective function at pointx(k)is as follows:

    The directional derivative can be regarded as the point product of the gradient vector and the unit directional vector[27], equal to the length of the vector ?f(x) projected ons.

    where cos(?f(x),s) is the cosine of the angle betweenands.kis the optimal serial number. The minute step alongsis decomposed into the minute step alongNcoordinate axes. Then, according to the definition of the partial derivative (7) can be obtained.

    According to the definition of directional derivative change rate, the numerical directional derivative atx(m), the midpoint ofx(k+1)andx(k)is as follows:

    Ifx(k+1)is sufficiently close tox(k), the numerical directional derivative at pointx(k)can be expressed approximately by (8). Calculate the objective function valuef(m)of pointx(m); then the numerical directional derivatives at pointx(k)are as follows:

    A linear equation of gradient vectors can be obtained from(7) and (8). By searching forNpoints near the current pointx(0)and establishing the above equations, a set ofN-ary linear equations can be obtained, and the numerical gradient of the objective function can be solved.

    The numerical negative gradient direction obtained from(10) has some theoretical errors due to the approximate directional derivative of (8). According to the quadratic function hypothesis, the linear equations can be obtained from(7) and (9).

    According to the method of taking points above, the coefficient matrix (10) and (11) is a real diagonal matrix, is non-singular, and its rank isN. It is almost impossible to find the same objective function at each point as that at pointx(0),so (10) and (11) have a unique nonzero solution. In the same way, if the first partial derivative of the objective function is known,Npoints can be found near the current pointx(0), and the aboveN-ary linear equations can be obtained, so that the second partial derivative matrix of the objective function can be solved by the least square method.

    IV. Construction Method of the Numerical Directional Derivative

    Fig. 1. The point drawing for obtaining mixed partial derivative by numerical differentiation.

    Then the second mixed partial derivative is obtained from the following equation.

    According to the quadratic function hypothesis, the relationship between the two coordinates has six undetermined coefficients, and the five points in Fig. 1 can not determine these coefficients, so a numerical point must be added. If we take pointfon the bisector of two coordinate axes, then according to (2) at the midpointgofcf, the partial derivative ofxiis as follows:

    The numerical second mixed partial derivative is derived from the assumption that the objective function is quadratic,so the results of (13) and (14) should be the same.

    The parameters (3), (4), (13) and (14) are not set according to the specific objective function, but determined, accurate and unchanged. Without the value of theoretical analysis,there is no need to discuss truncation error and rounding error.

    V. The Construction Method of a Local Orthogonal Coordinate System for Optimization

    If a local coordinate system is established in the design space, the partial derivative of the objective function can hardly be obtained, but the numerical partial derivative can be obtained using the numerical differential algorithm as in the global coordinate system. After optimizing in a certain direction, it is better to search in the direction of orthogonality again. An orthogonal direction vector group is set up whose last direction is the optimization direction. The orthogonal vector group can be regarded as the coordinate vector group of the local coordinate system. The coordinate vector group is then the coordinate transformation matrixAbetween the two coordinate systems. If the origin of the local coordinate system isx(0)in the fixed coordinate system, then a pointx′in the local coordinate system is shown in the global coordinate system as follows:

    The vectors′in the local coordinate system is shown in the global coordinate system as shown below:

    Set the previous optimization directionSas the last vector αN. Let the largest element ofS(the element with the largest absolute value) be the elementi. Then in theN×Nunit matrix, it corresponds to the column vectori. In order, the column vector of the unit matrix acts as other vectors; thus we can obtain the following linear independent vectors

    The modified Schmitt orthogonalization method is used to perform orthogonalization [28], [29]. We use the following definitions and change vector to unitization vectors with two norm

    Among them,kis an undetermined coefficient. According to the orthogonality between vectors, the following formula can be obtained. We then change the vectors to unitization vectors with two norm

    The square matrix composed of unitized vectors β1,β2,...,βNis the coordinate transformation matrix of (15)and (16).

    The orthogonal vector group can also makeSthe last coordinate axis of the local coordinate system after coordinate transformation, so as to obtain the last column elements ofAaccording to (16). Then, according to the type of coordinate transformation, other elements ofAcan be obtained.

    For a non-quadratic objective function, the optimization direction obtained from the numerical gradient and the second partial derivative matrix constructed by 1, 2, 3 in different local coordinate systems is different.

    VI. Large Stepsize Numerical Optimization Algorithm Based on Coordinate Transformation

    The second approximate direction directly points to the extreme point of the quadratic objective function. Such a numerical direction is good. Another new algorithm is as follows:

    In order to make full use of the optimization effect obtained by the detection point, after each detection of the best point,the best point is taken as the current pointx, and the originalxis taken as the reserve pointx(i)(iis the coordinate axis serial number). If neither probe point is good enough, halve the step size and continue to probe until the step size is less than the termination condition value. The next best point is taken as the reserved pointx(i), and then, from the current point (the best point) detect along the next coordinate axis.After sequential detection alongNcoordinate axes, the numerical negative gradient or numerical second approximation direction is constructed. A step optimization is achieved after one dimensional optimization until the current pointxis close enough to the reserve pointx(i).

    Due to the large distance between the detection points, the numerical first partial derivatives and second partial derivatives have large errors. Instead of calculating the mixed partial derivative and one-dimensional optimization in the new direction, the backward half-step searching algorithm[30] based on the blind path-finding idea is used to find better points according to the current step size. The optimization scheme is shown in Fig. 2.

    Because of the continuity of the feasible region, the initial point can be chosen arbitrarily. If the step size is too large, the optimization direction may deviate too much from the negative gradient direction and the optimization effect would not be good; if the step size is too small, it causes a sawtooth phenomenon because it is too close to the negative gradient direction. Therefore, the initial step size should be determined according to the specific optimization problem, and a better value can be determined by trial calculation.

    VII. Example Verification

    A. The Two-Dimensional Example

    The two-dimensional Rosenbrock function is the four square function, and it is one of the most difficult examples to test the optimization algorithm with. The unconstrained optimization problem whose objective function is as follows:

    The gradient of the objective function is as follows:

    The second partial derivative matrix of the objective function is as follows:

    The objective function consists of two components, each of which has a minimum value of zero. A system of two variables and quadratic equations can be obtained, and the extreme points and extremes can be obtained as follows:

    Fig. 2. The program diagram of large step numerical algorithm based on coordinate transformation.

    Take the initial point and the micro step length as follows:

    The objective function value, numerical gradient and second partial derivative matrix at this point are as follows:

    The relative errors between the numerical solutions obtained from (2), (4), (13), and (14) and the analytical solutions are all less than 1 0?5.

    In this paper, the numerical second approximate direction algorithm and the large step numerical algorithm based on coordinate transformation are realized by theClanguage. The blind searching method [31] is used in one dimensional optimization.

    B. The Test of Numerical Second Approximate Direction Method

    The optimal direction is obtained by constructing the numerical diagonal second partial derivative matrix from (4).After one-dimensional optimization is performed 3955 times,the objective function is calculated 156 517 times. The optimal point and optimal solution are obtained as follows:

    The optimization process is shown in Fig. 3, solid line 1.

    Fig. 3. The seeking process of the numerical second-order approximate direction method.

    The optimal direction is obtained by constructing the numerical symmetric second partial derivative matrix from(4), (13), (14). After one-dimensional optimization is performed 11 times, the objective function is calculated 445 times, and the optimal point and solution are obtained as follows:

    With the numerical second mixed partial derivative, the optimization calculation amount is 0.2843% of without it. The optimization process is shown in the gem-shaped point and dashed line 2 in Fig. 3. The result is similar to that of the analytical second approximate direction method.

    A local coordinate system is established for each optimization.After one-dimensional optimization is performed 244 times, the objective function is called 9884 times, and the optimal point and solutions are obtained as follows:

    Its optimization process is shown in Fig. 3, dotted line 3.Coordinate transformation is not suitable.

    C. The Test of the Constructing Search Direction in Local Coordinate System

    In order to test the influence of coordinate transformation on the numerical optimization direction, a local coordinate system is established withx(0)as the origin, and the local coordinate system is rotated counterclockwise 360°.

    The numerical direction derivative of the fixed coordinate system is used as the numerical partial derivative of the local coordinate system to construct the optimization direction. In the above rotation process, the trajectory of the numerical optimization direction in the local coordinate system is shown in the plus point and dashed line 1 in Fig. 4. It is not evenly distributed, indicating that it is not a fixed direction. In the fixed coordinate system, the trajectory is not a fixed direction,but varies within the range 103.05°, as shown by the asterisk point and dotted line 2 in Fig. 4. It is clear that the calculated results are consistent with those of the 5 analyses.

    Fig. 4. Numerical direction trajectories in the rotation process of local coordinate system.

    D. Test of the Large Step Numerical Optimization Algorithm

    The initial step size, the minimum initial step length, and the termination condition value are as follows:

    In the first optimization round, the following points are detected along theNaxis:

    According to (27), the analytic gradient of the objective function at this point is as follows:

    The numerical gradient obtained by (9) is as follows:

    The difference between them is quite large. Then we obtain theN(2) midpoint, and the numerical gradient obtained by(11) is as follows:

    Using (9), after 1437 detection rounds, the objective function is called 89 155 times, and the optimal point and solutions are obtained as follows:

    The optimization process is shown in Fig. 5, dotted line 1.Adopting (11), the objective function must be called 87 065 times in 1726 detection rounds.

    Fig. 5. The search process of the large step numerical algorithm based on coordinate transformation.

    Then the numerical optimization direction of (4) and (21) is as follows:

    After 1430 detection rounds, we call the objective function 70 621 times to get the optimal point and solution.

    The optimization process is shown in Fig. 5, solid line 2.

    The numerical gradient is obtained by means of 5 coordinate transformation and (9). After 144 detection rounds,we call the objective function 4717 times to get the optimal point and solution

    The optimization process is shown in the dashed line 3 of Fig. 5. Compared with the black dotted line, the computation amount is reduced by 94.71%. Due to the large step size, the numerical negative gradient direction obtained by (9) has a large error. If the numerical direction is rotated 90, after 243 rounds of detection, the objective function is called 7872 times, and the optimal point and solution are obtained as follows:

    The calculation amount is not large, which is consistent with the analysis of 7.3. In the process of approaching, the extremum point is skipped over for several times, but not arrived at step by step. After the numerical negative gradient direction is constructed by (11) and compared with the dotted line 1, the optimal solution is 2.4764×10?15after 14 975 detection rounds and 944 941 calls to the objective function.There is a very serious zigzag phenomenon because it is too close to the analytic negative gradient direction.

    E. The Three-Dimensional Example

    The three-dimensional Rosenbrock function is one of the most difficult multi-dimensional examples to test the optimization algorithm with. The unconstrained optimization problem whose objective function is as follows:

    The gradient of the objective function is as follows:

    The second partial derivative matrix of the objective function is as follows:

    The objective function consists of 4 components, each of which has a minimum value of 0. A system of three variable quadratic equations can be obtained, and the extreme points and extremes can be obtained as follows:we take the initial point as follows:

    Adopting the analytic 2 order approximation direction, after 16 rounds of one-dimensional optimization, the objective function is calculated 614 times, and the optimal point and solution are obtained as follows:

    Adopting the numerical two order approximation direction,if we do not calculate the numerical mixed partial derivative,after 446 rounds of one-dimensional optimization, the objective function is calculated 20 038 times, and the optimal point and solution are obtained as follows:

    Using (13) or (14) to calculate the numerical mixed partial derivative, after 16 rounds of one-dimensional optimization,the objective function is calculated 614 times, and the optimal point and solution are obtained as follows:

    Taking more than one point for each mixed partial derivative, it is then computed by using (13) or (14). After 16 rounds of one-dimensional optimization, the objective function is calculated 660 times, and the optimal point and solution are obtained as follows:

    VIII. Discussion

    The mathematical deduction in this paper is concise and easy to understand. A lot of tedious deduction processes are omitted. The key derivation steps and principles are explained in the text and it is similar to many classical numerical differentiation formulas (difference formula).

    The purpose of this paper is to broaden the scope of existing optimization methods. It should test the approximation between numerical and analytical solutions. If the two approaches are close enough, the optimization effect of the original optimization method will not be affected. The Rosenbrock function is four square, with a curved canyon. At the bottom of the canyon, the objective function contour is close to the parallel, and there are no characteristics of the quadratic function at all. The optimization methods with the optimal direction are all based on the quadratic objective function hypothesis. Some optimization algorithms cannot obtain the effective direction at the bottom of the valley.Therefore, this example is one of the most difficult examples to test the optimization algorithm with. The successful verification of the numerical example is sufficient to prove the optimization effect of numerical optimization direction.

    Conjugate directions needNto expand enough in optimization space, and more than three-dimensional examples are needed to verify the optimization effect. The research in this paper does not involve the verification of conjugate directions. Two-dimensional examples can not only illustrate the problem, but also visually represent the optimization process. Even if the numerical algorithm of the conjugate direction is studied, a three-dimensional numerical example is not necessary, because if the analytical direction is not good, the numerical direction will not be good.

    For multimodal objective functions, different initial points may have different optimization results. For a single peak objective function, the optimization algorithm with a certain direction will not yield different optimal points because of the different initial points. The initial point of the example in this paper is taken on a flat slope on the side of the canyon, which is far from the extreme value point. It not only checks the speed of searching to the canyon bottom, but also checks the effect of approaching the extreme value point at the canyon bottom.

    In this paper, we compare different algorithms using computation and optimization results, and we show that it is not necessary to use an elegant intelligent algorithm. This algorithm is not comparable with an evolutionary algorithm or swarm intelligence optimization algorithm. The latter is an optimization method based on other theories, and does not require gradient information such as in the Powell method[32]. However, there is no concept of optimizing direction; it is an algorithm that seeks order in disorder; it is an algorithm that explores necessity in occasionality. The algorithm has obvious advantages for multi-modal feasible regions, but for single-modal feasible regions, it is not as effective as an algorithm with a definite direction. The numerical direction constructed by the numerical method is not more accurate than that obtained by the analytical method, and both are suitable for a single-peak feasible region.

    IX. Conclusion

    Through the verification of several numerical algorithms with two-dimensional Rosenbrock objective function optimization, the following conclusions can be drawn:

    1) The numerical formulas for calculating the first, second and second mixed partial derivatives have higher accuracy.The numerical first partial derivatives are in agreement with the results obtained by the method of constructing Lagrange interpolation functions. Thus, these numerical formulas are correct. Because it is difficult to obtain second partial derivatives, research in optimization methods that only need the gradient of objective function have been explored. This paper continues the research in this area. At present, there are some research directions involving using numerical direction to improve the accuracy of optimization. The research based on the quadratic hypothesis also ends this direction.

    2) The numerical second approximate direction method with numerical second mixed partial derivatives is the best.

    3) The optimization direction obtained in a local coordinate system has strong randomness, and the large step numerical algorithm requires it.

    4) Based on the quadratic hypothesis and fitting method, the numerical algorithm proposed in this paper allows the derivatives, gradients, second partial derivatives and second mixed partial derivatives of the objective function to remain unresolved, which makes optimization methods that require this information more applicable. There is no derivable objective function for optimization problems in many engineering fields. This study provides more optimization methods for solving these optimization problems.

    5) Because constrained optimization problems are often encountered in control processes, future research on constrained optimization methods may become the focus of our research.

    高清在线国产一区| 无人区码免费观看不卡| 午夜成年电影在线免费观看| 欧美精品啪啪一区二区三区| 精品久久久久久,| 亚洲精品美女久久久久99蜜臀| 国产成人精品久久二区二区91| 99久久国产精品久久久| 色播在线永久视频| 好男人电影高清在线观看| 深夜精品福利| 好男人在线观看高清免费视频 | 国产精品一区二区在线不卡| 黄网站色视频无遮挡免费观看| 99re在线观看精品视频| 久久人妻熟女aⅴ| 极品人妻少妇av视频| 脱女人内裤的视频| 天天躁夜夜躁狠狠躁躁| 午夜福利免费观看在线| 99精品久久久久人妻精品| 最新美女视频免费是黄的| 国产精品99久久99久久久不卡| 精品国产一区二区三区四区第35| 亚洲aⅴ乱码一区二区在线播放 | 精品一区二区三区四区五区乱码| 日日摸夜夜添夜夜添小说| 妹子高潮喷水视频| 国产一区二区三区综合在线观看| 一进一出抽搐动态| 欧美日本视频| 国产精品久久久av美女十八| 91av网站免费观看| 老司机福利观看| 看片在线看免费视频| 亚洲精品国产区一区二| 国产亚洲欧美在线一区二区| 亚洲人成伊人成综合网2020| 咕卡用的链子| 啦啦啦免费观看视频1| 制服丝袜大香蕉在线| av片东京热男人的天堂| 成人永久免费在线观看视频| 久久狼人影院| 99久久精品国产亚洲精品| 欧美色视频一区免费| 香蕉久久夜色| 91老司机精品| 级片在线观看| 婷婷丁香在线五月| 久久性视频一级片| 大码成人一级视频| 色av中文字幕| www.精华液| 欧美另类亚洲清纯唯美| 最新美女视频免费是黄的| 色播在线永久视频| 老熟妇乱子伦视频在线观看| 国产99久久九九免费精品| 国产日韩一区二区三区精品不卡| 国产99久久九九免费精品| 搞女人的毛片| 一区二区三区激情视频| 男女午夜视频在线观看| 国产高清激情床上av| 在线免费观看的www视频| 黄色视频,在线免费观看| 热99re8久久精品国产| 在线国产一区二区在线| 巨乳人妻的诱惑在线观看| av视频在线观看入口| 人妻久久中文字幕网| 欧美日韩瑟瑟在线播放| 一区二区日韩欧美中文字幕| 好男人在线观看高清免费视频 | 无限看片的www在线观看| 无遮挡黄片免费观看| 叶爱在线成人免费视频播放| √禁漫天堂资源中文www| 国产精品99久久99久久久不卡| 国产精品久久久人人做人人爽| 亚洲精华国产精华精| 日韩欧美国产在线观看| 国产一区二区三区视频了| 久久精品影院6| 亚洲人成电影免费在线| 免费在线观看视频国产中文字幕亚洲| av视频免费观看在线观看| 一边摸一边抽搐一进一小说| 伦理电影免费视频| 午夜两性在线视频| 99国产精品免费福利视频| 中文字幕人妻熟女乱码| 午夜日韩欧美国产| 日韩欧美一区视频在线观看| 久久久久国内视频| 精品高清国产在线一区| 亚洲片人在线观看| av福利片在线| 午夜福利影视在线免费观看| 丰满的人妻完整版| 国产国语露脸激情在线看| 亚洲av成人一区二区三| 国产高清激情床上av| 桃红色精品国产亚洲av| 9191精品国产免费久久| 日韩欧美一区视频在线观看| a在线观看视频网站| 我的亚洲天堂| 日韩精品青青久久久久久| av在线天堂中文字幕| 狂野欧美激情性xxxx| 欧美不卡视频在线免费观看 | 日本欧美视频一区| 久久九九热精品免费| 国产99久久九九免费精品| 久久欧美精品欧美久久欧美| 亚洲欧美一区二区三区黑人| 亚洲在线自拍视频| 精品人妻1区二区| 18禁美女被吸乳视频| 国产av一区在线观看免费| 精品国产超薄肉色丝袜足j| 国产一区二区三区综合在线观看| 久久草成人影院| 国产精品免费视频内射| 国产精品一区二区三区四区久久 | 亚洲avbb在线观看| 不卡一级毛片| 亚洲av第一区精品v没综合| 久久伊人香网站| 不卡一级毛片| av视频在线观看入口| 制服人妻中文乱码| 99国产极品粉嫩在线观看| 国产高清有码在线观看视频 | 91国产中文字幕| 精品久久久久久久人妻蜜臀av | 日本一区二区免费在线视频| 老熟妇乱子伦视频在线观看| 亚洲久久久国产精品| 美女 人体艺术 gogo| 老司机深夜福利视频在线观看| 久热这里只有精品99| 久久性视频一级片| 国产熟女xx| 午夜a级毛片| 两性夫妻黄色片| 看片在线看免费视频| 国产精品久久久av美女十八| 69精品国产乱码久久久| 在线十欧美十亚洲十日本专区| av视频免费观看在线观看| 一区在线观看完整版| 在线观看舔阴道视频| 亚洲国产看品久久| 亚洲色图 男人天堂 中文字幕| 韩国av一区二区三区四区| 亚洲国产看品久久| 欧美色视频一区免费| 久久精品亚洲精品国产色婷小说| 天堂动漫精品| 色av中文字幕| 此物有八面人人有两片| 在线观看午夜福利视频| 欧美乱色亚洲激情| 天堂√8在线中文| 欧美黄色淫秽网站| 国产一区二区三区视频了| 国产av精品麻豆| 亚洲国产中文字幕在线视频| 日韩欧美国产一区二区入口| 91国产中文字幕| 黑人操中国人逼视频| 99re在线观看精品视频| 69精品国产乱码久久久| 久久国产精品影院| 性色av乱码一区二区三区2| 黄色丝袜av网址大全| 好男人电影高清在线观看| 久久久久久久精品吃奶| 欧美日本视频| АⅤ资源中文在线天堂| 国产成人啪精品午夜网站| 悠悠久久av| 18禁美女被吸乳视频| 黄片播放在线免费| 久久狼人影院| 极品人妻少妇av视频| 在线观看66精品国产| 久久久久久大精品| 亚洲欧美日韩高清在线视频| 男女下面进入的视频免费午夜 | 婷婷丁香在线五月| 国产精品影院久久| 欧美乱妇无乱码| 91九色精品人成在线观看| 久久精品国产亚洲av香蕉五月| 啪啪无遮挡十八禁网站| 午夜免费鲁丝| 欧美在线黄色| 国产麻豆69| 又紧又爽又黄一区二区| 免费女性裸体啪啪无遮挡网站| xxx96com| 国产不卡一卡二| 一级作爱视频免费观看| 久热爱精品视频在线9| videosex国产| 亚洲在线自拍视频| 国产蜜桃级精品一区二区三区| 免费在线观看黄色视频的| 日韩精品中文字幕看吧| av在线天堂中文字幕| 久久人妻av系列| 久久国产乱子伦精品免费另类| 午夜福利视频1000在线观看 | 禁无遮挡网站| 欧美日韩精品网址| 久久人妻福利社区极品人妻图片| 亚洲色图综合在线观看| 亚洲成人国产一区在线观看| 欧美成人免费av一区二区三区| 久久婷婷成人综合色麻豆| 国产成人精品久久二区二区91| 亚洲 欧美一区二区三区| 日韩 欧美 亚洲 中文字幕| 亚洲一卡2卡3卡4卡5卡精品中文| 一级a爱视频在线免费观看| 啦啦啦 在线观看视频| 日韩欧美在线二视频| 亚洲精品在线美女| 亚洲中文av在线| 欧美色欧美亚洲另类二区 | 日韩av在线大香蕉| 最新美女视频免费是黄的| 成人三级黄色视频| 精品一区二区三区四区五区乱码| 久久久久久免费高清国产稀缺| 午夜亚洲福利在线播放| 一级毛片女人18水好多| 精品乱码久久久久久99久播| 精品久久久久久成人av| 韩国精品一区二区三区| 精品熟女少妇八av免费久了| 少妇 在线观看| 叶爱在线成人免费视频播放| 热re99久久国产66热| 少妇被粗大的猛进出69影院| 亚洲欧美精品综合久久99| 国产亚洲av高清不卡| 久久人妻福利社区极品人妻图片| 在线国产一区二区在线| 亚洲精品国产精品久久久不卡| 日本三级黄在线观看| 欧美日本亚洲视频在线播放| 精品国产一区二区久久| 免费在线观看影片大全网站| 亚洲黑人精品在线| 欧美 亚洲 国产 日韩一| 国产一区二区激情短视频| 天堂影院成人在线观看| 国产又色又爽无遮挡免费看| 老司机午夜十八禁免费视频| 嫩草影视91久久| 禁无遮挡网站| 午夜两性在线视频| 久久亚洲真实| 少妇熟女aⅴ在线视频| 美女高潮到喷水免费观看| 国产aⅴ精品一区二区三区波| 天天躁狠狠躁夜夜躁狠狠躁| 中文字幕人妻熟女乱码| 欧美亚洲日本最大视频资源| 男人的好看免费观看在线视频 | 香蕉久久夜色| 视频区欧美日本亚洲| 九色国产91popny在线| 国产精品香港三级国产av潘金莲| 欧美一级毛片孕妇| 波多野结衣av一区二区av| 一级作爱视频免费观看| 国产午夜福利久久久久久| 亚洲精品中文字幕一二三四区| 免费一级毛片在线播放高清视频 | 免费一级毛片在线播放高清视频 | 亚洲av成人一区二区三| videosex国产| 麻豆一二三区av精品| 久久精品国产清高在天天线| 婷婷丁香在线五月| 国产又爽黄色视频| 大码成人一级视频| 色在线成人网| 免费不卡黄色视频| 精品人妻1区二区| av福利片在线| 国产午夜福利久久久久久| 亚洲aⅴ乱码一区二区在线播放 | 搡老熟女国产l中国老女人| 欧美黑人欧美精品刺激| 精品第一国产精品| 欧美激情 高清一区二区三区| 色综合欧美亚洲国产小说| 午夜视频精品福利| 18美女黄网站色大片免费观看| 无限看片的www在线观看| 久久久国产精品麻豆| 9色porny在线观看| 欧美激情 高清一区二区三区| √禁漫天堂资源中文www| 啦啦啦免费观看视频1| e午夜精品久久久久久久| 好男人电影高清在线观看| 两性夫妻黄色片| 国产精品99久久99久久久不卡| 日韩欧美免费精品| 熟妇人妻久久中文字幕3abv| 午夜福利免费观看在线| 9191精品国产免费久久| 国产一区二区三区在线臀色熟女| 人人澡人人妻人| 黄色女人牲交| 日韩欧美国产在线观看| 亚洲第一青青草原| 久久国产精品人妻蜜桃| 日本精品一区二区三区蜜桃| 中文字幕久久专区| 免费高清视频大片| 精品一区二区三区av网在线观看| 黄片小视频在线播放| 日本三级黄在线观看| av在线播放免费不卡| 婷婷丁香在线五月| av在线播放免费不卡| 欧美一级毛片孕妇| 国产不卡一卡二| 免费不卡黄色视频| 黄网站色视频无遮挡免费观看| 欧美日韩中文字幕国产精品一区二区三区 | 成人亚洲精品av一区二区| 又黄又爽又免费观看的视频| 91麻豆精品激情在线观看国产| 欧美丝袜亚洲另类 | 麻豆一二三区av精品| 国产欧美日韩一区二区三| 亚洲免费av在线视频| 久久精品91蜜桃| 男女下面插进去视频免费观看| 9色porny在线观看| 美女免费视频网站| 在线观看免费日韩欧美大片| 国产亚洲精品av在线| 亚洲熟妇熟女久久| 亚洲成国产人片在线观看| 精品福利观看| 国产成年人精品一区二区| 亚洲人成伊人成综合网2020| 麻豆成人av在线观看| 免费看美女性在线毛片视频| 国产人伦9x9x在线观看| av中文乱码字幕在线| 国产成人影院久久av| 香蕉久久夜色| 欧美黑人欧美精品刺激| 亚洲第一欧美日韩一区二区三区| 亚洲中文字幕一区二区三区有码在线看 | 国产精品日韩av在线免费观看 | 久久亚洲精品不卡| 校园春色视频在线观看| 日本 av在线| 校园春色视频在线观看| 两个人免费观看高清视频| svipshipincom国产片| 亚洲片人在线观看| 91大片在线观看| 1024香蕉在线观看| 50天的宝宝边吃奶边哭怎么回事| 国产一区二区激情短视频| 看片在线看免费视频| 久久久国产精品麻豆| 亚洲国产欧美网| 黄色视频,在线免费观看| 好看av亚洲va欧美ⅴa在| 欧美日韩亚洲综合一区二区三区_| 国产亚洲精品综合一区在线观看 | 黄色视频不卡| 国产99久久九九免费精品| av有码第一页| 午夜福利影视在线免费观看| 大型av网站在线播放| 国语自产精品视频在线第100页| a级毛片在线看网站| 黄色成人免费大全| 国产色视频综合| 丝袜在线中文字幕| 99久久国产精品久久久| 亚洲国产毛片av蜜桃av| 精品日产1卡2卡| 日韩 欧美 亚洲 中文字幕| 法律面前人人平等表现在哪些方面| 一区二区日韩欧美中文字幕| 黑人欧美特级aaaaaa片| 女性生殖器流出的白浆| 亚洲专区字幕在线| 母亲3免费完整高清在线观看| 又紧又爽又黄一区二区| 麻豆国产av国片精品| 999久久久国产精品视频| 国产麻豆成人av免费视频| 人人妻人人澡欧美一区二区 | 日韩欧美三级三区| 亚洲精品av麻豆狂野| 淫秽高清视频在线观看| 少妇熟女aⅴ在线视频| 亚洲国产欧美一区二区综合| 亚洲三区欧美一区| 女警被强在线播放| 美女大奶头视频| 午夜久久久在线观看| 国产精品久久久人人做人人爽| 最新美女视频免费是黄的| 免费高清视频大片| 搡老熟女国产l中国老女人| 日本a在线网址| 97人妻天天添夜夜摸| a在线观看视频网站| 免费在线观看影片大全网站| e午夜精品久久久久久久| 国产一区在线观看成人免费| 欧美大码av| 变态另类丝袜制服| 成人永久免费在线观看视频| 最近最新免费中文字幕在线| 国产成人欧美| e午夜精品久久久久久久| 亚洲国产看品久久| 国产精品一区二区免费欧美| or卡值多少钱| 老熟妇乱子伦视频在线观看| 成年人黄色毛片网站| 伊人久久大香线蕉亚洲五| 首页视频小说图片口味搜索| 一区二区日韩欧美中文字幕| 久久午夜综合久久蜜桃| 久久亚洲精品不卡| 久久久国产成人精品二区| 国产亚洲av嫩草精品影院| 亚洲av片天天在线观看| 国产欧美日韩一区二区三| 亚洲av电影不卡..在线观看| 波多野结衣av一区二区av| 99久久国产精品久久久| 一级a爱视频在线免费观看| av在线天堂中文字幕| 制服人妻中文乱码| 在线永久观看黄色视频| 精品日产1卡2卡| 久久人人精品亚洲av| 人人妻人人澡人人看| 999久久久精品免费观看国产| 亚洲五月婷婷丁香| 精品国产超薄肉色丝袜足j| 18禁裸乳无遮挡免费网站照片 | 精品一区二区三区四区五区乱码| 麻豆久久精品国产亚洲av| www.www免费av| 一二三四在线观看免费中文在| 欧美在线一区亚洲| 757午夜福利合集在线观看| 久久人妻福利社区极品人妻图片| 国产亚洲精品综合一区在线观看 | 老司机深夜福利视频在线观看| 黄色视频不卡| 欧美成人免费av一区二区三区| 在线观看免费视频网站a站| 国内久久婷婷六月综合欲色啪| 国产欧美日韩一区二区三| svipshipincom国产片| 欧美av亚洲av综合av国产av| 狂野欧美激情性xxxx| 男人舔女人的私密视频| av网站免费在线观看视频| 国产精品野战在线观看| 免费高清视频大片| 久久国产精品人妻蜜桃| 国产私拍福利视频在线观看| 一卡2卡三卡四卡精品乱码亚洲| 亚洲欧美日韩无卡精品| 极品人妻少妇av视频| 97碰自拍视频| 国产精品秋霞免费鲁丝片| www.自偷自拍.com| 热99re8久久精品国产| 999久久久国产精品视频| 欧美日韩福利视频一区二区| 女人被躁到高潮嗷嗷叫费观| 青草久久国产| 久久久久久免费高清国产稀缺| 欧美成狂野欧美在线观看| 亚洲中文字幕一区二区三区有码在线看 | 午夜福利在线观看吧| 亚洲欧美日韩另类电影网站| 亚洲国产欧美一区二区综合| 国产精品99久久99久久久不卡| 啪啪无遮挡十八禁网站| 欧美另类亚洲清纯唯美| 丝袜美足系列| 十八禁网站免费在线| 黑人欧美特级aaaaaa片| 人人妻人人爽人人添夜夜欢视频| 九色国产91popny在线| 88av欧美| 成人欧美大片| 亚洲精品国产精品久久久不卡| 免费一级毛片在线播放高清视频 | 成人三级做爰电影| 女人高潮潮喷娇喘18禁视频| 日本vs欧美在线观看视频| 99国产极品粉嫩在线观看| 国产成人av教育| 成人国产一区最新在线观看| 国产精品美女特级片免费视频播放器 | 国产成人精品无人区| 久久精品亚洲精品国产色婷小说| 亚洲中文字幕一区二区三区有码在线看 | 日韩大码丰满熟妇| 999精品在线视频| 热99re8久久精品国产| 国产三级在线视频| 婷婷丁香在线五月| 久热这里只有精品99| 中文字幕久久专区| 大陆偷拍与自拍| 激情在线观看视频在线高清| 午夜免费激情av| 18禁观看日本| 久久久久国产一级毛片高清牌| 国产av一区二区精品久久| 日本三级黄在线观看| 久久欧美精品欧美久久欧美| 亚洲人成电影免费在线| 人人妻,人人澡人人爽秒播| 女人被狂操c到高潮| 国产亚洲精品第一综合不卡| 亚洲欧美精品综合久久99| 午夜福利一区二区在线看| 精品第一国产精品| АⅤ资源中文在线天堂| 色综合婷婷激情| 午夜影院日韩av| 可以免费在线观看a视频的电影网站| av在线播放免费不卡| 黄色丝袜av网址大全| 日韩欧美三级三区| 国产人伦9x9x在线观看| 国产精品久久久av美女十八| 桃色一区二区三区在线观看| 亚洲第一av免费看| 精品一区二区三区四区五区乱码| 在线观看www视频免费| 国产熟女午夜一区二区三区| 亚洲伊人色综图| 欧美一级毛片孕妇| 他把我摸到了高潮在线观看| 国产精品一区二区免费欧美| 国产三级黄色录像| 国产精品亚洲美女久久久| 亚洲黑人精品在线| 在线观看www视频免费| 热99re8久久精品国产| 亚洲国产精品sss在线观看| 久久精品国产亚洲av高清一级| 麻豆成人av在线观看| 麻豆国产av国片精品| 国产野战对白在线观看| 午夜成年电影在线免费观看| 亚洲熟妇中文字幕五十中出| 色综合婷婷激情| 亚洲av熟女| 亚洲电影在线观看av| 欧美日韩福利视频一区二区| 涩涩av久久男人的天堂| 午夜成年电影在线免费观看| 国产一卡二卡三卡精品| 黄色毛片三级朝国网站| 村上凉子中文字幕在线| 日本在线视频免费播放| 长腿黑丝高跟| 国产熟女xx| 久久久久国内视频| 欧美+亚洲+日韩+国产| 精品午夜福利视频在线观看一区| 两个人免费观看高清视频| 国产成人免费无遮挡视频| 这个男人来自地球电影免费观看| 欧美av亚洲av综合av国产av| 国产1区2区3区精品| 在线观看午夜福利视频| 日本欧美视频一区| 国产成人精品久久二区二区免费| 高清毛片免费观看视频网站| 99re在线观看精品视频| 91在线观看av| 国产成人影院久久av| 精品国产一区二区三区四区第35| 亚洲av片天天在线观看| 国产av又大| 国产精品久久久久久精品电影 | 美女大奶头视频| 成人精品一区二区免费| 成人手机av| 一区二区日韩欧美中文字幕| 国产成+人综合+亚洲专区| 国产精品美女特级片免费视频播放器 |