• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    A deep reinforcement learning (DRL) based approach for well-testing interpretation to evaluate reservoir parameters

    2022-03-30 13:52:22PengDongZhiMingChenXinWeiLioWeiYu
    Petroleum Science 2022年1期

    Peng Dong ,Zhi-Ming Chen ,,* ,Xin-Wei Lio ,Wei Yu

    a State Key Laboratory of Petroleum Resources and Prospecting,China University of Petroleum Beijing (CUP),Beijing,102249,China

    b University of Texas at Austin,Austin,78731,Texas,USA

    Keywords:Well testing Deep reinforcement learning Automatic interpretation Parameter evaluation

    ABSTRACT Parameter inversions in oil/gas reservoirs based on well test interpretations are of great significance in oil/gas industry.Automatic well test interpretations based on artificial intelligence are the most promising to solve the problem of non-unique solution.In this work,a new deep reinforcement learning(DRL)based approach is proposed for automatic curve matching for well test interpretation,by using the double deep Q-network (DDQN).The DDQN algorithms are applied to train agents for automatic parameter tuning in three conventional well-testing models.In addition,to alleviate the dimensional disaster problem of parameter space,an asynchronous parameter adjustment strategy is used to train the agent.Finally,field applications are carried out by using the new DRL approaches.Results show that step number required for the DDQN to complete the curve matching is the least among,when comparing the naive deep Q-network (naive DQN) and deep Q-network (DQN).We also show that DDQN can improve the robustness of curve matching in comparison with supervised machine learning algorithms.Using DDQN algorithm to perform 100 curve matching tests on three traditional well test models,the results show that the mean relative error of the parameters is 7.58%for the homogeneous model,10.66%for the radial composite model,and 12.79%for the dual porosity model.In the actual field application,it is found that a good curve fitting can be obtained with only 30 steps of parameter adjustment.

    1.Introduction

    Well test analysis plays an important role in understanding the characteristics of wellbore,reservoir and boundary,especially in the discovery and development of some important oil and gas fields(Yao and Ge,2011;Chen et al.,2018;Mohammed et al.,2020).One of the most important means of well test interpretation is the type curve match.In the early days,curve matching was performed based on the chart(Earlougher,1977;Horne,1995).The parameters are obtained by moving the measured curve to find the most consistent theoretical curve on the chart.Since the number of curves on the chart is limited,this method will cause large errors.At present,manual parameter adjustment assisted by computer software become the primary method for curve matching(Bourdet,2002).However,the human bias,non-unique solutions,and low efficiency caused by artificial interpretation cannot adapt to the rapid development of oil and gas fields (AlMaraghi and El-Banbi,2015).

    In recent decades,with the improvement of calculation technology and test instrument precision,many optimization algorithms have been gradually applied to the automatic matching of well test curve.Among them,nonlinear regression is the classical method,and the least square method represented by Levenberg-Marquardt method is the most common solution method (Nanba and Horne,1992;Dastan,2010;Dastan and Horne,2011).However,the method is greatly affected by the initial value,and the parameters obtained are locally optimal.Therefore,some global optimization algorithms,such as genetic algorithm and particle swarm optimization algorithm,are also applied to the automatic matching of well test curve (Guyaguler et al.,2001;Gomez et al.,2014;Awotunde,2015).But the global optimization algorithm has the problem of low efficiency.

    With the great progress made in computer science in recent years,artificial intelligence (AI) algorithms have been widely used in the oil and gas industry (Zhu et al.,2019;Liu et al.,2020;Gao et al.,2021;Huang et al.,2021).Meanwhile,AI algorithms are also tried for automatic well test interpretation.Al-Kaabi and Lee(1990)firstly deployed artificial neural network (ANN) in well test model recognition.They used a number of independent ANN to calculate the probability that the curve fell into each type of well test model.Adibifard et al.(2014)trained an ANN to automatically identify well test curve parameters.The coefficients interpolated by Chebyshev polynomial on pressure derivative data are used as the input of ANN to improve the accuracy of parameter identification.Li et al.(2020) used the convolutional neural network to train the deep learning model for identifying curve parameters by taking the complete curve as input,without manually extracting curve features.In addition,the application of deep learning on well-test interpretation for identifying pressure behavior and characterizing reservoirs was also performed (Dong et al.,2021).However,training neural network requires the acquisition or synthesis of a large number of samples,and the interpretability of the parameter inversion results is weak.Sometimes unrealistic interpretation results are prone to appear and the reasons cannot be known(Zhang and Zhu,2018).To make the parameter inversion results more interpretable and reasonable,this work proposes for the first time a robust automatic well test curve matching method based on reinforcement learning (RL).

    RL is inspired by relevant principles of animal psychology(Sutton and Barto,2018).By imitating the trial-and-error mechanism of humans or animals,the agent can interact with the environment and learn the mapping relationship between state and behavior to obtain the maximum cumulative expected return.Compared with supervised learning algorithms,RL algorithms have the potential to achieve results beyond human performance through active learning,exploration and exploitation.Because of this,RL tends to be a harder learning task.Among RL algorithms,Q learning is one of the most popular(Gao et al.,2020).Recently,the original DRL algorithm naive DQN and its improved algorithm DQN,which combines Q learning with deep neural network,have been introduced and applied into Atari games to achieve automatic control at or beyond the human level (Mnih et al.,2013,2015).However,these two algorithms lead to overoptimistic value estimates,so van Hasselt et al.(2016) proposed DDQN algorithm to alleviate this problem.DDQN algorithm has been successfully used for battery energy storage system,power system,and stock trading(Arulkumaran et al.,2017;Bui et al.,2019;Shi et al.,2021).These studies show that the DDQN algorithm can avoid the agent trapped into local optimization and is suitable for the environment with large state space.

    At present,only limited work about RL has been done in the oil and gas industry.Hourfar et al.(2019) used the DRL algorithm to optimize reservoir water injection.By allowing the agent to dynamically adjusting the water injection rate,higher NPV can be obtained than the traditional optimization method.Miftakhov et al.(2020) use reservoir pressure and water saturation distribution as direct observations to train agents to optimize injection and production parameters.Li and Misra (2020) transformed the history matching problem into a continuous decision problem,and adapted the reservoir permeability using a RL algorithm to achieve automatic history fitting.Guevara et al.(2018)used RL to optimize the gas injection strategy in steam-assisted gravity drainage process.The field test results show that the RL method increases the NPV by at least 30%and reduces the calculation cost by more than 60%.Unfortunately,although the RL has lots of advantages,few work about RL has been done in well test interpretation.

    In this study,the DDQN algorithm based agent learns to match the well test data in optimal number of steps by iteratively adjusting the parameters of well test model.This work represents,to our knowledge,the first application of DRL approach for automatically match the well test curves.The new proposed method has a fast speed and reliable results in automatic well test interpretation,which is of great significance to improve the repeatability of well test interpretation.

    2.Theory

    2.1.Reinforcement learning

    The reinforcement learning (RL) is an important paradigm of machine learning,whose goal is to find an optimal strategy to obtain the largest cumulative expected return by train an agent.Markov Decision Process(MDP)is a basic theoretical framework to solve the problem of RL.Within this framework,machines that learn and implement decisions are called agent.Anything outside the agent that interacts with it is called the environment.In the interaction process,the agent observes the statesof the current environment and chooses the actionan under a certain policy π,the environment responds to the action,and the new states′and rewardrwill be fed back to the agent.Therefore,assuming that starting from the initial states0,executing the MDP will result in a sequence,s0,a0,r0,s1,a1,r1,...,sn,an,rn.

    The agent's job is to optimize the policy for taking action to maximize the cumulative expected return(Sutton and Barto,2018).The return at steptis the sum of the discount rewardswhere γ∈[0,1 is discount rate,which determines the present value of future rewards.In RL,it is the most important method to train agents for solving MDP problem based on action value functionQπ(s,a).Qπ(s,a)represents the expected return on the actionataken in accordance with policy π at states,as in Eq.(1).

    Qπ(s,a)calculates the value of the action in a certain state.Simply speaking,Qπ(s,a)expresses how good it is for an agent to be in a certain state (Sun,2020).Therefore,the optimal strategy is based on the optimal value of the action.Specifically,when the optimal action value functionQ*(s,a)=is obtained,the optimal policy π*(s)=(s,a)is to pick the action corresponding to the maximumQ*(s,a)in each state.In general,Q*(s,a)can be solved by Bellman optimality equation (Sutton and Barto,2018),as Eq.(2),which shows the relationship between the current optimal action value function and the subsequent optimal action value function.

    wheres′is the next state reached after taking actiona,anda′is the action taken in the next state.As long as iterates on Eq.(2),it will eventually converge to the optimal action value functionQ*(s,a)(Watkins and Dayan,1992).

    However,the action value function described above can only represent discrete states.For a continuous state space,like the well test curve,the function approximator,denoted asQ(s,a;θ)≈Q*(s,a),must be used to estimate the action value function.Q(s,a;θ)is usually designed as a neural network (NN) with weights θ,as shown in Fig.1,which method is called deep reinforcement learning(DRL).This study uses the DDQN algorithm,a value based DRL algorithm,proposed by van Hasselt et al.(2016) to train the agent for curve matching,and the details will be discussed in Section 3.

    Fig.1.Schematic diagram of NN function approximator.Wherein,the input layer accepts the state s consisting of the target and the predicted derivative pressure curve,and the output layer returns the action value Q(s,a).

    2.2.Well test model

    In this work,three classical well test models are used to construct an environment for training agents respectively,and the performance of agents in different environments is evaluated to verify the generalization ability of the proposed method.The three well test models are:homogeneous model,dual-porosity model,and radial composite model.To facilitate analysis,the parameters are dimensionless and are defined in Appendix 1.This section briefly introduces these three well test models.

    2.2.1.Homogeneous model

    It is assumed that a production well in homogeneous infinite formation is affected by wellbore storage effect and skin effect.According to the solution of Lee et al.(2003),the dimensionless bottom hole pressurepwDis a function oftD,CD,andS,which can be denoted as follows:

    whereCDis the dimensionless wellbore storage,Sis the skin factor,tDis the dimensionless time.Therefore,the required inversion parameters areCD,S.A typical log-log plots of bottomhole pressure and its derivative curve is shown in Fig.2.The derivative curve type used in this work is Bourdet pressure derivative curve (Bourdet et al.,1984).

    Fig.2.The typical well test curves for homogeneous model.

    2.2.2.Dual-porosity model

    It is assumed that there are two types of pore media in the formation:the fracture system,which is the fluid flow channel,and the matrix rock system,which is the fluid reservoir space.According to the solution of Lee et al.(2003),thepwDis a function oftD,CD,S,ω,and λ,which can be expressed as follows:

    where ω is the storativity ratio,which represents the storage capacity of the fracture;λ is the inter-porosity flow factor,which represents the communication between the fracture and the matrix;φ is porosity;Ctis the total compressibility,MPa-1;ais the parameter characteristic of the system geometry;rwis the well radius,m;kis permeability,mD;Subscript m represents matrix;Subscript f represents fracture.Therefore,the parameters need to be inverted areCD,S,ω,and λ.A typical log-log plots of pressure and its derivative curve is shown in Fig.3.

    2.2.3.Radial composite model

    It is assumed that there are two seepage zones with different physical parameters in the formation and there is no additional pressure drop at the interface of the two zones.According to the solution of Chu and Shank(1993),ThepwDis a function oftD,CD,S,M,andRi,which can be denoted as follows:

    with

    Fig.3.The typical well test curves for dual-porosity model.

    Fig.4.The typical well test curves for radial composite model.

    whereMis mobility ratio;Riis inner zone radius,m;μ is viscosity,mPa·s;Subscript ir represents inner zone;Subscript er represents outer zone.Therefore,the parameters to be inverted areCD,S,M,andRi.A typical log-log plots of pressure and its derivative curve is shown in Fig.4.

    3.Method

    Based on concepts of DRL,the well test curve matching process can be regarded as an MDP.Therefore,the automated well test curve matching process can be achieved by applying DRL algorithm.Fig.5 is a schematic diagram of automatic curve matching process based on MDP.In this process,the target and predicted pressure derivative curve are considered as statestat first.Here,the target curvecomes from the measured curve,and the predicted curveis calculated from the well test model.Following that,the agent provides the actionatof parameter adjustment and updates the policy based on the rewardrt.Next,the environment based on the well test (WT) model receives actionatand then updates the parameters of the WT model.Finally,the environment outputs new statesst+1and rewardsrt+1back to the agent.In this way,the agent continuously interacts with the environment to obtain more rewards,which will make the prediction parameters continuously update to the target parameters to complete the curve matching.

    In the remainder of this section,we will detail the agent and environment of the study in turn.

    Fig.5.Schematic diagram of agent interacting with the environment in MDP.The environment is encapsulated by the well test model (WT model).

    3.1.Agent

    This section will introduce the double deep Q-network(DDQN)algorithm used to train the agent,the policy used when the agent interacts with the environment,and the discrete mode of action space to avoid the curse of dimensionality in the parameter adjustment process.

    3.1.1.DDQN algorithm

    As noted earlier,in this work the double deep Q-network(DDQN)(van Hasselt et al.,2016)algorithm is considered.DDQN is a value-based DRL algorithm that can be used with continuous state spaces.DDQN uses the policy based on the online Q-networkQ(s,a;θ)to select actions that interact directly with the environment,and the target Q-networkQ(s,a;θ′)to evaluate the selected actions separately,as shown in Fig.6.Separating the action selection from the evaluation can alleviate the overly optimistic estimate of the value of the action and improves the accuracy of the action value estimation.In addition,to improve the utilization efficiency of data,the transition tuples at timestept,et=(st,at,rt,)generated by the interaction between the agent and the environment is stored in a buffer called experience replay memory Dt={e1,e2,...,et} (Mnih et al.,2015),as Fig.6.Therefore,by extracting experiments randomly in D,it is possible to use minibatch samples to update these two Q-networks.More precisely,the loss function,as Eq.(6),is used to updates the weights θ of online Q-networks,while the weights θ’of target Q-network is a delayed copy of the online Qnetworks,which is to copy after a certain number of training.The complete training process is discussed in Section 3.3.

    where θ is the weight of online Q-networks,θ′is the weight of target Q-networks,γ is discount factor.The notation(s,a,r,s′)U(D)represents a minibatch samples are uniformly sampled from the replay memory.

    Once the training of the agent is completed,the online Q-networks is used for prediction.For a particular state,the online Qnetworks will provide a Q-value for each of the possible actions for that state.Finally,the action is determined by the policy in terms of Q-value,which will be discussed in detail later.

    3.1.2.Policy

    Fig.6.Schema of the DDQN model in training process.

    When agents interact with the environment,an important problem is to trade-off exploration and exploitation.Exploitation means that the agent takes the most valuable action it estimates every time,and exploration means that the agent randomly takes actions to explore the environment.Proper exploration will help the agent find better policy and improve the accuracy of the action value estimation.However,the agent cannot simultaneously exploration and exploitation in the same action selection.Therefore,the selection of action needs to obey a certain probability distribution to make exploration and exploitation alternate.In Section 5.1,we discussed the influence of different policies on the training process.The results show that Boltzmann exploration policy(Derhami et al.,2008)can better balance the exploration and exploitation of agents in the curve matching task.In the Boltzmann exploration policy,the probability distribution of the agent's action selection is determined by Eq.(7).

    where τ ∈(0,1 is the temperature parameter,which controls the certainty of the action.

    3.1.3.Curse of dimension

    The DRL algorithm based on value functionQ(s,a)needs to clarify the actions that an agent can perform,which means that the action space of the agent is discrete.Therefore,when it is applied to a high-dimensional continuous action space,such as well test curve matching problem,the continuous action needs to be discrete.However,this will lead to huge combinatorial increase in the number of actions with the number of well-test parameter dimensions.For example,for the homogeneous model,it is assumed that the parameters to be inverted are dimensionless well storage coefficientCDand skin coefficientS.If the parameter space is discretized into 100 intervals,the action space of the two parametersCDandSwill reach 1002.For the agent,the value of this 1002action needs to be calculated,that is,the output dimension of the Q network is 1002.Generally,if the number of parameters to be inverted isNand the number of parameter discrete spaces isnd,then the total action space to be considered isWhen the parameter precision requirement is high or the number of parameters is large,the agent will not be able to handle the resulting huge action space.To alleviate the problem,an asynchronous parameter tuning method is proposed.To be more specific,the agent only adjusts one well test parameter at a time,and the action for each parameter is only set to increase and decrease the value of the parameter,and the step length of the increase or decrease is set to a fixed value.Obviously,the total action space that needs to be considered is only 2Nbased on this approach.This allows the current discrete action DRL algorithm to be applied to well test curve matching task.In this work,three typical well test models are used to verify the proposed method.The parameter range and action space of each well test model are shown in Table 1.In addition,the reward weights are given for different parameters to make the agent pay more attention to those parameters that are important in reflecting the reservoir characteristics,as shown in Table 1.In the next section,we will discuss how these weights are used.For the action step length,it can be set as follows.

    Table 1 Action spaces in three environments based on well test models.

    a) The well test model parameters ψ are uniformly distributed in the interval:

    b)The sensitivity of the double log curve to the parameter values in different ranges is different.We hope to have smaller step length in more sensitive areas.Therefore,some well test model parameters ψ are exponentially distributed in the interval:

    In this work,theSis set to uniformly distributed,and other parameters are set to exponentially distributed.ndis set to 100.

    3.2.Environment

    After the agent takes an action,its state will change and the environment will generate feedback (reward).The environment is constructed based on the test well model and is designed for the agents to interact with it for curve matching.In an environment,the state and reward provided to the agent are two basic elements.For the state,we define it as in Eq.(10).

    wherew(i)is the reward weight of theith well test model parameter,and its value is shown in Table 1.The interpretation results of important or insensitive parameters can be improved by setting different reward weights.δ is the maximum error when curve matching is completed.The definition oferris as follows:

    Equation(12)indicates that when the agent takes an action,the parameter error becomes smaller and the reward will be positive.Otherwise,the reward will be negative.In addition,negative rewards need to be greater than positive rewards,which helps the agent reach its goal faster and reduce unnecessary actions(Wiewiora,2003).The quality of reward design significantly affects the performance of the agent.Therefore,in Section 5.2,we discuss the influence of different reward design methods on the results.

    3.3.Training process and implementation details

    Algorithm 1 is the training process for the proposed method to accomplish automated pressure derivative curve matching.The process starts with an initial target pressure derivative curve at the beginning of each episode.We trained the agent to perform a total of 200 episodes of curve matching.The termination condition of each episode is that the curve matching is completed or the number of timesteps exceeds 300.To allow the agent to fully explore the environment,the agent's action in the first ten episodes are randomly sampled.After completing the exploration,the agent enters the training state.

    Algorithm 1.The training process of an agent in this work

    Fig.8.The result of 100 times curve matching of the agent on the homogeneous model.

    Fig.9.The result of 100 times curve matching of the agent on the dual porosity model.

    In this work,the Adam algorithm (Kingma and Ba,2014) was used to updates the weight θ of the online Q-network by minimizing the loss function in Eq.(6) with a learning rate of 0.0001.The weight θ′of target Q-network was updated every 1000 timesteps by the delayed copy.The discount factor γ in Eq.(6) is set to 0.99.The capacity of the replay memory is 105,and the minibatch size is set to 128.The temperature parameter τ in Eq.(7)is 1 during the training and 10 during the inference.

    The Q-network used in this work is a neural network composed of three hidden layers,and the number of neurons in each layer is 500,500,300,respectively.The hidden layer activation function is ReLU,while the output layer is not activated.The input layer takes the state with a feature number of 80 as input,and the output layer gets the Q-value for each action.The output dimension is determined by the environment.In this work,the output dimension based on the homogeneous model environment is 4,and others environment is 8.The agent is implemented based on PyTorch and trained on NVIDIA 1060Ti graphic processing unit.The average training time was 26 min,and the average time for each curve matching was only 2.86 s.In addition,the raw data,due to the data noise from human factors,measuring instruments,and production conditions,which will affect the results of automatic parameter inversion.In this work,wavelet threshold denoising method(Valencia et al.,2016) is used to denoise the original pressure and derivative data.

    4.Method verification

    4.1.Result verification

    To show the reliability of DDQN,Fig.7 compares the performance of DDQN on different well test models.To avoid the influence of randomness on the results (Mania et al.,2018;Henderson et al.,2018),we perform 5 random simulations and draw the training curve with 95%confidence interval.The results show that the agent trained based on the DDQN algorithm takes the fewest timesteps to complete the curve matching and obtains the highest reward in the homogeneous model.However,for the dual porosity and radial composite model,the agent has similar performance in both.This indicates that the number of parameters to be inverted,that is,the size of the action space,has a greater impact on the performance of the agent than the well test model itself.

    Fig.10.The result of 100 times curve matching of the agent on the radial composite model.

    In order to verify the reliability of the proposed method,the trained agent based on DDQN algorithm was asked to complete 100 times of curve matching on different well testing models,and the error of the finally obtained parameters was analyzed.In the inference process,the agent only selects actions and does not update the parameters of its action value network Q.The parameter inversion results on the three well test models are shown in Figs.8-10.The results show that the predicted parameters have a pronounced correlation with the actual parameters (consideringR2),which proves that the trained agent are able to invert the curve parameters accurately.Furthermore,we can observe that the accuracy of parameter inversion on the homogeneous model is higher than the other two models,which indicates that the parameter inversion error increases correspondingly with its number.

    We performed further statistical analysis of the errors in the parameter inversions.Table 2 is the statistical indicators used and their calculation formulas.The statistic results of parameter errors are reported in Table 3.Concretely,the mean of relative error(MRE)of the parameters is 7.58%for the homogeneous model,10.66%for the radial composite model,and 12.79% for the dual porosity model.In addition,the median is substantially lower than the mean,as Table 3,which indicates that the accuracy of the parameter inversion is quite high without considering the effect of extreme values.Standard deviation of relative error reflects the effect of parameter sensitivity on the inversion results.We can observe that the standard deviation of relative error of the skin factors (S) is generally larger due to its low sensitivity.Moreover,mean absolute error(MAE)and root mean square error(RMSE)are calculated to visualize the magnitude of the error,as shown in Table 3.Finally,coefficient of determinationR2are calculated to synthetically evaluate the performance of the parameter inversion.The results show that the agents perform well in these three models,and the error of parameter inversion and the parameter inversion results balance accuracy and stability.

    Table 2 The calculation formula of evaluation index. K is the number of inferences.

    Table 3 Error statistical of parameter inversion results.

    Table 4 Performance scores for all models.

    4.2.Result comparison

    In this section,to further show the advantages of DDQN,we compare its results with other two DRL algorithms and three classicmachine learning (ML) algorithms.

    The other two DRL algorithms are Naive DQN(Mnih et al.,2013),DQN (Mnih et al.,2015).Fig.11 shows the single-step rewards obtained by the Naive DQN (Mnih et al.,2013),DQN (Mnih et al.,2015),and DDQN algorithms when performing curve-fitting training on the homogeneous,dual porosity,and radial composite well test model.Fig.11 indicates that DDQN algorithm finally converges to a stable reward value on all well test models and the reward is the highest with the smallest variance.The Na?ve DQN single-step reward has been hovering at 0,indicating that a good strategy cannot be learned.The performance of DQN lies between these two algorithms.The training results show that the DDQN algorithm is more suitable for the problem of automatic curve matching.

    Three alternative supervised ML algorithms are:artificial neural network (ANN),random forest (RF),support vector regression(SVR)with multiple output.In this work,the input of ML algorithm is set to the measured pressure derivative curve with a feature number of 40,and the output is set to the unknown parameters ofthe well test model.The sample size of training set and validation set are 900 and 100,respectively.

    Fig.11.Performance curves of different agents on different well test models.

    Fig.12.Performance curves of different policies.(a) Cumulative reward curve for each episode,and (b) single-step reward curve.

    Fig.13.Histograms of steps required to complete curve matching 100 times by different policies.(a) Boltzmann exploration policy,and (b) ε-greedy policy.

    Fig.14.Performance curves of different reward designs.(a) Cumulative reward curve for each episode,and (b) single-step reward curve.

    Table 4 shows the comparison between the inference results of DDQN and the prediction results of ML algorithm in the validation set.The indicator in Table 4 are averages of all parameters for each well test model.We can observe that,for the homogeneous model,RF has obtained the best parameter inversion results (consideringR2),and DDQN has obtained the suboptimal results.For dual porosity model and radial composite model,DDQN achieved the optimal results,followed by RF,ANN and SVR.When the unknown parameters of the well test model are few (homogeneous model),the ML algorithm generally performs well.However,when the well test model becomes complex,the parameter inversion results of DDQN will be more robust,which can be observed from the fluctuation of Mean of Relative Error in different well test models.

    Table 6 Case 2 parameter inversion result.

    5.Results and discussion

    5.1.Impact of policy

    Fig.15.Histograms of steps required to complete curve matching 100 times by different reward designs.(a) Parameter-based reward design,and (b) curve-based reward design.

    Fig.16.Visualization of the step-by-step parameter adjustment process in Example 1. A:action index.

    In this section,we compare the effects of ε-greedy policy and bBoltzmann exploration policy,two commonly used exploration and balance methods,on the performance of the agent's well test curve matching.Among them,the action probability distribution of the ε-greedy policy is shown in Eq.(13).

    Fig.12 presents the training process under different policies.The results show that the Boltzmann exploration policy will enable the agent to obtain higher cumulative rewards and single-step rewards.In addition,the agent under Boltzmann exploration policy hasfewer steps when completing curve matching,as shown in Fig.13.

    This is because Boltzmann exploration policy is more suitable for a highly certain environment such as well test curve matching tasks.Besides,the use of Boltzmann exploration policy allows the agent to be biased towards exploration in the early stages of training and towards exploitation in the later stages.Therefore,the action certainty of the agent using Boltzmann exploration policy will improve with training,and converge to a better strategy.whereA*is the optimal action,a∈A(s),|A(s)|is the number of actions,ε is the probability of taking a random action.

    Fig.17.Visualization of the step-by-step parameter adjustment process in Example 2. A:action index.

    5.2.Reward design

    In this section,we discuss the impact of two different reward design methods on the agent performance in the task of automatic curve matching.In RL,the design of reward function will significantly affect the training effect of the agent (Ng et al.,1999;Laud,2004).Therefore,it is very important to find a reward function suitable for the current environment for the agent.For curve matching problems,an intuitive reward function design method is based on the error between the target curve and the prediction curve,as in Eq.(14).However,the reward design based on curve error is not a good practice.In fact,when the agent takes the action of parameter adjustment,the reduction of the error between the target curve and the prediction curve sometimes does not mean that the parameter error is reduced.As a result,this ambiguous reward design causes the agent's training to oscillate or diverge.Correspondingly,a more accurate reward design method is to directly use the error between the target curve parameters and the prediction curve parameters.So,we adopt this method to design the reward function,as shown in Eq.(12).

    Fig.14 compares the training process of the two reward design methods.The training process is based on Algorithm 1 and the reward value is still as shown in Eq.(11).It is shown that when the agent uses the curve error-based reward design,the cumulative reward is more,as in Fig.14(a),but the single-step reward is lower,as in Fig.14(b).This indicates that the agent has taken an action that can be rewarded but does not help accomplish the goal.Moreover,Fig.15 compares the histograms of steps required to complete curve matching 100 times by these two methods.The results indicate that setting the reward based on the parameter error will complete the curve matching task faster.

    6.Field application

    To further show the practicality of the proposed automatic matching method,case studies were carried out with 3 typical wells as examples.To make the agent suitable for reservoirs and wellbore under different conditions,the data used is dimensionless.

    6.1.Example 1

    Fig.18.Visualization of the step-by-step parameter adjustment process in Example 3. A:action index.

    Example 1 is an exploration well from the Dagang Oilfield.The permeability is interpreted as 5.4 mD.The well was produced 222 min before the test at a rate of 13.8 m3/d,shut in 621 min during the test,and a good build-up pressure curve was measured.The pressure derivative curve of this well shows the characteristics of a homogeneous reservoir.After dimensionless,the agent trained on the homogeneous model is used to automatically fit the curve.After 26 steps of parameter adjustments,the agent completed the matching process.To understand the curve matching process of the agent,the parameter adjustment action is shown in Fig.16 step by step.Finally,the automatic interpretation result isS=2.64,CD=79.In curve matching process,the agent prefers to adjustCDwith obvious features first,and then adjustS.Table 5 compares the results of automatic matching and computer-assisted manual matching,and the average relative error between them is 4.9%.It is shown that the automatic fitting method has the potential to reach the level of manual interpretation.

    6.2.Example 2

    Example 2 is a well test data from the Sichuan Basin.The well is a high-pressure gas well,and the reservoir has the characteristics of dual porosity.The permeability is interpreted as 0.23 mD.Due to the extremely low permeability of the reservoir,radial flow characteristics still did not appear after 934 h of shut-in testing.Afterdimensionless,the agent trained on dual porosity model is used to automatically fit the curve.The agent completed the curve matching by adjusting the parameters 27 times,as shown in Fig.17.The result of automatic interpretation isS=0.86,CD=2032,ω=0.13,λ=6.4 × 10-7.The results show that in the case of insufficient test time,the agent can also fit the curve well.Table 6 compares the results of automatic matching and computerassisted manual matching,and the average relative error between them is 13.92%.The error in Case 2 is higher than in Case 1,but still acceptable.

    Table 7 Case 3 parameter inversion result.

    6.3.Example 3

    Example 3 is a gas well in the Tarim Basin.The well was tested for 72 h of shut-in pressure build-up,and a pressure build-up curve has been obtained.The pressure derivative curve dropped at the end of the test,showing the characteristics of a composite formation.After dimensionless,the agent trained on radio composite model is used to automatically fit the curve.The agent completed the curve matching by gradually adjusting the parameters,as shown in Fig.18.The result of automatic interpretation isS=8,CD=300,M=0.23,Ri=40 m.The interpretation results show that the skin of the well is large.It turns out that there is pollution around the well and there is serious sanding problem.4.3.Table 7 compares the results of automatic matching and computerassisted manual matching,and the average relative error between them is 4.30%.The result of automatic interpretation in this case is very close to the result of manual interpretation,which proves the practical value of the proposed method.

    7.Conclusions

    In this work,we successfully applied DRL to the task of automatically interpreting well test data.In the automatic interpretation process,the agent interacts in an environment encapsulated based on the well test model to learn how to adjust the parameters to match the well test curve.By testing the performance of the DRL algorithms on different well test models,the following key conclusions are drawn:

    (1) By making the agent adjust the curve parameters asynchronously,the dimensioning disaster was alleviated,and DDQN algorithm was successfully used in the automatic curve matching task on different well test models.Using DDQN algorithm to perform 100 curve matching tests on three well test models,the results show that the mean relative error of the parameters is 7.58% for the homogeneous model,10.66% for the radial composite model,and 12.79% for the dual porosity model.

    (2) Comparing the performance of Na?ve DQN,DQN,DDQN algorithms on the homogeneous model,radial composite model,and dual porosity model,it is shown that the agent based on the DDQN algorithm obtains the highest cumulative reward on these three well test models.In addition,compared with the supervised ML algorithm,DDQN has the least fluctuation of evaluation index on different well test models,which reflects its robustness in curve matching.

    (3) The experimental results show that the use of parameterbased reward design can achieve better training results.In addition,Boltzmann exploration policy is more suitable for agents to balance exploration and exploitation on curve matching tasks.

    (4) In the three field case tests,the agent completed the curve matching within 30 steps.By visualizing the process of the step-by-step parameter tuning,it was verified that the agent learned the correct strategy.

    Acknowledgements

    This work received funding support from National Natural Science Foundation of China (52074322),Beijing Natural Science Foundation (3204052),Science Foundation of China University of Petroleum,Beijing (No.2462018YJRC032),and National Major Project of China (2017ZX05030002-005).The authors sincerely thank the colleagues at State Key Laboratory of Petroleum Resources for their helpful support.

    Appendix 1.Dimensionless definitions

    日韩欧美 国产精品| 亚洲 欧美一区二区三区| 蜜桃久久精品国产亚洲av| 女人被狂操c到高潮| 国模一区二区三区四区视频 | 成人鲁丝片一二三区免费| 久久久久性生活片| 美女cb高潮喷水在线观看 | 一级毛片高清免费大全| 黑人欧美特级aaaaaa片| 制服人妻中文乱码| xxxwww97欧美| 午夜精品久久久久久毛片777| 男人的好看免费观看在线视频| 69av精品久久久久久| 老汉色av国产亚洲站长工具| 悠悠久久av| 久久精品国产综合久久久| 两人在一起打扑克的视频| 国产欧美日韩精品亚洲av| 日韩免费av在线播放| 亚洲av中文字字幕乱码综合| 三级毛片av免费| 午夜亚洲福利在线播放| 99re在线观看精品视频| 亚洲av五月六月丁香网| 18禁黄网站禁片午夜丰满| 90打野战视频偷拍视频| 日本免费一区二区三区高清不卡| 午夜精品一区二区三区免费看| 欧美xxxx黑人xx丫x性爽| 国产蜜桃级精品一区二区三区| 手机成人av网站| 国内精品久久久久精免费| 色综合婷婷激情| 亚洲av中文字字幕乱码综合| 亚洲av片天天在线观看| 特级一级黄色大片| 一个人看的www免费观看视频| 亚洲一区二区三区色噜噜| 女人高潮潮喷娇喘18禁视频| 少妇丰满av| 亚洲美女视频黄频| 亚洲av免费在线观看| 亚洲精品粉嫩美女一区| 久久香蕉国产精品| 欧美黄色片欧美黄色片| 亚洲精品国产精品久久久不卡| 亚洲无线在线观看| 亚洲av日韩精品久久久久久密| 少妇裸体淫交视频免费看高清| 亚洲国产欧美人成| 十八禁网站免费在线| 18禁国产床啪视频网站| 国产精品,欧美在线| 丰满人妻一区二区三区视频av | 中文字幕av在线有码专区| 麻豆一二三区av精品| 欧美乱码精品一区二区三区| 色老头精品视频在线观看| 757午夜福利合集在线观看| 熟妇人妻久久中文字幕3abv| 搞女人的毛片| 国内精品一区二区在线观看| 国产毛片a区久久久久| 小说图片视频综合网站| 亚洲一区二区三区色噜噜| 午夜视频精品福利| 亚洲乱码一区二区免费版| 国产成人系列免费观看| 99re在线观看精品视频| 在线永久观看黄色视频| 国产私拍福利视频在线观看| www.999成人在线观看| 亚洲无线在线观看| 叶爱在线成人免费视频播放| 亚洲av美国av| 麻豆国产97在线/欧美| 国产一区二区三区在线臀色熟女| 亚洲精品中文字幕一二三四区| 免费观看人在逋| 国产成人福利小说| 99在线人妻在线中文字幕| 九九热线精品视视频播放| 色播亚洲综合网| 嫁个100分男人电影在线观看| 后天国语完整版免费观看| 热99re8久久精品国产| 久久欧美精品欧美久久欧美| 国产三级在线视频| 悠悠久久av| 天天躁狠狠躁夜夜躁狠狠躁| 人人妻人人澡欧美一区二区| 美女午夜性视频免费| 88av欧美| 99在线人妻在线中文字幕| 一个人免费在线观看电影 | 一本综合久久免费| 国产精品99久久99久久久不卡| 成年版毛片免费区| 九九在线视频观看精品| 国内精品美女久久久久久| 午夜福利在线在线| 免费看a级黄色片| 少妇丰满av| 成人三级黄色视频| 黄色 视频免费看| 亚洲自拍偷在线| 一区福利在线观看| avwww免费| 中文字幕人妻丝袜一区二区| 国产一区二区在线av高清观看| 观看美女的网站| 99久久无色码亚洲精品果冻| 99精品久久久久人妻精品| www国产在线视频色| 亚洲欧美日韩高清在线视频| 亚洲欧美日韩东京热| 久久久久免费精品人妻一区二区| 不卡av一区二区三区| 99视频精品全部免费 在线 | 1024手机看黄色片| xxxwww97欧美| www.精华液| 麻豆一二三区av精品| 白带黄色成豆腐渣| 欧美三级亚洲精品| 亚洲国产精品sss在线观看| 亚洲国产看品久久| 日韩三级视频一区二区三区| 国产成人精品久久二区二区91| 午夜精品在线福利| 又爽又黄无遮挡网站| 婷婷丁香在线五月| 国产av麻豆久久久久久久| 亚洲色图av天堂| 国产精品综合久久久久久久免费| 婷婷六月久久综合丁香| 丰满的人妻完整版| av在线天堂中文字幕| 日本一二三区视频观看| 国产单亲对白刺激| av黄色大香蕉| 亚洲av片天天在线观看| 国产精品亚洲一级av第二区| 国产三级在线视频| 国产高清视频在线观看网站| 床上黄色一级片| a级毛片在线看网站| 国产成人一区二区三区免费视频网站| 法律面前人人平等表现在哪些方面| 亚洲第一欧美日韩一区二区三区| 看免费av毛片| 一二三四在线观看免费中文在| 欧美+亚洲+日韩+国产| 久久天堂一区二区三区四区| 国产午夜精品久久久久久| 91久久精品国产一区二区成人 | 观看美女的网站| 国产精华一区二区三区| aaaaa片日本免费| 男女那种视频在线观看| 欧美日韩瑟瑟在线播放| www.自偷自拍.com| 啦啦啦免费观看视频1| 欧美日韩中文字幕国产精品一区二区三区| 色播亚洲综合网| 国产69精品久久久久777片 | 久久天堂一区二区三区四区| 香蕉丝袜av| 女警被强在线播放| 国产av一区在线观看免费| 91九色精品人成在线观看| 精品一区二区三区视频在线观看免费| 欧美在线黄色| 国产精品爽爽va在线观看网站| 一本久久中文字幕| 国产v大片淫在线免费观看| 啦啦啦韩国在线观看视频| av福利片在线观看| 午夜福利18| 久久精品综合一区二区三区| 高清毛片免费观看视频网站| 色综合欧美亚洲国产小说| 99国产精品99久久久久| 全区人妻精品视频| 精华霜和精华液先用哪个| 久久久久久久午夜电影| 十八禁人妻一区二区| 欧美一级毛片孕妇| 日韩欧美在线乱码| 人妻夜夜爽99麻豆av| 一进一出好大好爽视频| 亚洲精品中文字幕一二三四区| 国产精品美女特级片免费视频播放器 | 国模一区二区三区四区视频 | 脱女人内裤的视频| 欧美三级亚洲精品| 亚洲,欧美精品.| 后天国语完整版免费观看| 国产探花在线观看一区二区| 全区人妻精品视频| 18禁国产床啪视频网站| 国产成人影院久久av| 一本精品99久久精品77| 99视频精品全部免费 在线 | 精品久久久久久久久久免费视频| 亚洲 欧美 日韩 在线 免费| 国产欧美日韩精品亚洲av| 麻豆国产av国片精品| 日韩大尺度精品在线看网址| 午夜视频精品福利| 长腿黑丝高跟| 亚洲aⅴ乱码一区二区在线播放| 两人在一起打扑克的视频| 麻豆一二三区av精品| 亚洲一区二区三区色噜噜| 99热只有精品国产| a在线观看视频网站| 老司机午夜福利在线观看视频| 成人永久免费在线观看视频| 人人妻,人人澡人人爽秒播| 国产一区二区在线av高清观看| 麻豆国产av国片精品| 亚洲精品456在线播放app | 国产一级毛片七仙女欲春2| 丰满的人妻完整版| cao死你这个sao货| 精品日产1卡2卡| 黑人操中国人逼视频| 此物有八面人人有两片| 久久久久久国产a免费观看| 一级作爱视频免费观看| 久久久精品大字幕| 成年女人看的毛片在线观看| 久久精品国产清高在天天线| 少妇裸体淫交视频免费看高清| 国产成人欧美在线观看| 99国产精品99久久久久| 大型黄色视频在线免费观看| 18禁国产床啪视频网站| 嫁个100分男人电影在线观看| 欧美极品一区二区三区四区| 国产伦在线观看视频一区| 此物有八面人人有两片| 久久精品综合一区二区三区| av女优亚洲男人天堂 | 亚洲熟妇熟女久久| 91久久精品国产一区二区成人 | 在线观看午夜福利视频| 午夜成年电影在线免费观看| 又紧又爽又黄一区二区| 久久精品aⅴ一区二区三区四区| 日本 欧美在线| 欧美丝袜亚洲另类 | 老汉色∧v一级毛片| 一进一出好大好爽视频| 久久性视频一级片| 欧美一级毛片孕妇| 在线十欧美十亚洲十日本专区| 五月伊人婷婷丁香| 欧美在线一区亚洲| 人妻久久中文字幕网| 亚洲色图av天堂| 在线观看一区二区三区| 久久亚洲真实| 欧美国产日韩亚洲一区| 女同久久另类99精品国产91| 久久中文看片网| 欧美成狂野欧美在线观看| 麻豆一二三区av精品| 国产真人三级小视频在线观看| 天天添夜夜摸| 国产亚洲欧美98| 亚洲va日本ⅴa欧美va伊人久久| 国产99白浆流出| 国产高清视频在线观看网站| 久久性视频一级片| 欧美一级a爱片免费观看看| 999久久久精品免费观看国产| 国产精品,欧美在线| 淫秽高清视频在线观看| 午夜激情福利司机影院| 国产伦精品一区二区三区四那| 国产高潮美女av| 亚洲熟妇熟女久久| 三级毛片av免费| 久久久久免费精品人妻一区二区| 久久性视频一级片| 久久久国产成人免费| 真人做人爱边吃奶动态| 欧美又色又爽又黄视频| 日本精品一区二区三区蜜桃| 亚洲色图 男人天堂 中文字幕| 一级毛片女人18水好多| 久久久久久九九精品二区国产| 99在线人妻在线中文字幕| 精品久久久久久久末码| 丰满人妻熟妇乱又伦精品不卡| 看黄色毛片网站| 人妻夜夜爽99麻豆av| 精品国产三级普通话版| 亚洲,欧美精品.| 美女cb高潮喷水在线观看 | 99re在线观看精品视频| 99国产精品一区二区蜜桃av| 日韩欧美在线二视频| 国产精品亚洲一级av第二区| 国产欧美日韩一区二区三| 欧美xxxx黑人xx丫x性爽| 两个人看的免费小视频| 久久精品人妻少妇| 91在线精品国自产拍蜜月 | 校园春色视频在线观看| 99视频精品全部免费 在线 | 窝窝影院91人妻| 动漫黄色视频在线观看| 色吧在线观看| 日韩有码中文字幕| 香蕉久久夜色| 热99在线观看视频| 高清毛片免费观看视频网站| 国产伦精品一区二区三区四那| 老熟妇仑乱视频hdxx| 色综合站精品国产| 老司机午夜福利在线观看视频| 麻豆成人av在线观看| 精品乱码久久久久久99久播| bbb黄色大片| 国产亚洲精品综合一区在线观看| av国产免费在线观看| 国产精华一区二区三区| 最新美女视频免费是黄的| 成人av一区二区三区在线看| 国产精品自产拍在线观看55亚洲| 又爽又黄无遮挡网站| 亚洲精品一卡2卡三卡4卡5卡| 99在线人妻在线中文字幕| 国产视频内射| 特级一级黄色大片| 黄色女人牲交| 成人av在线播放网站| 国产午夜精品论理片| 美女黄网站色视频| 国产真实乱freesex| 曰老女人黄片| 波多野结衣高清作品| 色视频www国产| 少妇熟女aⅴ在线视频| 久久婷婷人人爽人人干人人爱| 给我免费播放毛片高清在线观看| 日韩欧美在线乱码| 日韩 欧美 亚洲 中文字幕| 欧美成狂野欧美在线观看| 精品国产乱子伦一区二区三区| 久久欧美精品欧美久久欧美| 日日干狠狠操夜夜爽| 国产又黄又爽又无遮挡在线| 午夜福利高清视频| 免费观看人在逋| 国产高清videossex| 午夜福利高清视频| 亚洲欧美日韩卡通动漫| 午夜精品一区二区三区免费看| 久久久久久九九精品二区国产| 国产高清有码在线观看视频| 精品一区二区三区av网在线观看| 亚洲真实伦在线观看| 午夜福利在线观看吧| 国产成人精品久久二区二区免费| 男女下面进入的视频免费午夜| www.www免费av| 啦啦啦观看免费观看视频高清| 日本与韩国留学比较| 欧美成人一区二区免费高清观看 | 亚洲av美国av| 欧美中文日本在线观看视频| 一区二区三区激情视频| 欧美日韩乱码在线| 99久久成人亚洲精品观看| 午夜福利高清视频| 亚洲国产精品成人综合色| 脱女人内裤的视频| 啦啦啦免费观看视频1| 真人一进一出gif抽搐免费| 真人做人爱边吃奶动态| avwww免费| 美女 人体艺术 gogo| 网址你懂的国产日韩在线| 亚洲精品国产精品久久久不卡| 成人三级做爰电影| 成人国产综合亚洲| 国产欧美日韩精品亚洲av| 亚洲欧美精品综合久久99| 悠悠久久av| 日韩av在线大香蕉| 免费一级毛片在线播放高清视频| 精品国产三级普通话版| 国产美女午夜福利| 国产精品亚洲一级av第二区| av在线天堂中文字幕| 国产激情欧美一区二区| 色综合亚洲欧美另类图片| 1024手机看黄色片| 成年版毛片免费区| 波多野结衣高清作品| 首页视频小说图片口味搜索| 全区人妻精品视频| 一个人免费在线观看电影 | 丰满人妻一区二区三区视频av | 亚洲精华国产精华精| 丁香六月欧美| 99久久精品热视频| 亚洲欧美激情综合另类| 国产精品久久久久久久电影 | 一个人观看的视频www高清免费观看 | 两性夫妻黄色片| 少妇熟女aⅴ在线视频| 哪里可以看免费的av片| 国产精品亚洲av一区麻豆| 亚洲精品一卡2卡三卡4卡5卡| 嫩草影院精品99| 亚洲国产精品久久男人天堂| 亚洲 国产 在线| 欧美xxxx黑人xx丫x性爽| 国产免费男女视频| 午夜福利高清视频| 日韩高清综合在线| 免费在线观看日本一区| 午夜a级毛片| 亚洲成a人片在线一区二区| 精品国产超薄肉色丝袜足j| 我要搜黄色片| 日本一本二区三区精品| 亚洲欧美日韩无卡精品| 女警被强在线播放| 国产极品精品免费视频能看的| 国产成人精品久久二区二区91| 小蜜桃在线观看免费完整版高清| 岛国视频午夜一区免费看| 欧美中文日本在线观看视频| 久久久久久九九精品二区国产| 久久天躁狠狠躁夜夜2o2o| 亚洲自偷自拍图片 自拍| 90打野战视频偷拍视频| 国产精品 欧美亚洲| 最好的美女福利视频网| www.自偷自拍.com| 国内揄拍国产精品人妻在线| 免费观看的影片在线观看| 国产成+人综合+亚洲专区| 久久久水蜜桃国产精品网| 国产精品久久电影中文字幕| 久久人人精品亚洲av| 操出白浆在线播放| 国产精品98久久久久久宅男小说| 叶爱在线成人免费视频播放| 日日摸夜夜添夜夜添小说| 久久精品国产清高在天天线| 欧美另类亚洲清纯唯美| 人人妻,人人澡人人爽秒播| 亚洲人成电影免费在线| 亚洲最大成人中文| av在线天堂中文字幕| 精品久久久久久久久久久久久| 亚洲欧美日韩无卡精品| 午夜激情欧美在线| 不卡一级毛片| 丝袜人妻中文字幕| 一个人看视频在线观看www免费 | 午夜免费观看网址| 校园春色视频在线观看| 天天躁狠狠躁夜夜躁狠狠躁| 日韩有码中文字幕| 精品福利观看| 制服丝袜大香蕉在线| 亚洲成av人片在线播放无| 变态另类丝袜制服| 最近最新中文字幕大全电影3| 久久亚洲精品不卡| 欧美3d第一页| 97碰自拍视频| 1024手机看黄色片| 成年女人看的毛片在线观看| 这个男人来自地球电影免费观看| 黑人操中国人逼视频| 久久精品综合一区二区三区| 精品一区二区三区av网在线观看| 色综合亚洲欧美另类图片| 老司机午夜福利在线观看视频| 日本五十路高清| 非洲黑人性xxxx精品又粗又长| 每晚都被弄得嗷嗷叫到高潮| 91在线精品国自产拍蜜月 | 又黄又爽又免费观看的视频| 免费看美女性在线毛片视频| 在线观看舔阴道视频| 日本精品一区二区三区蜜桃| 制服人妻中文乱码| 久久婷婷人人爽人人干人人爱| 精品久久久久久成人av| 91字幕亚洲| 99精品久久久久人妻精品| 999久久久精品免费观看国产| 又大又爽又粗| 精品福利观看| 亚洲aⅴ乱码一区二区在线播放| 91九色精品人成在线观看| 亚洲无线在线观看| 国产激情欧美一区二区| 国产精品九九99| 亚洲国产精品久久男人天堂| 色噜噜av男人的天堂激情| 久久草成人影院| 人人妻人人澡欧美一区二区| 亚洲成人久久爱视频| 久久久久久人人人人人| 国产亚洲精品久久久com| 国产亚洲av高清不卡| 日本 欧美在线| 黑人操中国人逼视频| 久久久久久国产a免费观看| 窝窝影院91人妻| 女生性感内裤真人,穿戴方法视频| 男女那种视频在线观看| 97超视频在线观看视频| 99热这里只有是精品50| 精品国产三级普通话版| 欧美中文综合在线视频| 亚洲av成人不卡在线观看播放网| 成人性生交大片免费视频hd| 国产精品综合久久久久久久免费| 成人av在线播放网站| 一区二区三区高清视频在线| 麻豆成人午夜福利视频| 12—13女人毛片做爰片一| 99精品久久久久人妻精品| 不卡一级毛片| 男女视频在线观看网站免费| 亚洲精品一卡2卡三卡4卡5卡| 亚洲精品久久国产高清桃花| 亚洲成人免费电影在线观看| 日本一二三区视频观看| 九九在线视频观看精品| 国产亚洲av嫩草精品影院| 两性午夜刺激爽爽歪歪视频在线观看| 叶爱在线成人免费视频播放| 一个人看视频在线观看www免费 | 俄罗斯特黄特色一大片| 国产成+人综合+亚洲专区| 观看美女的网站| 国内精品一区二区在线观看| 亚洲国产精品成人综合色| 亚洲欧美精品综合一区二区三区| 一夜夜www| 成人高潮视频无遮挡免费网站| 亚洲黑人精品在线| 变态另类成人亚洲欧美熟女| 丝袜人妻中文字幕| 国产午夜福利久久久久久| 午夜精品久久久久久毛片777| 午夜福利高清视频| 熟女人妻精品中文字幕| 国产高清视频在线播放一区| 国产成人一区二区三区免费视频网站| 国产精品久久久久久亚洲av鲁大| 99精品久久久久人妻精品| www.自偷自拍.com| 亚洲人与动物交配视频| 日韩欧美国产一区二区入口| 精品一区二区三区四区五区乱码| 伊人久久大香线蕉亚洲五| 色视频www国产| 美女午夜性视频免费| 国产激情久久老熟女| 97人妻精品一区二区三区麻豆| 欧美日韩亚洲国产一区二区在线观看| av欧美777| 日本黄大片高清| 一进一出抽搐动态| av天堂在线播放| 久久久久国产一级毛片高清牌| 日本 av在线| 午夜久久久久精精品| 美女高潮喷水抽搐中文字幕| 精品国内亚洲2022精品成人| 天天添夜夜摸| 国内精品一区二区在线观看| a在线观看视频网站| 精品一区二区三区视频在线观看免费| av女优亚洲男人天堂 | 欧美日韩福利视频一区二区| 97碰自拍视频| 色综合欧美亚洲国产小说| 久久精品国产清高在天天线| 国内久久婷婷六月综合欲色啪| 免费电影在线观看免费观看| 日日夜夜操网爽| 国内毛片毛片毛片毛片毛片| 亚洲自偷自拍图片 自拍| 亚洲黑人精品在线| 一区二区三区激情视频| bbb黄色大片| www日本黄色视频网| 久久久久精品国产欧美久久久| 亚洲av成人av| 欧美日韩国产亚洲二区| 最近最新免费中文字幕在线| 国产精品1区2区在线观看.| 给我免费播放毛片高清在线观看| 欧美+亚洲+日韩+国产| 亚洲精品456在线播放app | 听说在线观看完整版免费高清| 国产1区2区3区精品| 叶爱在线成人免费视频播放|