• <tr id="yyy80"></tr>
  • <sup id="yyy80"></sup>
  • <tfoot id="yyy80"><noscript id="yyy80"></noscript></tfoot>
  • 99热精品在线国产_美女午夜性视频免费_国产精品国产高清国产av_av欧美777_自拍偷自拍亚洲精品老妇_亚洲熟女精品中文字幕_www日本黄色视频网_国产精品野战在线观看 ?

    Parallel Extraction of Marine Targets Applying OIDA Architecture

    2022-06-14 05:21:00LIULinLIWanwuZHANGJixianSUNYiandCUIYumeng
    Journal of Ocean University of China 2022年3期

    LIU Lin, LI Wanwu, *, ZHANG Jixian, SUN Yi, and CUI Yumeng

    Parallel Extraction of Marine Targets Applying OIDA Architecture

    LIU Lin1), LI Wanwu1), *, ZHANG Jixian2), SUN Yi1), and CUI Yumeng1)

    1),,266590,2),100830,

    Computing resources are one of the key factors restricting the extraction of marine targets by using deep learning. In order to increase computing speed and shorten the computing time, parallel distributed architecture is adopted to extract marine targets. The advantages of two distributed architectures, Parameter Server and Ring-allreduce architecture, are combined to design a parallel distributed architecture suitable for deep learning–Optimal Interleaved Distributed Architecture (OIDA). Three marine target extraction methods including OTD_StErf, OTD_Loglogistic and OTD_Sgmloglog are used to test OIDA, and a total of 18 experiments in 3 categories are carried out. The results show that OIDA architecture can meet the timeliness requirements of marine target extraction. The average speed of target parallel extraction with single-machine 8-core CPU is 5.75 times faster than that of single-machine single-core CPU, and the average speed with 5-machine 40-core CPU is 20.75 times faster.

    parallel computing; distributed architecture; deep learning; target extraction; PolSAR image

    1 Introduction

    The effect of marine target detection based on deep learning (DL) largely depends on the amount of training data and training scale. Therefore, with the help of parallel distributed architecture for model training, the training time of deep learning is reduced, which is beneficial to optimize the neural network model rapidly (Goyal, 2017), thus improving the speed and accuracy of marine target detection.

    Parallel architecture has been applied to related researches of target detection and extraction, and has achieved certain results. Parallel architecture and parallel tempering algorithm were used to conduct multi-target tracking, and the results show that it can make full use of parallel com- puting capabilities of GPU and improve the computational efficiency of the algorithm (Wen, 2017). In order to solve the timeliness problem of video processing, parallel architectures were designed to detect moving targets in the vi- deos, which significantly improved the computing speed (Peng., 2014; Lou., 2016; Zeng, 2017). Ling. (2016) proposed a parallel algorithm using sparse pulse- coupled neural network to detect moving targets based on Nvidia compute unified device architecture (CUDA), which improved the computational efficiency of the algorithm. Zhang (2017) changed traditional image processing algorithm based on CPU serial, and adopted GPU parallel algorithm to realize ship target detection in optical images. You (2016) derived fast detection algorithm for abnormal targets, proposed a parallel processing method under GPU architecture, and accelerated the calculation of hyperspectral data through CUDA, up to 33.2 times (Li, 2015). Different distributed architectures have been proposed to improve image processing speed (Hu., 2016; Quirita., 2016; Wu., 2016; Chen., 2017; Ye., 2017; Huang., 2018). And the parallel algorithm was enhanced from different angles to improve its applicabi- lity in deep learning (Aytekin, 2016; Mamidala., 2018; Cheng and Xu, 2019; Shen., 2019; Liu., 2020). Alqahtani and Demirbas (2019) established three parallel models of different system architectures including Parameter Server (PS), Peer-to-Peer, and Ring Allreduce (RA). Thao Nguyen. (2019) proposed two hierarchical distributed memory multi-lead AllReduce algorithms for GPU accelerated clusters optimization, and evaluated it on the discrete-event simulation simgrid. Bouzidi (2019) used MapReduce model to realize the parallel distributed algorithm based on the computing engine SPARK.

    On the basis of existing research results, the PS archi-tecture and the RA architecture are synthesized and a moreefficient parallel distributed architecture is designed according to the characteristics of the polarized Synthetic Aperture Radar (SAR) image and its deep learning data set.A single computer uses multi-core CPU/GPU to achieve pa- rallel computing, while multiple computers pre-load model training parameters through the parameter server to achieve uninterrupted distributed computing. In this way, parallel experiments are conducted on multiple extraction models of marine targets to test the performance of Optimal Interleaved Distributed Architecture (OIDA).

    2 Proposed Architecture

    2.1 Parallel Distributed Architecture

    Parallel distributed architectures mainly include two types: PS architecture and RA architecture.

    In PS architecture, the nodes of cluster are divided into two categories: PS and Worker. PS stores the parameters of the model, and Worker calculates the gradient of the parameters. In each iterative process, Worker obtains the parameters from PS, and then returns the calculated gradient to PS. PS aggregates the gradients returned from Worker, then updates the parameters and passes the new parameters to Worker. PS architecture is the most common-ly used distributed training architecture for DL. PS architecture using synchronous SGD is shown in Fig.1.

    The devices of ring-allreduce architecture are all Workers and form a ring, as shown in Fig.2, and there is no central node to aggregate the gradients calculated by all Workers. In the iterative process, each Worker calculates the gradient after completing its own mini-batch training, and passes it to the next Worker in the ring. Worker also re- ceives the gradient passed by the previous Worker at the same time. Each Worker updates the model parameters after receiving the gradients of all Workers except for itself. Compared with PS architecture, ring-allreduce architecture is bandwidth-optimized, which can make full use of the cha-racteristics of BP algorithm to reduce training time.

    2.2 OIDA Architecture

    The PS and RA architectures are integrated in the paper, and on this basis OIDA is designed, as shown in Fig.3. Its working principle is as follows: A single computer makes full use of multi-core CPU and GPU to achieve parallel computing, while multiple computers pre-load intermediate training parameters of model through the PS to achieve uninterrupted distributed computing. In an iterative pro- cess, each Worker of the child node calculates the gradientafter completing its own batch training, passes it to PServer, and reads the model from the model folder to continue training. The PServer of the child node passes the gradient and other parameters calculated by Worker at this node to the central node chief, receives the model parameters of the central node at the same time, updates the model in themodel folder of the node for Worker to call, and passes the model parameters to the upper and lower nodes connected to it. The upper and lower nodes check the model para- meters from the child nodes and compare with the model parameters of their own nodes to determine whether to up- date the models. PServer in the central node chief monitors and receives the model parameters passed by each child Node, passes them to Worker_DS in central node to test and evaluate model parameters, and updates the model parameters in the model folder for the child nodes to call.

    For OIDA deployment, first a cluster is created then job and task are assigned. For each task, host addresses are assigned, and a server is created. Cluster must be imported when sever is created so that each server knows which hosts are included in the cluster it is in, and then servers can communicate each other. Servers must be created on their own host. Once all servers are created on their respective hosts, the entire cluster is constructed, and servers be- tween clusters can communicate with each other. Each Ser- ver contains two components: Master and Worker. Master can mainly provide remote access (RPC protocol) to each device in the cluster. At the same time, it serves as a target for creating TF. Session. Worker can execute calculation sub- graphs with local equipments.

    Fig.1 Synchronous SGD training method in PS architecture.

    Fig.2 Ring-allreduce architecture.

    Fig.3 OIDA architecture.

    3 Models and Methods

    3.1 DL Model Construction

    A DL model OceanTDA9 is constructed to detect marine targets in this paper. OceanTDA9 contains 4 convolutional layers, 1 convolutional group and 3 fully connected layers, as shown in Fig.4. The first 4 convolution forms are same, and each group is as Convolution2D-relu-Dropout-Max- pool. The organization form of the intermediate convolution group is as (Convolution2D-relu-Dropout)*2-Max- pool. The last three are fully connected dense layers, among which the first two sets of dense are as Dense-relu-Drop- out, and the last fully connected dense layer only has dense. The kernel of each convolutional layer is a 3×3 small convolution kernel, and Dropout is set to 0.2. All pooling adopts maxpooling with a core size of 2×2 and a sliding step stride of 2.

    3.2 Target Extraction Method

    The marine target extraction methods are proposed in this paper, including Constant False Alarm Rate (CFAR) method based on the initial detection (OTD_StErf), marine target extraction method based on loglogistic (OTD_ Loglogistic), the method based on Adjoint Covariance Cor- rection Model (OTD_Sgmloglog) for complex sea conditions.

    Fig.4 DL model OceanTDA9 for marine target detection.

    The OTD_StErf method is to first adopt DL model OceanTDA9 for initial detection, and save the obtained suspected targets. 3×3 images, which contains 28×28 pixel,are constructed with the suspected target points as the cen- ter in turn. Then, the two-parameter CFAR method was used to perform n tests on the 3×3 window. The value ofis calculated by Eq. (1), and the critical pixel valueXis found by the target detection model (Eq. (2)). The pixels whose values are larger than the critical value in the 3×3 image window are determined as suspected targets.

    OTD_Loglogistic method first adopts the DL model OceanTDA9 for initial detection, and constructs a 3×3 image window with the obtained suspected target as the center, and saves the pixels that are not the targets in the window to the array. Calculate the frequency of each pixel inand draw a histogram, extract the effective pi-xels and corresponding frequencies from the smallest pixel to the largest pixel in the histogram and save them to the array2 and2 as fitting data. Call loglogistic probability density function(,,) (Eq.(3)) to fit2 and2, obtain the parameter values of (,)for the fitted curve, and evaluate the goodness of fit such as Chi-square and absolute value error.

    The principle of the OTD_Sgmloglog ocean target extraction method is similar to the above two methods, except that the calling model is different. It calls Adjoint Covariance Correction Model(,,) constructed by the author to fit2,2, as shown in Eq. (4), where?[0, 255],>0,>0, the modified parameter?[0, 1], andis the standard deviation of the fitted curve. Finally, the para- meter values of,andof the fitted curve are calculated, and the parameters indicating the goodness of fit such as Chi-square and absolute value error are calculated.

    4 Parallel Experiment

    In the paper, the experimental area for marine target parallel extraction is in the Bohai Sea, located in 37?07?– 40?56?N, 117?33?–122?08?E, and the polarized SAR data from Sentinel-1 in this area are used as the experimental data. Through experimental comparison, it is found that the VV (vertical transmission, vertical reception) polarizedSAR image from Sentinel-1 IW (Interferometric Wide Swath Mode) is more suitable for marine target extraction. So VV polarized SAR image is selected for experiments. After the VV polarized SAR image is preprocessed by the pro- cedure such as split, calibration, speckle filter, multi-look- ing, terrain correction, resampling and data conversion, the marine target detection data set is obtained, which is used as the data source of parallel distributed experiments for marine target extraction.

    The marine target detection DL model OceanTDA9 is adopted to learn and train the preprocessed PolSAR dataset according to the above parallel distributed architectures, and the parameters of the neural network model for marine target detection are obtained. The model parameters are called to detect the research area, and the obtained sus- pected targets are saved in the library files. The suspected targets detected in the area with the range of 39?52?– 40?13?N, 120?44?–121?03?E are marked by the magenta boxes in Fig.5, where the upper left corner is assigned as 0 rows and 0 columns, corresponding to 40?13?N and 120?44?E, the lower right corner is at 2715 rows and 2519 columns, corresponding to 39?52?N and 121?03?E. Similar situations in this paper will not be explained again. The OIDA architecture designed in this paper is adopted to do experiments on three methods, OTD_StErf, OTD_Log- logistic, and OTD_Sgmloglog, and extract distributed tar- gets in the research area.

    Fig.5 Suspected targets after initial detection (magenta box).

    4.1 Experiment Design

    The overall process of marine target extraction by using parallel distributed architecture is shown in Fig.6. Each computer in the cluster is configured with a 4-core 8- thread CPU with 32GB of memory, and two GPUs with 2GB of video memory. All computers form a Gigabit Eth- ernet. Each computer in the cluster is configured with network parameter models and PolSAR image data after DL to ensure that the necessary dynamic data with extre- mely low redundancy are transmitted in the network.

    After the software and hardware environment is set up, start the service, and designate a node as chief, which is responsible for managing each node, coordinating the ope- ration between each node, and completing the distribution of tasks, the collection of intermediate results and the inte- gration and visualization of results. After the task of chief is received by other Workers, the corresponding function is initialized according to the task requirements. The tar- get fitting parameter extraction function is used to com- plete the clustering of suspected targets, extract the position parameters such as the center coordinates and inclina-tion of the suspected targets, and extract the shape parameters such as length, width, and area, then use a straight line to fit the central axis and use an ellipse to fit the shape of the suspected target. Three target parameter extraction me-thods including OTD_StErf method, OTD_Loglogistic me- thod, and OTD_Sgmloglog method were used.

    After each Worker node in the cluster complete initial detection task assigned by chief, the suspected targets in the unit of DL sub-image (28 pixels×28 pixels) are saved to the corresponding folder, and chief is notified that the tasks have been completed. According to the load condition of Workers in the cluster, the collected suspected targets are grouped and distributed to the corresponding Workers by chief. After the task of chief is received by Worker nodes, the target parameters are called to extract the function. According to the pre-set maximum and minimum length and aspect ratio, the target parameters that meet the conditions

    are extracted and saved to the corresponding folder, and Chief is notified that the tasks have been completed. The target parameters extracted by each Worker in the cluster are collected by chief and dynamically visualized after integration.

    Fig.6 Marine target extraction procedure by using parallel distributed architecture.

    4.2 Experiment on OTD_StErf

    For comparison, single-machine single-core device is used to extract marine target parameters for 35 suspected targets detected initially by OTD_StErf. It takes 381.45s to extract 22 targets. The time sequence is shown in Fig.7, and each target extraction task takes approximately 6s. In order to ensure the display effect and data storage, each target is set to delay 5s after displayed and then turned off.

    The OIDA designed in this paper is used for marine target extraction experiment based on OTD_StErf. The times for every CPU in every node of the experiment to display the suspected targets and extract targets are shown in Fig.8,where Fig.8(a) is the parallel computing with single-ma- chine 8-core CPU. The 35 suspected targets are allocated randomly to 0–7-core CPU, 4–5 tasks for one core. After each CPU receives the tasks, it will display positions of suspected targets in the research marine area in turn, and extract targets according to the target size and aspect ratio specified by the user.From Fig.8(a), it can be found that the suspected target detection takes approximately 6–7s, and each target extraction takes approximately 5–7s.

    Fig.8(b) is the parallel computing with dual-machine 16-core CPU, and 1–3 tasks are randomly allocated to 0–15 core CPU. The times for suspected target detection and each target extraction are approximately 6–7s. The first to complete the task is the 5-core and 6-core CPU, which are allocated 2 tasks and take 10s. The last one to complete the task is 1-core CPU, which takes approximately 33s. The CPU is assigned a total of 3 tasks, that is, 3 suspected tar- gets need to be detected, of which 2 suspected targets meet the conditions. The extraction times of these 2 targets are 10s and 25s, respectively.

    Fig.8(c) is the parallel computing with 3-machine 24- core CPU, and each CPU is randomly assigned 1–2 tasks. The first CPU that completes tasks takes 3s, and the last takes approximately 27s. Fig.8(d) is the parallel computing with 4-machine 32-core CPU, 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 23s. Fig.8(e) is the parallel computing with 5-machine 40-core CPU. Since there are 35 tasks, 35 CPU cores are involved in the calculation, and 1 task is randomly allocated to one CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 9s. The 0–7 core CPUs in the calculationshown in Fig.8(b–e) are the chief CPUs. Since chief is responsible for assigning tasks to other Workers, the time for the chief CPU to display the first suspected target is when the program runs for 3s, and the time for other Workers to display the first suspected target is when the program runs for 1s, 0 or 3s, 1 or 2s, 1 or 2s.

    The CPU task execution status of each node is shown in Table 1. In single-machine 8-core CPU parallel architecture, each CPU takes approximately 53.77s, all CPUs take a total of 430.23s, and the total time to complete the task is 67.07s, which is 17.58% of the time consumed by a single machine with one core,., 82.49% faster. In 5- machine 40-core CPU parallel architecture, the cluster com- posed of CPUs take 17.66s to complete the task, which is 4.62% of the time consumed by a single machine, and 95.38% faster. The CPU execution status in other clusters are shown in Table 1.

    Fig.7 The display time of target parameter extraction when OTD_StErf method is carried out under single-machine single- core condition. The suspected targets (Starget) are represented by the red dot ‘?’, and the extracted targets (Etarget) are represented by the blue plus ‘+’. Here in after the same.

    Fig.8 Parallel extraction CPU-t diagram for OTD_StErf.

    Table 1 Parallel extraction CPU status for OTD_StErf

    4.3 Experiment on OTD_Loglogistic

    Single-machine single-core architecture is adopted to extract marine target parameters for 35 suspected targets, detected initially after DL, by OTD_Loglogistic, which takes 389.02s and extracts 22 targets. As shown in Fig.9, each extraction takes approximately 6–9s, and the first sus- pected target is displayed when the program runs for 1s.

    The OIDA framework designed in this paper is used to extract marine target parameters by OTD_Loglogistic me- thod. The times that the CPU of each core at each node dis- plays the suspected targets and extracts them are shown in Fig.10. Among them, Fig.10(a) is the parallel computing with single-machine 8-core CPU. The 35 suspected targets are randomly allocated to 0–7 core CPU, 4–5 tasks for one core. The first CPU that completes tasks takes 22 s, and the last takes approximately 60s. Fig.10(b) is the parallel computing with dual-machine 16-core CPU. 1–3 tasks are randomly allocated to 0–15 core CPU. The first CPU that completes tasks takes 3s, and the last takes approximately 33s. Fig.10(c) is the parallel computing with 3-machine 24-core CPU. 1–2 tasks are randomly allocat- ed to each CPU. The first CPU that completes tasks takes 7s, and the last takes approximately 25s. Fig.10(d) is the parallel computing with 4-machine 32-core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 2s, and the last takes approximately 24s. Fig.10(e) is the parallel computing with 5-machine 40-core CPU. 1 task is randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 11s. In Figs.10b–e, the first suspect- ed target is displayed by the chief CPU when the program runs for 3s or 2s, and the first suspected target is display- ed by the other working nodes when the program runs for 2s, 1s, 2s, 1s or 2s, respectively.

    In the parallel extraction process of marine targets by single-machine multi-core/multi-machine multi-core OTD_ Loglogistic method, the CPU task execution status at each node is shown in Table 2. When the single-machine 8- core CPU is used, each CPU takes approximately 54.89s, all CPUs take a total of 439.18s, and the total time to complete the task is 68.40s, which is 17.58% of the time consumed by a single machine with single core, and 82.42% faster. The cluster composed of 5-machine 40-core CPU takes 18.91s to complete the task, which is 4.86% of the time consumed by a single machine with single core, and 95.14% faster. The CPU execution status in other clusters are shown in Table 2.

    4.4 Experiment on OTD_Sgmloglog

    An OTD_Sgmloglog method for marine target extraction in complex sea conditions is proposed by the author. The designed OIDA architecture is tested on this propos- ed method and the results are compared with the single- machine single-core mode.

    Single-machine single-core is used to extract marine target parameters by OTD_Sgmloglog method for 35 sus- pected targets detected initially after DL. It takes 389.52s to extract 22 targets. As shown in Fig.11, each target extraction takes approximately 6–7s, and the first suspected target is displayed when the program runs for 1s.

    Fig.9 Display time of target parameter extraction for single-machine single-core OTD_Loglogistic method.

    Fig.10 Parallel extraction CPU-t diagram for OTD_ Loglogistic.

    Table 2 Parallel extraction CPU status for OTD_Loglogistic method

    The designed OIDA architecture is used to extract marine target parameters by OTD_ Sgmloglog method. The times for each CPU at each node to display the suspected targets and extract targets are shown in Fig.12. Fig.12(a) is the parallel computing with single-machine 8-core CPU. The 35 suspected targets are randomly allocated to 0–7 core CPU, 4–5 tasks for one core. The first CPU that com- pletes tasks takes 23s, and the last takes approximately 60s. Fig.12(b) is the parallel computing with dual-machine 16-core CPU. 1–3 tasks are randomly allocated to 0–15 core CPU. The first CPU that completes tasks takes 3s, and the last takes approximately 33s. Fig.12(c) is the paral- lel computing with 3-machine 24-core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that com- pletes tasks takes 3s, and the last takes approximately 25s. Fig.12(d) is the parallel computing with 4-machine 32- core CPU. 1–2 tasks are randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 24s. Fig.12(e) is the parallel computing with 5-machine 40-core CPU. 1 task is randomly allocated to each CPU. The first CPU that completes tasks takes 1s, and the last takes approximately 12s. The 0–7 core CPU in Figs.12b–e is the chief CPU and the first suspected target is displayed when the program runs for 3 or 2s. The first suspected target for other working nodes is displayed when the program runs for 2s, 1s or 2 s, 1s or 2s, 1s or 2s, respectively.

    In the parallel extraction of marine target parameters by single-machine multi-core/multi-machine multi-core OTD_ Sgmloglog method, the CPU status of each node is shown in Table 3. Each CPU of the single-machine 8-core takes approximately 54.23s, all CPUs take a total of 433.85s, and the total time to complete the task is 67.40s, which is 17.30% of the time consumed by a single machine with single core, and 82.70% faster. The cluster composed of 5-machine 40-core CPU takes 20.17s to complete the task, which is 5.18% of the time consumed by a single machine with single core, and 94.82% faster. The CPU execution status in the other clusters is shown in Table 3.

    Fig.11 CPU-t diagram of parameter extraction for single-machine single-core OTD_ Sgmloglog method.

    Fig.12 Parallel extraction CPU-t diagram for OTD_ Sgmloglog method.

    Table 3 Parallel extraction CPU status for OTD_Sgmloglog method

    4.5 Results for Marine Target Extraction

    The above method is used to extract target parameters from the relevant SAR data in the research marine area, and the parameters of target position and shape are fitted by ellipses. The fitting results are shown in Fig.13, and the area marked by blue box is enlarged in Fig.14. The detailed ellipse fitting parameters of the extracted targets are listed in Table 4.

    5 Discussion and Conclusions

    The OIDA parallel distributed architecture is designed in the paper, and comparison experiments are conducted on three extraction methods of marine targets: OTD_ StErf, OTD_Loglogistic, and OTD_Sgmloglog. Each me- thod is applied under 6 deployments of single-machine single-core CPU, single-machine 8-core CPU, dual-ma-chine 16-core CPU, 3-machine 24-core CPU, 4-machine 32-core CPU, and 5-machine 40-core CPU, including 18 experiments. Experimental results show that the parallel distributed OIDA architecture which is designed in this paper is efficient in the applications of marine target extraction. The average speed of parallel target extraction for a single-machine 8-core CPU is 5.75 times the speed of a single-machine single-core CPU, where the highest is 5.84 times, and the lowest is 5.78 times. The average speed of parallel target extraction for a 5-machine 40-core CPU is 20.75 times the speed of a single-machine single-core CPU, where the highest is 21.53 times and the lowest is 19.31 times. The parallel distributed OIDA architecture shortens the time required for the operation of the marine target extraction method. The average response time per thousand square kilometers is approximately 2s, with the fastest response time of 1s, and the slowest response time of 3s. The designed parallel distributed architecture OIDA realizes the high-precision rapid extraction of target features.

    Fig.13 Marine target extraction results.

    Fig.14 Local enlargement of marine target extraction results.

    Acknowledgements

    The study was supported by the Natural Science Foundation of Shandong Province (No. ZR2019MD034), and the Education Reform Project of Shandong Province (No. M2020266). This study was supported by data from European Space Agency (ESA) and the Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences.

    Alqahtani, S., and Demirbas, M., 2019. Performance analysis and comparison of distributed machine learning systems.: 1909.02061.

    Aytekin, A., Feyzmahdavian, H. R., and Johansson, M., 2016. Analysis and implementation of an asynchronous optimization algorithm for the parameter server.: 1610.05507.

    Bouzidi, S., 2019. Parallel and distributed implementation on SPARK of a spectral-spatial classifier for hyperspectral images., 13 (3): 034501.

    Chen, H., Wei, A., and Zhang, Y., 2017. Three-level parallel-set partitioning in hierarchical trees coding based on the collaborative CPU and GPU for remote sensing images compression., 11 (4): 045015.

    Cheng, Z., and Xu, Z., 2019. Bandwidth reduction using importance weighted pruning on ring allreduce.: 1901.01544.

    Goyal, P., Dollár, P., Girshick, R., and Noordhuis, P., 2017. Accurate, large minibatch SGD: Training imagenet in 1hour.: 1706.02677.

    Hu, H., Shu, H., Hu, Z., and Xu, J., 2016. Using compute unified device architecture-enabled graphic processing unit to accelerate fast fourier transform-based regression kriging interpolation on a MODIS land surface temperature image., 10 (2): 026036.

    Huang, Y., Jin, T., Wu, Y., Cai, Z., and Cheng, J., 2018. Flexps: Flexible parallelism control in parameter server architecture., 11 (5): 566-579.

    Li, T., 2015. Parallel optimization and application research on moving object detection and recognition algorithms. Master thesis.National University of Defense Technology.

    Ling, B., Deng, Y., and Yu, S. B., 2016. Processing for accelerated sparse PCNN moving target detection algorithm with CUDA., 37 (12): 3300- 3305.

    Liu, Y. H., Zhou, J., Qi, W. H., Li, X. L., Gross, L., Shao, Q.,., 2020. ARC-Net: An efficient network for building extraction from high resolution aerial images., 8: 154997- 155010, DOI: 10.1109/ACCESS.2020.3015701.

    Lou, X. H., Guo, C. S., Song, S. L., and Qi, L. Q., 2016. Parallel implementation of video moving object detecion algorithm based on CUDA., 36 (3): 23-26.

    Mamidala, A. R., Kollias, G., Ward, C., and Artico, F., 2018. MXNET-MPI: Embedding MPI parallelism in parameter servertask model for scaling deep learning.: 18 01.03855.

    Peng, B., Zhang, C. Y., Zheng, S. B., and Tian, G., 2014. Multi- level parallel optimization of moving object detecion and fea- ture extraction algorithm., 38 (13): 173-177.

    Quirita, V., da Costa, G., Happ, P., Feitosa, P., da Silva Ferreyra R., Oliveira, D.,., 2016. A new cloud computing architecture for the classification of remote sensing data., 10 (2): 409-416.

    Shen, S., Xu, L., Liu, J., Liang, X., and Cheng, Y., 2019. Faster distributed deep net training: Computation and communication decoupled stochastic gradient descent.: 1906.12043,

    Thao Nguyen, T., Wahib, M., and Takano, R., 2019. Efficient MPI-allreduce for large-scale deep learning on GPU-clusters.: e5574.

    Wen, S., 2017. The parallel design and implementation of MCMC multi-object tracking algorithm. Master thesis. Xidian Univer- sity.

    Wu, Z., Li, Y., Plaza, A., Li, J., Xiao, F., and Wei, Z., 2016. Pa- rallel and distributed dimensionality reduction of hyperspectral data on cloud computing architectures., 9 (6): 2270-2278.

    Ye, L. H., Wang, L., Sun, Y. X., Zhao, L. P., and Wei, Y. W., 2017. Parallel multi-stage features fusion of deep convolutional neural networks for aerial scene classification., 9: 3, 294-303.

    You, W., 2016. Research on hyperspectral remote sensing target detection parallel processing. Master thesis. Harbin Engineering University.

    Zeng, T., 2017. Research on parallel algorithm for moving object contour extraction in video sequences. Master thesis. Cen- tral China Normal University.

    Zhang, F., 2017. Target recognition and parallel acceleration with GPU in marine remote sensing image. Master thesis. Changchun Institute of Optics, Fine Mechanics and Physics,Chinese Academy of Sciences.

    J. Ocean Univ. China(Oceanic and Coastal Sea Research)

    https://doi.org/10.1007/s11802-022-4839-z

    ISSN 1672-5182, 2022 21 (3): 737-747

    (November 16, 2020;

    December 24, 2020;

    January 13, 2021)

    ? Ocean University of China, Science Press and Springer-Verlag GmbH Germany 2022

    Corresponding author. E-mail: liwanwuqd@126.com

    (Edited by Chen Wenwen)

    精品少妇一区二区三区视频日本电影| 纯流量卡能插随身wifi吗| 韩国精品一区二区三区| 深夜精品福利| 狠狠狠狠99中文字幕| 国产精华一区二区三区| 性欧美人与动物交配| 黄色a级毛片大全视频| 757午夜福利合集在线观看| 少妇粗大呻吟视频| av视频免费观看在线观看| 欧美激情高清一区二区三区| 99久久精品国产亚洲精品| 窝窝影院91人妻| 亚洲欧美激情在线| 亚洲自拍偷在线| 国产成年人精品一区二区| 99国产精品免费福利视频| 亚洲国产看品久久| 成人欧美大片| 久久人人爽av亚洲精品天堂| 亚洲精品国产色婷婷电影| 亚洲情色 制服丝袜| 成人精品一区二区免费| 久久精品国产综合久久久| 又黄又粗又硬又大视频| 国产精品一区二区精品视频观看| 色哟哟哟哟哟哟| 成人亚洲精品av一区二区| x7x7x7水蜜桃| 99在线视频只有这里精品首页| 嫁个100分男人电影在线观看| 久99久视频精品免费| 黄片大片在线免费观看| 成人手机av| 成人av一区二区三区在线看| 成人av一区二区三区在线看| 神马国产精品三级电影在线观看 | АⅤ资源中文在线天堂| 亚洲精品国产精品久久久不卡| 成人亚洲精品一区在线观看| 性色av乱码一区二区三区2| 中国美女看黄片| 中出人妻视频一区二区| 50天的宝宝边吃奶边哭怎么回事| 精品久久久精品久久久| 亚洲成人免费电影在线观看| 免费在线观看影片大全网站| 精品欧美一区二区三区在线| 青草久久国产| 女警被强在线播放| 精品久久蜜臀av无| 成人18禁高潮啪啪吃奶动态图| 国内精品久久久久精免费| 久久热在线av| 日日摸夜夜添夜夜添小说| 欧美日本视频| 在线十欧美十亚洲十日本专区| 曰老女人黄片| 中文亚洲av片在线观看爽| 午夜视频精品福利| 日韩精品青青久久久久久| 久久国产乱子伦精品免费另类| 在线视频色国产色| 18禁国产床啪视频网站| or卡值多少钱| 黄频高清免费视频| 多毛熟女@视频| 色老头精品视频在线观看| 巨乳人妻的诱惑在线观看| 久久人妻熟女aⅴ| 久久久久久亚洲精品国产蜜桃av| 少妇熟女aⅴ在线视频| 精品久久久久久,| 叶爱在线成人免费视频播放| 婷婷丁香在线五月| 变态另类成人亚洲欧美熟女 | x7x7x7水蜜桃| 757午夜福利合集在线观看| 巨乳人妻的诱惑在线观看| 欧美成人性av电影在线观看| 国产av一区在线观看免费| av免费在线观看网站| 久久精品aⅴ一区二区三区四区| 亚洲一码二码三码区别大吗| 最近最新中文字幕大全电影3 | 国产99白浆流出| 亚洲男人天堂网一区| 午夜福利视频1000在线观看 | 成人亚洲精品一区在线观看| 久久久精品国产亚洲av高清涩受| 国产亚洲精品av在线| 最好的美女福利视频网| 国产精品 国内视频| √禁漫天堂资源中文www| 亚洲精品一卡2卡三卡4卡5卡| 亚洲精品久久成人aⅴ小说| 国产精品 国内视频| 伦理电影免费视频| 老司机靠b影院| 国产一区二区三区综合在线观看| 精品无人区乱码1区二区| 午夜日韩欧美国产| 国产三级在线视频| а√天堂www在线а√下载| 欧美人与性动交α欧美精品济南到| 制服丝袜大香蕉在线| 色综合亚洲欧美另类图片| 黄色视频,在线免费观看| 欧美成狂野欧美在线观看| 热99re8久久精品国产| 国产97色在线日韩免费| 99久久精品国产亚洲精品| 女人精品久久久久毛片| 精品无人区乱码1区二区| 三级毛片av免费| 老鸭窝网址在线观看| 一a级毛片在线观看| 国产欧美日韩一区二区三区在线| 国产精品亚洲美女久久久| 脱女人内裤的视频| 午夜久久久在线观看| 人妻久久中文字幕网| 大码成人一级视频| e午夜精品久久久久久久| 久久人妻熟女aⅴ| 国产亚洲精品av在线| 老汉色∧v一级毛片| 一a级毛片在线观看| 日本a在线网址| 首页视频小说图片口味搜索| 国产亚洲精品一区二区www| 999精品在线视频| 男女床上黄色一级片免费看| 欧美成人一区二区免费高清观看 | 自拍欧美九色日韩亚洲蝌蚪91| 欧美中文综合在线视频| 又黄又爽又免费观看的视频| 亚洲精品一区av在线观看| 久久香蕉国产精品| 国产欧美日韩一区二区精品| 亚洲欧美日韩高清在线视频| 亚洲欧美日韩另类电影网站| 午夜精品在线福利| 中出人妻视频一区二区| 老熟妇乱子伦视频在线观看| 国产在线观看jvid| 丝袜美腿诱惑在线| 中文字幕色久视频| 99久久99久久久精品蜜桃| 国产精品亚洲一级av第二区| 一进一出好大好爽视频| 香蕉丝袜av| 波多野结衣一区麻豆| 一卡2卡三卡四卡精品乱码亚洲| 波多野结衣巨乳人妻| 不卡av一区二区三区| 亚洲全国av大片| 亚洲第一青青草原| 满18在线观看网站| 久9热在线精品视频| 亚洲avbb在线观看| 免费在线观看视频国产中文字幕亚洲| 亚洲中文字幕一区二区三区有码在线看 | 女警被强在线播放| 色av中文字幕| 国产免费男女视频| 亚洲 国产 在线| 日韩欧美一区二区三区在线观看| 满18在线观看网站| 久久草成人影院| 欧美乱码精品一区二区三区| 亚洲国产毛片av蜜桃av| 免费女性裸体啪啪无遮挡网站| 99久久精品国产亚洲精品| 一进一出抽搐gif免费好疼| 国产伦一二天堂av在线观看| 久久久久久人人人人人| 99re在线观看精品视频| 国产精品精品国产色婷婷| 很黄的视频免费| 国产精品电影一区二区三区| 少妇 在线观看| 99re在线观看精品视频| 天天添夜夜摸| 色综合站精品国产| 欧美人与性动交α欧美精品济南到| 国内久久婷婷六月综合欲色啪| 成人国产综合亚洲| 国产在线观看jvid| 久久中文字幕人妻熟女| 欧美日韩一级在线毛片| 国产在线精品亚洲第一网站| 一级黄色大片毛片| 一a级毛片在线观看| 国内毛片毛片毛片毛片毛片| 99久久精品国产亚洲精品| 精品久久久精品久久久| 长腿黑丝高跟| 久久伊人香网站| 天天躁夜夜躁狠狠躁躁| 日韩大码丰满熟妇| 国产精品乱码一区二三区的特点 | 午夜久久久久精精品| 91字幕亚洲| 大型av网站在线播放| 国产99久久九九免费精品| 欧美av亚洲av综合av国产av| 国产精品美女特级片免费视频播放器 | 国产午夜精品久久久久久| 啦啦啦 在线观看视频| 久久午夜亚洲精品久久| 国产精品98久久久久久宅男小说| 国产蜜桃级精品一区二区三区| 免费搜索国产男女视频| 97超级碰碰碰精品色视频在线观看| 黄色a级毛片大全视频| 成人特级黄色片久久久久久久| 国产免费av片在线观看野外av| а√天堂www在线а√下载| 久久久国产成人精品二区| 亚洲国产欧美网| 每晚都被弄得嗷嗷叫到高潮| 黄色女人牲交| 亚洲国产精品成人综合色| 日本一区二区免费在线视频| www.精华液| 我的亚洲天堂| 九色亚洲精品在线播放| 在线永久观看黄色视频| 久99久视频精品免费| 亚洲精品一卡2卡三卡4卡5卡| 亚洲无线在线观看| 精品国产乱子伦一区二区三区| 两个人免费观看高清视频| 国产熟女午夜一区二区三区| 777久久人妻少妇嫩草av网站| videosex国产| 精品国内亚洲2022精品成人| 亚洲,欧美精品.| 精品久久蜜臀av无| 亚洲少妇的诱惑av| av福利片在线| 757午夜福利合集在线观看| 国产三级黄色录像| 人人澡人人妻人| 久久狼人影院| 最新在线观看一区二区三区| 伊人久久大香线蕉亚洲五| 这个男人来自地球电影免费观看| 欧美绝顶高潮抽搐喷水| videosex国产| 中文字幕av电影在线播放| 两个人免费观看高清视频| 女人精品久久久久毛片| 日韩精品青青久久久久久| 亚洲五月婷婷丁香| 国产精品国产高清国产av| videosex国产| 黄频高清免费视频| 91在线观看av| 久久久国产精品麻豆| 最新美女视频免费是黄的| 久久国产乱子伦精品免费另类| 69精品国产乱码久久久| 动漫黄色视频在线观看| 亚洲aⅴ乱码一区二区在线播放 | 欧美人与性动交α欧美精品济南到| 中文字幕最新亚洲高清| 女同久久另类99精品国产91| 丰满的人妻完整版| 国产成人影院久久av| 色综合欧美亚洲国产小说| 欧美日韩精品网址| 久久精品aⅴ一区二区三区四区| 欧美成人一区二区免费高清观看 | 午夜激情av网站| 国产精品久久久久久精品电影 | 国产国语露脸激情在线看| 色播在线永久视频| 亚洲欧美日韩无卡精品| 热99re8久久精品国产| 在线观看www视频免费| 一个人观看的视频www高清免费观看 | 人人妻人人澡欧美一区二区 | 亚洲国产毛片av蜜桃av| 操美女的视频在线观看| 91字幕亚洲| 两性夫妻黄色片| 手机成人av网站| 俄罗斯特黄特色一大片| 黄色毛片三级朝国网站| 国产午夜精品久久久久久| 最近最新中文字幕大全电影3 | 亚洲成av片中文字幕在线观看| 精品日产1卡2卡| 波多野结衣高清无吗| tocl精华| 不卡一级毛片| 午夜精品在线福利| 亚洲中文字幕一区二区三区有码在线看 | 真人做人爱边吃奶动态| 老汉色av国产亚洲站长工具| 精品国产乱子伦一区二区三区| 丰满的人妻完整版| 伦理电影免费视频| 久久天堂一区二区三区四区| 久久伊人香网站| 美女免费视频网站| 精品欧美一区二区三区在线| 波多野结衣av一区二区av| 国内精品久久久久久久电影| 99国产极品粉嫩在线观看| 久久香蕉激情| 纯流量卡能插随身wifi吗| 国产色视频综合| 国产精品亚洲一级av第二区| 亚洲人成电影观看| 一边摸一边做爽爽视频免费| 免费看十八禁软件| 欧美成人午夜精品| 男人的好看免费观看在线视频 | 香蕉久久夜色| 美女高潮到喷水免费观看| 国产在线精品亚洲第一网站| 国产精品二区激情视频| 国内精品久久久久精免费| 少妇被粗大的猛进出69影院| 三级毛片av免费| 手机成人av网站| 亚洲av成人av| 操出白浆在线播放| 一边摸一边抽搐一进一出视频| 成人免费观看视频高清| 美女扒开内裤让男人捅视频| 免费在线观看影片大全网站| 国产精品av久久久久免费| 色av中文字幕| 亚洲中文字幕一区二区三区有码在线看 | 免费看a级黄色片| av视频在线观看入口| 久久久久久亚洲精品国产蜜桃av| 99国产精品一区二区三区| 18禁国产床啪视频网站| 免费在线观看黄色视频的| 一本大道久久a久久精品| 18禁国产床啪视频网站| 桃色一区二区三区在线观看| www.熟女人妻精品国产| 久久国产亚洲av麻豆专区| 日韩欧美国产一区二区入口| 久久亚洲真实| 欧美一级a爱片免费观看看 | 别揉我奶头~嗯~啊~动态视频| 欧美乱码精品一区二区三区| 国产高清激情床上av| 精品欧美国产一区二区三| 午夜福利影视在线免费观看| 国产免费av片在线观看野外av| 热99re8久久精品国产| 午夜激情av网站| 日韩三级视频一区二区三区| 亚洲av电影不卡..在线观看| 国产午夜福利久久久久久| 亚洲九九香蕉| 夜夜躁狠狠躁天天躁| 亚洲av美国av| 一级毛片女人18水好多| 琪琪午夜伦伦电影理论片6080| 日日夜夜操网爽| 国产欧美日韩精品亚洲av| 一级毛片女人18水好多| 亚洲av日韩精品久久久久久密| 国产成人免费无遮挡视频| АⅤ资源中文在线天堂| 国产成人精品久久二区二区免费| 久久人妻福利社区极品人妻图片| 老司机深夜福利视频在线观看| 叶爱在线成人免费视频播放| 精品久久久久久久久久免费视频| av在线播放免费不卡| 欧美日韩福利视频一区二区| 成人国产综合亚洲| 51午夜福利影视在线观看| 天天一区二区日本电影三级 | 精品福利观看| 叶爱在线成人免费视频播放| 丁香六月欧美| 免费av毛片视频| 狂野欧美激情性xxxx| 亚洲全国av大片| a在线观看视频网站| 成人永久免费在线观看视频| 美女高潮喷水抽搐中文字幕| www.精华液| 亚洲欧美日韩无卡精品| 香蕉国产在线看| 欧美绝顶高潮抽搐喷水| 亚洲熟妇中文字幕五十中出| 99久久综合精品五月天人人| 国产一区二区三区视频了| 欧美在线一区亚洲| 夜夜躁狠狠躁天天躁| 亚洲av成人不卡在线观看播放网| 18美女黄网站色大片免费观看| 色播在线永久视频| 欧美老熟妇乱子伦牲交| www国产在线视频色| 老熟妇仑乱视频hdxx| av超薄肉色丝袜交足视频| 亚洲专区字幕在线| www.999成人在线观看| 亚洲国产精品久久男人天堂| 国产野战对白在线观看| 99精品久久久久人妻精品| 久久精品亚洲熟妇少妇任你| 久久久久久国产a免费观看| 欧美大码av| 国产不卡一卡二| 很黄的视频免费| 色哟哟哟哟哟哟| 一级,二级,三级黄色视频| 免费在线观看视频国产中文字幕亚洲| 在线天堂中文资源库| 天堂√8在线中文| 国产精品秋霞免费鲁丝片| 午夜激情av网站| 国内毛片毛片毛片毛片毛片| 亚洲 欧美一区二区三区| 国产精品久久久人人做人人爽| av片东京热男人的天堂| 午夜福利欧美成人| 91大片在线观看| 亚洲精品美女久久av网站| 真人做人爱边吃奶动态| 51午夜福利影视在线观看| 国产精品99久久99久久久不卡| 大香蕉久久成人网| 亚洲av成人不卡在线观看播放网| 色综合欧美亚洲国产小说| 免费av毛片视频| 久久青草综合色| 亚洲狠狠婷婷综合久久图片| 欧美不卡视频在线免费观看 | 国产精品一区二区三区四区久久 | 国产激情欧美一区二区| 亚洲视频免费观看视频| 色综合欧美亚洲国产小说| 村上凉子中文字幕在线| 久久久久精品国产欧美久久久| 天天一区二区日本电影三级 | 精品国内亚洲2022精品成人| 叶爱在线成人免费视频播放| 欧美日本视频| 电影成人av| 一区二区三区激情视频| 人人妻人人澡人人看| 欧美av亚洲av综合av国产av| 亚洲天堂国产精品一区在线| 久久久水蜜桃国产精品网| 色播在线永久视频| av网站免费在线观看视频| 亚洲电影在线观看av| 国产成人av教育| 欧美av亚洲av综合av国产av| 国产精品一区二区在线不卡| 色综合站精品国产| 色老头精品视频在线观看| 国产成人影院久久av| 国产欧美日韩一区二区三区在线| 久久草成人影院| 国产欧美日韩一区二区三| 99精品在免费线老司机午夜| 精品免费久久久久久久清纯| 午夜两性在线视频| 亚洲人成伊人成综合网2020| 一区二区日韩欧美中文字幕| 精品国产乱子伦一区二区三区| 国产精品精品国产色婷婷| 老司机在亚洲福利影院| 麻豆国产av国片精品| 最好的美女福利视频网| www.熟女人妻精品国产| 国产亚洲欧美在线一区二区| 日韩精品中文字幕看吧| 精品欧美国产一区二区三| 人人妻人人澡欧美一区二区 | 在线视频色国产色| 两人在一起打扑克的视频| 最好的美女福利视频网| 国产成人一区二区三区免费视频网站| 色哟哟哟哟哟哟| 一边摸一边抽搐一进一小说| 国产精品国产高清国产av| 午夜亚洲福利在线播放| 99久久99久久久精品蜜桃| 19禁男女啪啪无遮挡网站| 精品一区二区三区av网在线观看| 国产91精品成人一区二区三区| 多毛熟女@视频| 国产蜜桃级精品一区二区三区| 色尼玛亚洲综合影院| 日韩欧美一区二区三区在线观看| 国产精品一区二区在线不卡| 日韩欧美三级三区| 黄色a级毛片大全视频| 亚洲久久久国产精品| 色综合婷婷激情| 老鸭窝网址在线观看| 久久久久久人人人人人| 午夜福利成人在线免费观看| 99在线人妻在线中文字幕| 不卡av一区二区三区| 给我免费播放毛片高清在线观看| 精品人妻1区二区| 高清在线国产一区| 亚洲人成77777在线视频| 老司机在亚洲福利影院| 18禁黄网站禁片午夜丰满| 中文字幕人成人乱码亚洲影| 欧美最黄视频在线播放免费| 18禁美女被吸乳视频| 精品不卡国产一区二区三区| 国内精品久久久久久久电影| 欧美午夜高清在线| 欧美日本视频| 一区二区三区精品91| 午夜影院日韩av| 九色国产91popny在线| 亚洲九九香蕉| 日本黄色视频三级网站网址| 久久久国产成人精品二区| 欧美黄色片欧美黄色片| 国产免费男女视频| 好男人在线观看高清免费视频 | 免费女性裸体啪啪无遮挡网站| 在线观看www视频免费| 欧美黄色淫秽网站| 精品乱码久久久久久99久播| 香蕉久久夜色| 给我免费播放毛片高清在线观看| 成人三级黄色视频| 免费在线观看日本一区| 国产成人精品无人区| 国产精品综合久久久久久久免费 | 色在线成人网| 我的亚洲天堂| 久久久久久久精品吃奶| 亚洲全国av大片| svipshipincom国产片| 成人亚洲精品一区在线观看| 女同久久另类99精品国产91| 亚洲免费av在线视频| 免费在线观看完整版高清| 国产精品亚洲美女久久久| 欧美成狂野欧美在线观看| 巨乳人妻的诱惑在线观看| 亚洲色图综合在线观看| 夜夜爽天天搞| 亚洲精品在线观看二区| 9色porny在线观看| 日日干狠狠操夜夜爽| 亚洲国产欧美日韩在线播放| 精品日产1卡2卡| 美女高潮到喷水免费观看| 黄色成人免费大全| 亚洲人成伊人成综合网2020| 亚洲,欧美精品.| 亚洲va日本ⅴa欧美va伊人久久| 久久香蕉激情| 黄色女人牲交| 女人爽到高潮嗷嗷叫在线视频| 青草久久国产| 一a级毛片在线观看| 精品高清国产在线一区| 一本久久中文字幕| 搡老岳熟女国产| 久久 成人 亚洲| 免费看十八禁软件| 亚洲欧美日韩无卡精品| 99精品在免费线老司机午夜| 国产精品久久视频播放| 欧美日本亚洲视频在线播放| 怎么达到女性高潮| 亚洲黑人精品在线| 精品人妻1区二区| 精品一区二区三区av网在线观看| 国产成人精品无人区| 日韩精品青青久久久久久| 最新美女视频免费是黄的| 51午夜福利影视在线观看| 18禁观看日本| 国产aⅴ精品一区二区三区波| 一夜夜www| 999久久久国产精品视频| www.熟女人妻精品国产| 欧美日韩乱码在线| 国产欧美日韩一区二区三区在线| 成年女人毛片免费观看观看9| 变态另类丝袜制服| 亚洲五月色婷婷综合| 久久草成人影院| 色综合站精品国产| 日本黄色视频三级网站网址| 国产亚洲欧美精品永久| 90打野战视频偷拍视频| 在线观看www视频免费| 亚洲男人的天堂狠狠| or卡值多少钱| 妹子高潮喷水视频| 日韩欧美在线二视频| 免费看美女性在线毛片视频| 一进一出抽搐gif免费好疼| 天堂影院成人在线观看| 欧美成人性av电影在线观看|