Project Articles

    Frontier and comprehensive applications

    Default Latest Most Read
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Review of application analysis and research progress of deep learning in weather forecasting
    Runting DONG, Li WU, Xiaoying WANG, Tengfei CAO, Jianqiang HUANG, Qin GUAN, Jiexia WU
    Journal of Computer Applications    2023, 43 (6): 1958-1968.   DOI: 10.11772/j.issn.1001-9081.2022050745
    Abstract1226)   HTML91)    PDF (1570KB)(1423)       Save

    With the advancement of technologies such as sensor networks and global positioning systems, the volume of meteorological data with both temporal and spatial characteristics has exploded, and the research on deep learning models for Spatiotemporal Sequence Forecasting (STSF) has developed rapidly. However, the traditional machine learning methods applied to weather forecasting for a long time have unsatisfactory effects in extracting the temporal correlations and spatial dependences of data, while the deep learning methods can extract features automatically through artificial neural networks to improve the accuracy of weather forecasting effectively, and have a very good effect in encoding long-term spatial information modeling. At the same time, the deep learning models driven by observational data and Numerical Weather Prediction (NWP) models based on physical theories are combined to build hybrid models with higher prediction accuracy and longer prediction time. Based on these, the application analysis and research progress of deep learning in the field of weather forecasting were reviewed. Firstly, the deep learning problems in the field of weather forecasting and the classical deep learning problems were compared and studied from three aspects: data format, problem model and evaluation metrics. Then, the development history and application status of deep learning in the field of weather forecasting were looked back, and the latest progress in combining deep learning technologies with NWP was summarized and analyzed. Finally, the future development directions and research focuses were prospected to provide a certain reference for future deep learning research in the field of weather forecasting.

    Table and Figures | Reference | Related Articles | Metrics
    Path planning method of unmanned aerial vehicle based on chaos sparrow search algorithm
    TANG Andi, HAN Tong, XU Dengwu, XIE Lei
    Journal of Computer Applications    2021, 41 (7): 2128-2136.   DOI: 10.11772/j.issn.1001-9081.2020091513
    Abstract1037)      PDF (1479KB)(1432)       Save
    Focusing on the issues of large alculation amount and difficult convergence of Unmanned Aerial Vehicle (UAV) path planning, a path planning method based on Chaos Sparrow Search Algorithm (CSSA) was proposed. Firstly, a two-dimensional task space model and a path cost model were established, and the path planning problem was transformed into a multi-dimensional function optimization problem. Secondly, the cubic mapping was used to initialize the population, and the Opposition-Based Learning (OBL) strategy was used to introduce elite particles, so as to enhance the diversity of the population and expand the scope of the search area. Then, the Sine Cosine Algorithm (SCA) was introduced, and the linearly decreasing strategy was adopted to balance the exploitation and exploration abilities of the algorithm. When the algorithm fell into stagnation, the Gaussian walk strategy was adopted to make the algorithm jump out of the local optimum. Finally, the performance of the proposed improved algorithm was verified on 15 benchmark test functions and applied to solve the path planning problem. Simulation results show that CSSA has better optimization performance than Particle Swarm Optimization (PSO) algorithm, Beetle Swarm Optimization (BSO) algorithm, Whale Optimization Algorithm (WOA), Grey Wolf Optimizer (GWO) algorithm and Sparrow Search Algorithm (SSA), and can quickly obtain a safe and feasible path with optimal cost and satisfying constraints, which proves the effectiveness of the proposed method.
    Reference | Related Articles | Metrics
    High-accuracy localization algorithm based on fusion of two-dimensional code vision and laser lidar
    LUAN Jianing, ZHANG Wei, SUN Wei, ZHANG Ao, HAN Dong
    Journal of Computer Applications    2021, 41 (5): 1484-1491.   DOI: 10.11772/j.issn.1001-9081.2020081162
    Abstract890)      PDF (2182KB)(910)       Save
    Traditional laser localization algorithms such as Monte Carlo localization algorithm have the problems of low accuracy and poor anti-robot kidnapping performance, and traditional two-dimensional code localization algorithms have complex environmental layout and strict limitation to robot's trajectory. In order to solve these problems, a mobile robot localization algorithm based on two-dimensional code vision and laser lidar data was proposed. Firstly, the computer vision technology was used by the robot to detect two-dimensional codes in the test environment, and the poses of detecting two-dimensional codes were transformed to map coordinates respectively, and they were fused to generate the prior pose information. Then the optimized pose was obtained by the point cloud alignment with the generated information as the initial poses. At the same time, the odometry-vision supervising mechanism was introduced to effectively solve the problems brought by the environmental factors such as the information lack of two-dimensional codes and the wrong recognition of the two-dimensional codes as well as ensure the smoothness of the poses. Finally, experimental results based on mobile robot show that, the proposed algorithm has the average error of lidar sampling points reduced by 92%, the average time spent per pose calculation reduced by 88% compared with the classical Adaptive Monto Carlo Localization (AMCL) algorithm, and it solves robot kidnapping problem effectively. This algorithm can be applied to the indoor robots such as storage robot.
    Reference | Related Articles | Metrics
    Fall detection algorithm integrating motion features and deep learning
    CAO Jianrong, LYU Junjie, WU Xinying, ZHANG Xu, YANG Hongjuan
    Journal of Computer Applications    2021, 41 (2): 583-589.   DOI: 10.11772/j.issn.1001-9081.2020050705
    Abstract863)      PDF (1348KB)(860)       Save
    In order to use computer vision technology to accurately detect the fall of the elderly, aiming at the incompleteness of existing fall detection algorithms caused by artificial designing of features and the problems in the fall detection process such as the difficulty of separating foreground and background, the confusion of objects, the loss of moving objects, and the low accuracy of fall detection, a deep learning fall detection algorithm with the fusion of human motion information was proposed to detect the fall state of human body. Firstly, foreground and background were separated by the improved YOLOv3 network, and human object was marked by minimum bounding rectangle according to the detection results of YOLOv3 network. Then, by analyzing the motion features in the process of human fall, the motion features of human body were vectorized and transformed into the motion weight information between 0 and 1 through the Sigmoid activation function. Finally, in order to classify human falls, the motion features and the features extracted by Convolutional Neural Network (CNN) were spliced and fused through the fully connected layer. The proposed fall detection algorithm was compared with human object detection algorithms such as background difference, Gaussian mixture, VIBE (VIsual Background Extractor), Histogram of Oriented Gradient (HOG) and human fall judgment schemes such as threshold method, grading method, Support Vector Machine (SVM) classification, CNN classification, and tested under different lighting conditions and the interference of mixed daily noise motion. The results show that the proposed algorithm is superior to traditional human fall detection algortihms in environmental adaptability and fall detection accuracy. The proposed algorithm can effectively detect the human body in the video and accurately detect the fall state of human body, which further verifies the feasibility and efficiency of the deep learning recognition method with the fusion of motion information in the video fall behavior analysis.
    Reference | Related Articles | Metrics
    Plant leaf disease recognition method based on lightweight convolutional neural network
    JIA Heming, LANG Chunbo, JIANG Zichao
    Journal of Computer Applications    2021, 41 (6): 1812-1819.   DOI: 10.11772/j.issn.1001-9081.2020091471
    Abstract687)      PDF (1486KB)(469)       Save
    Aiming at the problems of low accuracy and poor real-time performance of plant leaf disease recognition in the field of agricultural information, a plant leaf disease recognition method based on lightweight Convolutional Neural Network (CNN) was proposed. The Depthwise Separable Convolution (DSC) and Global Average Pooling (GAP) methods were introduced in the original network to replace the standard convolution operation and the fully connected layer part at the end of the network respectively. At the same time, the technique of batch normalization was also applied to the process of training network to improve the intermediate layer data distribution and increase the convergence speed. In order to comprehensively and reliably evaluate the performance of the proposed method, experiments were conducted on the open plant leaf disease image dataset PlantVillage, and loss function convergence curve, test accuracy, parameter memory demand and other indicators were selected to verify the effectiveness of the improved strategy. Experimental results show that the improved network has higher disease recognition accuracy (99.427%) and smaller memory space occupation (6.47 MB), showing that it is superior to other leaf recognition technologies based on neural network, and has strong engineering practicability.
    Reference | Related Articles | Metrics
    Attention-based object detection with millimeter wave radar-lidar fusion
    LI Chao, LAN Hai, WEI Xian
    Journal of Computer Applications    2021, 41 (7): 2137-2144.   DOI: 10.11772/j.issn.1001-9081.2020081334
    Abstract673)      PDF (1710KB)(557)       Save
    To address problems of missing occluded objects, distant objects and objects in extreme weather scenarios when using lidar for object detection in autonomous driving, an attention-based object detection method with millimeter wave radar-lidar feature fusion was proposed. Firstly, the scan frame data of millimeter wave radar and lidar were aggregated into their respective labeled frames, and the points of millimeter wave radar and lidar were spatially aligned, then PointPillar was employed to encode both the millimeter wave radar and lidar data into pseudo images. Finally, the features of both millimeter wave radar and lidar sensors were extracted by the middle convolution layer, and the features maps of them were fused by attention mechanism, and the fused feature map was passed through a single-stage detector to obtain detection results. Experimental results on nuScenes dataset show that compared to the basic PointPillar network, the mean Average Precision (mAP) of the proposed attention fusion algorithm is higher, which performs better than concatenation fusion, multiply fusion and add fusion methods. The visualization results show that the proposed method is effective and can improve the robustness of the network for detecting occluded objects, distant objects and objects surrounded by rain and fog.
    Reference | Related Articles | Metrics
    Research progress on driver distracted driving detection
    QIN Binbin, PENG Liangkang, LU Xiangming, QIAN Jiangbo
    Journal of Computer Applications    2021, 41 (8): 2330-2337.   DOI: 10.11772/j.issn.1001-9081.2020101691
    Abstract671)      PDF (2153KB)(454)       Save
    With the rapid development of the vehicle industry and world economy, the number of private cars continues to increase, which results in more and more traffic accidents, and traffic safety problem has become a global hotpot. The research of driver distracted driving detection is mainly divided into two types:traditional Computer Vision (CV) algorithms and deep learning algorithms. In the driver distraction detection based on traditional CV algorithm, image features are extracted by the feature operators such as Scale-Invariant Feature Transform (SIFT) and Histogram of Oriented Gradient (HOG), then Support Vector Machine (SVM) is combined to build model and classify the images. However, the traditional CV algorithms have disadvantages of high requirements for the environment, narrow application range, large amount of parameters and high computational complexity. In recent years, deep learning has shown excellent performance such as fast speed and high precision in extracting data features. Therefore, the researchers began to introduce deep learning into driver distracted driving detection. The methods based on deep learning can realize the end-to-end distracted driving detection network with high accuracy. The research status of the traditional CV algorithms and deep learning algorithms in driver distracted driving detection was introduced. Firstly, the situations of the traditional CV algorithms used in the image field and the research of driver distracted driving detection were elaborated. Secondly, the research of driver distracted driving based on deep learning was introduced. Thirdly, the accuracies and model parameters of different driver distracted driving detection methods were compared and analyzed. Finally, the existing research was summarized and three problems that driver distracted driving detection need to solve in the future were put forward:the driver's distraction state and the distraction degree division standards need to be further improved, three aspects of person-car-road need to be considered comprehensively, and how to reduce neural network parameters more effectively.
    Reference | Related Articles | Metrics
    Survey of unmanned aerial vehicle cooperative control
    MA Ziyu, HE Ming, LIU Zujun, GU Lingfeng, LIU Jintao
    Journal of Computer Applications    2021, 41 (5): 1477-1483.   DOI: 10.11772/j.issn.1001-9081.2020081314
    Abstract671)      PDF (1364KB)(1485)       Save
    Unmanned Aerial Vehicle (UAV) cooperative control means that a group of UAVs based on inter-aircraft communication complete a common mission with rational division of labor and cooperation by using swarm intelligence as the core. UAV swarm is a multi-agent system in which many UAVs with certain independence ability carry out various tasks based on local rules. Compared with a single UAV, UAV swarm has great advantages such as high efficiency, high flexibility and high reliability. In view of the latest developments of UAV cooperative control technology in recent years, firstly, the application prospect of multi-UAV technology was illustrated by giving examples from the perspectives of civil use and military use. Then, the differences and development statuses of the three mainstream cooperative control methods:consensus control, flocking control and formation control were compared and analyzed. Finally, some suggestions on delay, obstacle avoidance and endurance of cooperative control were given to provide some help for the research and development of UAV collaborative control in the future.
    Reference | Related Articles | Metrics
    Runoff forecast model based on graph attention network and dual-stage attention mechanism
    Hexuan HU, Huachao SUI, Qiang HU, Ye ZHANG, Zhenyun HU, Nengwu MA
    Journal of Computer Applications    2022, 42 (5): 1607-1615.   DOI: 10.11772/j.issn.1001-9081.2021050829
    Abstract582)   HTML11)    PDF (2505KB)(167)       Save

    To improve the accuracy of watershed runoff volume prediction, and considering the lack of model transparency and physical interpretability of data-driven hydrological model, a new runoff forecast model named Graph Attention neTwork and Dual-stage Attention mechanism-based Long Short-Term Memory network (GAT-DALSTM) was proposed. Firstly, based on the hydrological data of watershed stations, graph neural network was introduced to extract the topology of watershed stations and generate the feature vectors. Secondly, according to the characteristics of hydrological time series data, a runoff forecast model based on dual-stage attention mechanism was established to predict the watershed runoff volume, and the reliability and transparency of the proposed model were verified by the model evaluation method based on attention coefficient heat map. On the Tunxi watershed dataset, the proposed model was compared with Graph Convolution Neural network (GCN) and Long Short-Term Memory network (LSTM) under each prediction step. Experimental results show that, the Nash-Sutcliffe efficiency coefficient of the proposed model is increased by 3.7% and 4.9% on average respectively, which verifies the accuracy of GAT-DALSTM runoff forecast model. By analyzing the heat map of attention coefficient from the perspectives of hydrology and application, the reliability and practicability of the proposed model were verified. The proposed model can provide technical support for improving the prediction accuracy and model transparency of watershed runoff volume.

    Table and Figures | Reference | Related Articles | Metrics
    UAV cluster cooperative combat decision-making method based on deep reinforcement learning
    Lin ZHAO, Ke LYU, Jing GUO, Chen HONG, Xiancai XIANG, Jian XUE, Yong WANG
    Journal of Computer Applications    2023, 43 (11): 3641-3646.   DOI: 10.11772/j.issn.1001-9081.2022101511
    Abstract560)   HTML12)    PDF (2944KB)(403)       Save

    When the Unmanned Aerial Vehicle (UAV) cluster attacks ground targets, it will be divided into two formations: a strike UAV cluster that attacks the targets and a auxiliary UAV cluster that pins down the enemy. When auxiliary UAVs choose the action strategy of aggressive attack or saving strength, the mission scenario is similar to a public goods game where the benefits to the cooperator are less than those to the betrayer. Based on this, a decision method for cooperative combat of UAV clusters based on deep reinforcement learning was proposed. First, by building a public goods game based UAV cluster combat model, the interest conflict problem between individual and group in cooperation of intelligent UAV clusters was simulated. Then, Muti-Agent Deep Deterministic Policy Gradient (MADDPG) algorithm was used to solve the most reasonable combat decision of the auxiliary UAV cluster to achieve cluster victory with minimum loss cost. Training and experiments were performed under conditions of different numbers of UAV. The results show that compared to the training effects of two algorithms — IDQN (Independent Deep Q-Network) and ID3QN (Imitative Dueling Double Deep Q-Network), the proposed algorithm has the best convergence, its winning rate can reach 100% with four auxiliary UAVs, and it also significantly outperforms the comparison algorithms with other UAV numbers.

    Table and Figures | Reference | Related Articles | Metrics
    Application of deep learning in histopathological image classification of aortic medial degeneration
    SUN Zhongjie, WAN Tao, CHEN Dong, WANG Hao, ZHAO Yanli, QIN Zengchang
    Journal of Computer Applications    2021, 41 (1): 280-285.   DOI: 10.11772/j.issn.1001-9081.2020060895
    Abstract541)      PDF (1150KB)(541)       Save
    Thoracic Aortic Aneurysm and Dissection (TAAD) is one of the life-threatening cardiovascular diseases, and the histological changes of Medial Degeneration (MD) have important clinical significance for the diagnosis and early intervention of TAAD. Focusing on the issue that the diagnosis of MD is time-consuming and prone to poor consistency because of the great complexity in histological images, a deep learning based classification method of histological images was proposed, and it was applied to four types of MD pathological changes to verify its performance. In the method, an improved Convolutional Neural Network (CNN) model was employed based on the GoogLeNet. Firstly, transfer learning was adopted for applying the prior knowledge to the expression of TAAD histopathological images. Then, Focal loss and L2 regularization were utilized to solve the data imbalance problem, so as to optimize the model performance. Experimental results show that the proposed model is able to achieve the average accuracy of four-class classification of 98.78%, showing a good generalizability. It can be seen that the proposed method can effectively improve the diagnostic efficiency of pathologists.
    Reference | Related Articles | Metrics
    Prediction method of capacity data in telecom industry based on recurrent neural network
    DING Yin, SANG Nan, LI Xiaoyu, WU Feizhou
    Journal of Computer Applications    2021, 41 (8): 2373-2378.   DOI: 10.11772/j.issn.1001-9081.2020101677
    Abstract503)      PDF (1094KB)(377)       Save
    In the capacity prediction process of telecom operation and maintenance, there are problems of too many capacity indicators and deployed business classes. Most of the existing researches do not consider the difference of indicator data types, and use the same prediction method for all types of data, which results in both good and bad prediction effects. In order to improve the efficiency of indicator prediction, a classification method of data type was proposed, and the data types were divided into trend type, periodic type and irregular type. Aiming at the prediction of periodical data, a periodic capacity indicator prediction model based on Bi-directional Recurrent Neural Network (BiRNN), called BiRNN-BiLSTM-BI, was proposed. Firstly, In order to analyze the periodic characteristics of capacity data, a busy and idle distribution analysis algorithm was proposed. Secondly, a Recurrent Neural Network (RNN) model was built, which included a layer of BiRNN and a layer of Bi-directional Long Short-Term Memory network (BiLSTM). Finally, the output of BiRNN was optimized by the system's busy and idle distribution information. Experimental results compared with the best one among Holt-Winters, AutoRregressive Integrated Moving Average (ARIMA) model and Back Propagation (BP) neural network model show that, the proposed BiRNN-BiLSTM-BI model has the Mean Square Error (MSE) reduced by 15.16% and 45.67% on the unified log dataset and the distributed cache service dataset respectively, showing that the prediction accuracy is greatly improved.
    Reference | Related Articles | Metrics
    Motion control method of two-link manipulator based on deep reinforcement learning
    WANG Jianping, WANG Gang, MAO Xiaobin, MA Enqi
    Journal of Computer Applications    2021, 41 (6): 1799-1804.   DOI: 10.11772/j.issn.1001-9081.2020091410
    Abstract483)      PDF (875KB)(615)       Save
    Aiming at the motion control problem of two-link manipulator, a new control method based on deep reinforcement learning was proposed. Firstly, the simulation environment of manipulator was built, which includes the two-link manipulator, target and obstacle. Then, according to the target setting, state variables as well as reward and punishment mechanism of the environment model, three kinds of deep reinforcement learning models were established for training. Finally, the motion control of the two-link manipulator was realized. After comparing and analyzing the three proposed models, Deep Deterministic Policy Gradient (DDPG) algorithm was selected for further research to improve its applicability, so as to shorten the debugging time of the manipulator model, and avoided the obstacle to reach the target smoothly. Experimental results show that, the proposed deep reinforcement learning method can effectively control the motion of two-link manipulator, the improved DDPG algorithm control model has the convergence speed increased by two times and the stability after convergence enhances. Compared with the traditional control method, the proposed deep reinforcement learning control method has higher efficiency and stronger applicability.
    Reference | Related Articles | Metrics
    3D shale digital core reconstruction method based on deep convolutional generative adversarial network with gradient penalty
    WANG Xianwu, ZHANG Ting, JI Xin, DU Yi
    Journal of Computer Applications    2021, 41 (6): 1805-1811.   DOI: 10.11772/j.issn.1001-9081.2020091367
    Abstract471)      PDF (2129KB)(458)       Save
    Aiming at the problems of high cost, poor reusability and low reconstruction quality in traditional digital core reconstruction technology, a 3D shale digital core reconstruction method based on Deep Convolutional Generation Adversarial Network with Gradient Penalty (DCGAN-GP) was proposed. Firstly, the neural network parameters were used to describe the distribution probability of the shale training image, and the feature extraction of the training image was completed. Secondly, the trained network parameters were saved. Finally, the 3D shale digital core was constructed by using the generator. The experimental results show that, compared to the classic digital core reconstruction technologies, the proposed DCGAN-GP obtains the image closer to the training image in porosity, variogram, as well as pore size and distribution characteristics. Moreover, DCGAN-GP has the CPU usage less than half of the classic algorithms, the memory peak usage only 7.1 GB, and the reconstruction time reached 42 s per time, reflecting the characteristics of high quality and high efficiency of model reconstruction.
    Reference | Related Articles | Metrics
    Cleaning scheduling model with constraints and its solution
    FAN Xiaomao, XIONG Honglin, ZHAO Gansen
    Journal of Computer Applications    2021, 41 (2): 577-582.   DOI: 10.11772/j.issn.1001-9081.2020050735
    Abstract466)      PDF (876KB)(383)       Save
    Cleaning tasks of the cleaning service company often have the characteristics such as different levels, different durations and different cycles, and lack a general cleaning scheduling problem model. At present, the solving of cleaning scheduling problem is mainly relies on manual scheduling scheme, causing the problems such as time-consuming, labor-consuming and unstable scheduling quality. Therefore, a mathematical model of cleaning scheduling problem with constraints, which is a NP-hard problem, was proposed, then Simulated Annealing algorithm (SA), Bee Colony Optimization algorithm (BCO), Ant Colony Optimization algorithm (ACO), and Particle Swarm Optimization algorithm (PSO) were utilized to solve the proposed constrained cleaning scheduling problem. Finally, an empirical analysis was carried out by using the real scheduling state of a cleaning service company. Experimental results show that compared with the manual scheduling scheme, the heuristic intelligent optimization algorithms have obvious advantages in solving the constrained cleaning scheduling problem, and the manpower demand of the obtained cleaning schedule reduced significantly. Specifically, these algorithms can make the cleaning manpower in one year scheduling cycle be saved by 218.62 hours to 513.30 hours compared to manual scheduling scheme. It can be seen that the mathematical models based on heuristic intelligent optimization algorithms are feasible and efficient in solving cleaning scheduling problem with constraints, and provide making-decision supports for the scientific management of the cleaning service company.
    Reference | Related Articles | Metrics
    Spatial-temporal prediction model of urban short-term traffic flow based on grid division
    Haiqi WANG, Zhihai WANG, Liuke LI, Haoran KONG, Qiong WANG, Jianbo XU
    Journal of Computer Applications    2022, 42 (7): 2274-2280.   DOI: 10.11772/j.issn.1001-9081.2021050838
    Abstract463)      PDF (2906KB)(374)       Save

    Accurate traffic flow prediction is very important in helping traffic management departments to take effective traffic control and guidance measures and travelers to plan routes reasonably. Aiming at the problem that the traditional deep learning models do not fully consider the spatial-temporal characteristics of traffic data, a CNN-LSTM prediction model based on attention mechanism, namely STCAL (Spatial-Temporal Convolutional Attention-LSTM network), was established under the theoretical frameworks of Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM) unit and with the combination of the spatial-temporal characteristics of urban traffic flow. Firstly, the fine-grained grid division method was used to construct the spatial-temporal matrix of traffic flow. Secondly, CNN model was used as a spatial component to extract the spatial characteristics of urban traffic flow in different periods. Finally, the LSTM model based on attention mechanism was used as a dynamic time component to capture the temporal characteristics and trend variability of traffic flow, and the prediction of traffic flow was realized. Experimental results show that compared with Gated Recurrent Unit (GRU) and Spatio-Temporal Residual Network (ST-ResNet), STCAL model has the Root Mean Square Error (RMSE) index reduced by 17.15% and 7.37% respectively, the Mean Absolute Error (MAE) index reduced by 22.75% and 9.14% respectively, and the coefficient of determination (R2) index increased by 11.27% and 2.37% respectively. At the same time, it is found that the proposed model has the prediction effect on weekdays with high regularity higher than that on weekends, and has the best prediction effect of morning peak on weekdays, showing that it can provide a basis for short-term urban regional traffic flow change monitoring.

    Reference | Related Articles | Metrics
    Short-term trajectory prediction model of aircraft based on attention mechanism and generative adversarial network
    Yuli CHEN, Qiang TONG, Tongtong CHEN, Shoulu HOU, Xiulei LIU
    Journal of Computer Applications    2022, 42 (10): 3292-3299.   DOI: 10.11772/j.issn.1001-9081.2021081387
    Abstract458)   HTML19)    PDF (1549KB)(258)       Save

    Single Long Short-Term Memory (LSTM) network cannot effectively extract key information and cannot accurately fit data distribution in trajectory prediction. In order to solve the problems, a short-term trajectory prediction model of aircraft based on attention mechanism and Generative Adversarial Network (GAN) was proposed. Firstly, different weights were assigned to the trajectory by introducing attention mechanism, so that the influence of important features in the trajectory was able to be improved. Secondly, the trajectory sequence features were extracted by using LSTM, and the convergence net was used to gather all aircraft features within the time step. Finally, the characteristic of GAN optimizing continuously in adversarial game was used to optimize the model in order to improve the model accuracy. Compared with Social Generative Adversarial Network (SGAN), the proposed model has the Average Displacement Error (ADE), Final Displacement Error (FDE) and Maximum Displacement Error (MDE) reduced by 20.0%, 20.4% and 18.3% respectively on the dataset during climb phase. Experimental results show that the proposed model can predict future trajectories more accurately.

    Table and Figures | Reference | Related Articles | Metrics
    LSTM and artificial neural network for urban bus travel time prediction based on spatiotemporal eigenvectors
    ZHANG Xinhuan, LIU Hongjie, SHI Junqing, MAO Chengyuan, MENG Guolian
    Journal of Computer Applications    2021, 41 (3): 875-880.   DOI: 10.11772/j.issn.1001-9081.2020060467
    Abstract448)      PDF (859KB)(540)       Save
    Aiming at the problem that "with the increase of the prediction distance, the prediction of travel time becomes more and more difficult", a comprehensive prediction model of Long Short Term Memory (LSTM) and Artificial Neural Network (ANN) based on spatiotemporal eigenvectors was proposed. Firstly, 24 hours were segmented into 288 time slices to generate time eigenvectors. Secondly, the LSTM time window model was established based on the time slices. This model was able to solve the window movement problem of long-time prediction. Thirdly, the bus line was divided into multiple space slices and the average velocity of the current space slice was used as the instantaneous velocity. At the same time, the predicted time of each space slice would be used as the spatial eigenvector and sent to the new hybrid neural network model named LSTM-A (Long Short Term Memory Artificial neural network). This model combined with the advantages of the two prediction models and solved the problem of bus travel time prediction. Finally, based on the experimental dataset, experiments and tests were carried out:the prediction problem between bus stations was divided into sub-problems of line slice prediction, and the concept of real-time calculation was introduced to each related sub-problem, so as to avoid the prediction error caused by complex road conditions. Experimental results show that the proposed algorithm is superior to single neural network models in both accuracy and applicability. In conclusion, the proposed new hybrid neural network model LSTM-A can realize the long-distance arrival time prediction from the dimension of time feature and the short-distance arrival time prediction from the dimension of spatial feature, thus effectively solving the problem of urban bus travel time prediction and avoiding the remote dependency and error accumulation of buses.
    Reference | Related Articles | Metrics
    Segmentation of ischemic stroke lesion based on long-distance dependency encoding and deep residual U-Net
    HUANG Li, LU Long
    Journal of Computer Applications    2021, 41 (6): 1820-1827.   DOI: 10.11772/j.issn.1001-9081.2020111788
    Abstract446)      PDF (1812KB)(516)       Save
    Segmenting stroke lesions automatically can provide valuable support to the clinical decision process. However, this is a challenging task due to the diversity of lesion size, shape, and location. Previous works have failed to capture global context information which is helpful to handle the diversity. To solve the problem of segmentation of ischemic stroke lesions with small sample size, an end-to-end neural network combing with residual block and non-local block on the basis of traditional U-Net was proposed to predict stroke lesion from multi-modal Magnetic Resonance Imaging (MRI) image. In this method, based on the encoder-decoder architecture of U-Net, residual blocks were stacked to solve the degradation problem and avoid the overfitting, and the non-local blocks were added to effectively encode the long-distance dependencies and provide global context information for the feature extraction process. The proposed method and its variants were evaluated on the Ischemic Stroke Lesion Segmentation (ISLES) 2017 dataset. The results showed that the proposed residual U-Net (Dice=0.29±0.23, ASSD=7.66±6.41, HD=43.71±22.11) and Residual Non-local U-Net (RN-UNet) (Dice=0.29±0.23, ASSD=7.61±6.62, HD=45.36±24.75) achieved significant improvement in all metrics compared to the baseline U-Net (Dice=0.25±0.23, ASSD=9.45±7.36, HD=54.59 ±21.19); compared with the state-of-the-art methods from ISLES website, the two methods both achieved better segmentation results, so that they can help doctors to quickly and objectively evaluate the condition of patients in clinical practices.
    Reference | Related Articles | Metrics
    Classification of steel surface defects based on lightweight network
    SHI Yangxiao, ZHANG Jun, CHEN Peng, WANG Bing
    Journal of Computer Applications    2021, 41 (6): 1836-1841.   DOI: 10.11772/j.issn.1001-9081.2020081244
    Abstract444)      PDF (981KB)(358)       Save
    Defect classification is an important part of steel surface defect detection. When the Convolutional Neural Network (CNN) has achieved good results, the increasing number of network parameters consumes a lot of computing cost, which brings great challenges to the deployment of defect classification tasks on personal computers or low computing power devices. Focusing on the above problem, a novel lightweight network model named Mix-Fusion was proposed. Firstly, two operations of group convolution and channel-shuffle were used to reduce the computational cost while maintaining the accuracy. Secondly, a narrow feature mapping was used to fuse and encode the information between the groups, and the generated features were combined with the original network, so as to effectively solve the problem that "sparse connection" convolution hindered the information exchange between the groups. Finally, a new type of Mixed depthwise Convolution (MixConv) was used to replace the traditional DepthWise Convolution (DWConv) to further improve the performance of the model. Experimental results on NEU-CLS dataset show that, the number of floating-point operations and classification accuracy of Mix-Fusion network in defect classification task is 43.4 Million FLoating-point Operations Per second (MFLOPs) and 98.61% respectively. Compared to the networks of ShuffleNetV2 and MobileNetV2, the proposed Mix-Fusion network reduces the model parameters and compresses the model size effectively, as well as obtains the better classification accuracy.
    Reference | Related Articles | Metrics
    Acceleration compensation based anti-swaying flight control for unmanned aerial vehicle with slung-load
    JIAO Hailin, GUO Yuying, ZHU Zhengwei
    Journal of Computer Applications    2021, 41 (2): 604-610.   DOI: 10.11772/j.issn.1001-9081.2020050740
    Abstract443)      PDF (3609KB)(355)       Save
    In order to reduce the load swing during the slung-load flight of quadrotor Unmanned Aerial Vehicle (UAV), a new anti-sway control method based on acceleration compensation was developed. First, the nonlinear dynamic equation of the quadrotor UAV hanging system was established based on the Lagrange method, and the energy functions were proposed to design the flight control systems, making the quadrotor UAV track the reference trajectory. Then, the generalized error of the motion trajectory of the slung-load was used to design the anti-sway controller, the acceleration compensation of the quadrotor UAV was carried out to modify the motion trajectory of the quadrotor UAV, thereby reducing the slung-load swing caused by the rapid motion of quadrotor UAV. Finally, some simulations were carried out to compare and analyze the effect of the slung-load flight control before and after acceleration compensation. Simulation results show that, the flight control method based on acceleration compensation can not only ensure the stability of the quadrotor UAV hanging flight, but also provide sufficient stability margin for the flight control system.
    Reference | Related Articles | Metrics
    Citrus disease and insect pest area segmentation based on superpixel fast fuzzy C-means clustering and support vector machine
    YUAN Qianqian, DENG Hongmin, WANG Xiaohang
    Journal of Computer Applications    2021, 41 (2): 563-570.   DOI: 10.11772/j.issn.1001-9081.2020050645
    Abstract439)      PDF (1737KB)(607)       Save
    Focused on the existing problems that there are few image datasets of citrus diseases and insect pests, the targets of diseases and pests are complex and scattered, and are difficult to realize automatic location and segmentation, a segmentation method of agricultural citrus disease and pest areas based on Superpixel Fast Fuzzy C-means Clustering (SFFCM) and Support Vector Machine (SVM) was proposed. This method made full use of the advantages of SFFCM algorithm, which was fast and robust, and integrated the characteristics of spatial information, meanwhile, it did not require manual selection of samples in image segmentation like the traditional SVM. Firstly, the improved SFFCM segmentation algorithm was used to pre-segment the image to be segmented to obtain the foreground and background regions. Then, the erosion and dilation operations in morphology were used to narrow these two areas, and the training samples were automatically selected for SVM model training. Finally, the trained SVM classifier was used to segment the entire image. Experimental results show that compared with the following three methods:Fast and Robust Fuzzy C-means Clustering (FRFCM), the original SFFCM and Edge Guidance Network (EGNet), the proposed method has the average recall of 0.937 1, average precision of 0.941 8 and the average accuracy of 0.930 3, all of which are better than those of the comparison methods.
    Reference | Related Articles | Metrics
    Spatial frequency divided attention network for ultrasound image segmentation
    SHEN Xuewen, WANG Xiaodong, YAO Yu
    Journal of Computer Applications    2021, 41 (6): 1828-1835.   DOI: 10.11772/j.issn.1001-9081.2020091470
    Abstract435)      PDF (1917KB)(391)       Save
    Aiming at the problems of medical ultrasound images such as many noisy points, fuzzy boundaries, and difficulty in defining the cardiac contours, a new Spatial Frequency Divided Attention Network for ultrasound image segmentation (SFDA-Net) was proposed. Firstly, with the help of Octave convolution, the high and low-frequency parallel processing of image in the entire network was realized to obtain more diverse information. Then, the Convolutional Block Attention Module (CBAM) was added for paying more attention to the effective information when image feature recovered, so as to reduce the loss of segmenting the entire target area. Finally Focal Tversky Loss was considered as the objective function to reduce the weights of simple samples and pay more attention on difficult samples, as well as decrease the errors introduced by pixel misjudgment between the categories. Through multiple sets of comparative experiments,it can be seen that with the parameter number lower than that of the original UNet++, SFDA-Net has the segmentation accuracy increased by 6.2 percentage points, Dice sore risen by 8.76 percentage points, mean Pixel Accuracy (mPA) improved to 84.09%, and mean Intersection Over Union (mIoU) increased to 75.79%. SFDA-Net steadily improves the network performance while reducing parameters, and makes the echocardiographic segmentation more accurate.
    Reference | Related Articles | Metrics
    Multi-person collaborative creation system of building information modeling drawings based on blockchain
    SHEN Yumin, WANG Jinlong, HU Diankai, LIU Xingyu
    Journal of Computer Applications    2021, 41 (8): 2338-2345.   DOI: 10.11772/j.issn.1001-9081.2020101549
    Abstract434)      PDF (1810KB)(408)       Save
    Multi-person collaborative creation of Building Information Modeling (BIM) drawings is very important in large building projects. However, the existing methods of multi-person collaborative creation of BIM drawings based on Revit and other modeling software or cloud service have the confusion of BIM drawing version, difficulty of traceability, data security risks and other problems. To solve these problems, a blockchain-based multi-person collaborative creation system for BIM drawings was designed. By using the on-chain and off-chain collaborative storage method, the blockchain and database were used to store BIM drawings information after each creation in the BIM drawing creation process and the complete BIM drawings separately. The decentralization, traceability and anti-tampering characteristics of the blockchain were used to ensure that the version of the BIM drawings is clear, and provide a basis for the future copyright division. These characteristics were also used to enhance the data security of BIM drawings information. Experimental results show that the average block generation time of the proposed system in the multi-user concurrent case is 0.467 85 s, and the maximum processing rate of the system is 1 568 transactions per second, which prove the reliability of the system and that the system can meet the needs of actual application scenarios.
    Reference | Related Articles | Metrics
    Battery state-of-charge prediction method based on one-dimensional convolutional neural network combined with long short-term memory network
    NI Shuiping, LI Huifang
    Journal of Computer Applications    2021, 41 (5): 1514-1521.   DOI: 10.11772/j.issn.1001-9081.2020071097
    Abstract420)      PDF (2218KB)(463)       Save
    Focused on the issues of accuracy and stability of battery State-Of-Charge (SOC) prediction and gradient disappearance of deep neural network, a battery SOC prediction method based on the combination of one-Dimensional Convolutional Neural Network (1D CNN) and Long Short-Term Memory (LSTM) Recurrent Neural Network (RNN) named 1D CNN-LSTM (1D CNN combined with LSTM) model was proposed. The current, voltage and resistance of the battery were mapped to the target value SOC by 1D CNN-LSTM model. Firstly, a one-dimensional convolutional layer was used to extract the high-level data features from the sample data and make full use of the feature information of the input data. Secondly, a LSTM layer was used to save the historical input information, so as to effectively prevent the loss of important information. Finally, the prediction results of the battery SOC were outputted through a fully connected layer. The proposed model was trained with the experimental data of multiple cycles of charge-discharge of the battery, the prediction effects of the 1D CNN-LSTM model under different hyperparameter settings were analyzed and compared, and the weight coefficients and bias parameters of the model were adjusted through training the model, so that the optimal model setting was determined. Experimental results show that the 1D CNN-LSTM model has accurate and stable prediction effect of battery SOC. The Mean Absolute Error (MAE), Mean Square Error (MSE) and maximum prediction error of this model are 0.402 7%, 0.002 9% and 0.99% respectively.
    Reference | Related Articles | Metrics
    Improved pyramid evolution strategy for solving split delivery vehicle routing problem
    LI Huafeng, HUANG Zhangcan, ZHANG Qiang, ZHAN Hang, TAN Qing
    Journal of Computer Applications    2021, 41 (1): 300-306.   DOI: 10.11772/j.issn.1001-9081.2020050615
    Abstract417)      PDF (948KB)(401)       Save
    To solve the Split Delivery Vehicle Routing Problem (SDVRP) more reasonably, overcome the shortcoming that the traditional two-stage solution method of first route and then optimization is easy to fall into local optimization, and handle the problem that the intelligent optimization algorithm fails to integrate competition and cooperation organically in the optimization stage, an Improved Pyramid Evolution Strategy (IPES) was proposed with the shortest delivery path and the least delivery vehicles as the optimization objectives. Firstly, based on the pyramid, the encoding and decoding methods and hierarchical cooperation strategy were proposed to solve SDVRP. Secondly, according to the characteristics such as the random of genetic algorithm, high parallelism of "survival of the fittest" and self-adaption, as well as the different labor division of different layers of pyramid structure, an adaptive neighborhood operator suitable for SDVRP was designed to make the algorithm converge fast to the optimum. Finally, the optimal solution was obtained. Compared with the piecewise solving algorithm, clustering algorithm, particle swarm algorithm, artificial bee colony algorithm, taboo search algorithm,the results of four simulation experiments show that, when solving the optimal path of each case, the proposed IPES has the solution accuracy improved by at least 0.92%, 0.35%, 3.07%, 9.40% respectively, which verifies the good performance of IPES in solving SDVRP.
    Reference | Related Articles | Metrics
    Ultrasound thyroid segmentation network based on feature fusion and dynamic multi-scale dilated convolution
    HU Yishan, QIN Pinle, ZENG Jianchao, CHAI Rui, WANG Lifang
    Journal of Computer Applications    2021, 41 (3): 891-897.   DOI: 10.11772/j.issn.1001-9081.2020060783
    Abstract417)      PDF (1326KB)(1472)       Save
    Concerning the the size and morphological diversity of thyroid tissue and the complexity of surrounding tissue in thyroid ultrasound images, an ultrasound thyroid segmentation network based on feature fusion and dynamic multi-scale dilated convolution was proposed. Firstly, the dilated convolutions with different dilation rates and dynamic filters were used to fuse the global semantic information of different receptive domains and the semantic information in the context details with different ranges, so as to improve the adaptability and accuracy of the network to multi-scale targets. Then, the hybrid upsampling method was used to enhance the spatial information of high-dimensional semantic features and the context information of low-dimensional spatial features during feature dimensionality reduction. Finally, the spatial attention mechanism was introduced to optimize the low-dimensional features of the image, and the method of fusing high- and low-dimensional features was applied to retain the useful features of high- and low-dimensional feature information with the elimination of the redundant information and improve the network's ability to distinguish the background and foreground of the image. Experimental results show that the proposed method has an accuracy rate of 0.963±0.026, a recall rate of 0.84±0.03 and a dice coefficient of 0.79±0.03 in the public dataset of thyroid ultrasound images. It can be seen that the proposed method can solve the problems of large difference of tissue morphology and complex surrounding tissues.
    Reference | Related Articles | Metrics
    Macroscopic fundamental diagram traffic signal control model based on hierarchical control
    WANG Peng, LI Yanwen, YANG Di, YANG Huamin
    Journal of Computer Applications    2021, 41 (2): 571-576.   DOI: 10.11772/j.issn.1001-9081.2020050758
    Abstract410)      PDF (1351KB)(677)       Save
    Aiming at the problem of coordinated control within urban traffic sub-areas and boundary intersections, a traffic signal control model based on Hierarchical multi-granularity and Macroscopic Fundamental Diagram (HDMF) was proposed. First, the hierarchical multi-granularity characteristic of the urban traffic system and the rough set theory were used to describe the real-time states of the traffic elements. Then, combined with the distributed intersection signal control based on backpressure algorithm and the dynamic characteristics of the traffic elements, the pressures of the intersection phases were calculated and the phase decision was made. Finally, Macroscopic Fundamental Diagram (MFD) was used to achieve the maximum total flow of vehicles driving out of the area and the optimal number of vehicles in each sub-area. Experimental results showed that HDMF model had the average queue length reduced by 6.35% and 10.01% respectively, and had the average travel time reduced by 6.55% and 11.15% respectively compared with EMP (Extended cooperative Max-Pressure control) model and HGA model based on MFD and hybrid genetic simulated annealing algorithm. It can be seen that the propsed HDMF model can effectively relieve interior and boundary traffic congestions of sub-areas and maximize the traffic flow of the whole road network.
    Reference | Related Articles | Metrics
    Cerebral infarction image recognition based on semi-supervised method
    OU Lili, SHAO Fengjing, SUN Rencheng, SUI Yi
    Journal of Computer Applications    2021, 41 (4): 1221-1226.   DOI: 10.11772/j.issn.1001-9081.2020071034
    Abstract410)      PDF (1167KB)(585)       Save
    In the field of image recognition, images with insufficient label data cannot be well recognized by the supervised method model. In order to solve this problem, a semi-supervised method model based on Generative Adversarial Network(GAN) was proposed. That is, by combining the advantages of semi-supervised GANs and deep convolutional GANs, and replacing the sigmoid activation function with softmax in the output layer, the Semi-Supervised Deep Convolutional GAN(SS-DCGAN) model was established. Firstly, the generated samples were defined as pseudo-samples and used to guide the training process. Secondly, the semi-supervised training method was adopted to update the parameters of the model. Finally, the recognition of abnormal(cerebral infarction) images was realized. Experimental results show that the SS-DCGAN model can recognize abnormal images well with little label data, which achieves 95.05% recognition rates. Compared with Residual Network 32(ResNet32) and Ladder networks, the SS-DCGAN model has significant advantages.
    Reference | Related Articles | Metrics
    Cattle eye image feature extraction method based on improved DenseNet
    ZHENG Zhiqiang, HU Xin, WENG Zhi, WANG Yuhe, CHENG Xi
    Journal of Computer Applications    2021, 41 (9): 2780-2784.   DOI: 10.11772/j.issn.1001-9081.2020101533
    Abstract410)      PDF (1024KB)(341)       Save
    To address the problem of low recognition accuracy caused by vanishing gradient and overfitting in the cattle eye image feature extraction process, an improved DenseNet based cattle eye image feature extraction method was proposed. Firstly, the Scaled exponential Linear Unit (SeLU) activation function was used to prevent the vanishing gradient of the network. Secondly, the feature blocks of cattle eye images were randomly discarded by DropBlock, so as to prevent overfitting and strengthen the generalization ability of the network. Finally, the improved dense layers were superimposed to form an improved Dense convolutional Network (DenseNet). Feature information extraction recognition experiments were conducted on the self-built cattle eyes image dataset. Experimental results show that the recognition accuracy, precision and recall of the improved DenseNet are 97.47%, 98.11% and 97.90% respectively, and compared to the network without improvement, the above recognition accuracy rate, precision rate, recall rate are improved by 2.52 percentage points, 3.32 percentage points, 2.94 percentage points respectively. It can be seen that the improved network has higher precision and robustness.
    Reference | Related Articles | Metrics
2024 Vol.44 No.3

Current Issue
Archive
Honorary Editor-in-Chief: ZHANG Jingzhong
Editor-in-Chief: XU Zongben
Associate Editor: SHEN Hengtao XIA Zhaohui
Domestic Post Distribution Code: 62-110
Foreign Distribution Code: M4616
Address:
No. 9, 4th Section of South Renmin Road, Chengdu 610041, China
Tel: 028-85224283-803
  028-85222239-803
Website: www.joca.cn
E-mail: bjb@joca.cn
WeChat
Join CCF