A dynamic Unmanned Aerial Vehicle (UAV) path planning method based on Modified Whale Optimization Algorithm (MWOA) was proposed for the problem of UAV path planning in environments with complex terrains. Firstly, by analyzing the mountain terrain, dynamic targets, and threat zones, a three-dimensional dynamic environment and a UAV route model were established. Secondly, an adaptive step size Gaussian walk strategy was proposed to balance the algorithm’s abilities of global exploration and local exploitation. Finally, a supplementary correction strategy was proposed to correct the optimal individual in the population, and combined with differential evolution strategy, the population was avoided from falling into local optimum while improving convergence accuracy of the algorithm. To verify the effectiveness of MWOA, MWOA and intelligent algorithms such as Whale Optimization Algorithm (WOA), and Artificial Hummingbird Algorithm (AHA) were used to solve the CEC2022 test functions, and validated in designed UAV dynamic environment model. The comparative analysis of simulation results shows that compared with the traditional WOA, MWOA improves the convergence accuracy by 6.1%, and reduces the standard deviation by 44.7%. The above proves that the proposed MWOA has faster convergence and higher accuracy, and can handle UAV path planning problems effectively.
In response to the issue of low accuracy in handling complex work conditions in rolling bearing fault diagnosis, a Multi-Task Learning (MTL) model naming as Multi-pathed Hierarchical Mixture-of-Experts (MHMoE), and the corresponding hierarchical training mode were proposed. In this model, by combining multi-stage, multi-task joint training, a hierarchical information sharing mode was achieved. The model's generalization and fault recognition accuracy were further improved on the basis of the ordinary MTL mode, enabling the model to perform tasks on both complex and simple datasets excellently. Meanwhile, by incorporating the bottleneck layer structure of one-dimensional ResNet, the depth of the network was ensured while avoiding issues such as vanishing and exploding gradients, so as to extract relevant features of the dataset fully. Experimental results on the Paderborn University bearing fault dataset (PU) as the test dataset demonstrate that under varying degrees of working complexity, compared to the OMoE (One-gate Mixture-of-Experts) -ResNet18 model without MTL, the proposed model has the accuracy improved by 5.45 to 9.30 percentage points. Compared to the models such as Ensemble Empirical Mode Decomposition Hilbert spectral transform (EEMD-Hilbert), MMoE (Multi-gate Mixture-of-Experts), and Multi-Scale multi-Task Attention Convolutional Neural Network (MSTACNN), the proposed model has the accuracy improved by 3.21 to 16.45 percentage points at least.
Federated Learning(FL) has experienced rapid development due to its advantages in distributed structure and privacy security. However, the fairness issues caused by large-scale FL affect the sustainability of FL systems. In response to the fairness issues in FL, recent researches on fairness in FL was reviewed systematically and analyzed deeply. Firstly, the workflow and definitions of FL were explained, and biases and fairness concepts in FL were summarized. Secondly, commonly used datasets in fairness research of FL were detailed, and the challenges faced by fairness research were discussed. Finally, the advantages, disadvantages, applicable scenarios, and experimental setting of relevant research work were summed up from four aspects: data source selection, model optimization, contribution evaluation, and incentive mechanism, and the future research directions and trends in fairness of FL were prospected.
Federated learning is a distributed learning approach for solving the data sharing problem and privacy protection problem in machine learning, in which multiple parties jointly train a machine learning model and protect the privacy of data. However, there are security threats inherent in federated learning, which makes federated learning face great challenges in practical applications. Therefore, analyzing the attacks faced by federation learning and the corresponding defensive measures are crucial for the development and application of federation learning. First, the definition, process and classification of federated learning were introduced, and the attacker model in federated learning was introduced. Then, the possible attacks in terms of both robustness and privacy of federated learning systems were introduced, and the corresponding defense measures were introduced as well. Furthermore, the shortcomings of the defense schemes were also pointed out. Finally, a secure federated learning system was envisioned.
In response to the problems that the existing Transformer-based Multivariate Long-Term Series Forecasting (MLTSF) models mainly extract features from the time domain, and it is difficult to find out reliable dependencies directly from the dispersed time points of the long-term series, a new decomposition and frequency domain feature extraction model was proposed. Firstly, a periodic term-trend term decomposition method based on the frequency domain was proposed, which reduced the time complexity of the decomposition process. Then, based on the extraction of trend features using periodic term-trend term decomposition, a Transformer network performing frequency domain feature extraction based on Gabor transform was utilized to capture periodic dependencies, which enhanced the stability and robustness of forecasting. Experimental results on five benchmark datasets show that compared with the current state-of-the-art methods, the proposed model has the Mean Squared Error (MSE) in MLTSF is reduced by an average of 7.6% with a maximum reduction of 18.9%, which demonstrates that the proposed model improves forecasting accuracy effectively.
Addressing challenges in time series prediction tasks, such as high-dimensional features, large-scale data, and the demand for high prediction accuracy, a multi-scale trend-period decomposition model based on a multi-head gated dilated convolutional network was proposed. A multi-scale decomposition approach was employed to decompose the original covariate sequence and the prediction variable sequence into their respective periodic terms and trend terms, thereby enabling independent prediction. For the periodic terms, the multi-head gated dilated convolutional network encoder was introduced to extract respective periodic information; in the decoder stage, channel information interaction and fusion were performed through the utilization of a cross-attention mechanism, and after sampling and aligning the periodic information of the prediction variables, the periodic prediction was performed through time attention and channel fusion information. The trend terms prediction was executed by using an autoregressive approach. Finally, the prediction sequence was obtained by incorporating the trend prediction results with the periodic prediction results. Compared with multiple mainstream benchmark models such as Long Short-Term Memory (LSTM) and Informer, on five datasets including ETTm1 and ETTh1, a reduction in Mean Squared Error (MSE) is observed, ranging from 19.2% to 52.8% on average, a decrease in Mean Absolute Error (MAE) is noted, ranging from 12.1% to 33.8% on average. Ablation experiments confirm that the proposed multi-scale decomposition module, multi-head gated dilation convolution, and time attention module can enhance the accuracy of time series prediction.
Aiming at the problem of poor denoising effect and long training period in image denoising, an image denoising model based on approximate U-shaped network structure was proposed. Firstly, the original linear network structure was modified to an approximate U-shaped network structure by using convolutional layers with different strides. Then, the image information of different receptive fields was superimposed on each other to preserve the original information of the image as much as possible. Finally, the deconvolutional network layer was introduced for image restoration and further noise removal. Experimental results show that on Set12 and BSD68 test sets: compared with Denoising Convolutional Neural Network (DnCNN) model, the proposed model has an average increase of 0.04 to 0.14 dB on Peak Signal-to-Noise Ratio (PSNR), and an average reduction of 41% on training time, verifying that the proposed model has better denoising effect and shorter training time.
As the existing malicious code detection methods based on deep learning have problems of insufficiency and low accuracy of feature extraction, a malicious code detection method based on attention mechanism and Residual Network (ResNet) called ARMD was proposed. To support the training of this method, the hash values of 47 580 malicious and benign codes were obtained from Kaggle website, and the APIs called by each code were extracted by analysis tool VirusTotal. After that, the called APIs were integrated into 1 000 non-repeated APIs as the detection features, and the training sample data was constructed through these features. Then, the sample data was labeled by determining the benignity and maliciousness based on the VirusTotal analysis results, and the SMOTE (Synthetic Minority Over-sampling Technique) enhancement algorithm was used to equalize the data samples. Finally, the ResNet injecting with the attention mechanism was built and trained to complete the malicious code detection. Experimental results show that the accuracy of malicious code detection of ARMD is 97.76%, and compared with the existing detection methods based on Convolutional Neural Network (CNN) and ResNet models, ARMD has the average precision improved by at least 2%, verifying the effectiveness of ARMD.
Heuristic and machine learning based code smell detection methods have been proved to have limitations, and most of these methods focus on the common code smells. In order to solve these problems, a deep learning based method was proposed to detect three relatively rare code smells which are related to coupling, those are Intensive Coupling, Dispersed Coupling and Shotgun Surgery. First, the metrics of three code smells were extracted, and the obtained data were processed. Second, a deep learning model combining Convolutional Neural Network (CNN) and attention mechanism was constructed, and the introduced attention mechanism was able to assign weights to the metric features. The datasets were extracted from 21 open source projects, and the detection methods were validated in 10 open source projects and compared with CNN model. Experimental results show that the proposed model achieves the better performance with the code smell precisions of 93.61% and 99.76% for Intensive Coupling and Dispersed Coupling respectively, and the CNN model achieves the better results with the code smell precision of 98.59% for Shotgun Surgery.
Influence maximization is one of the important issues in social network analysis, which aims to identify a small group of seed nodes. When these nodes act as initial spreaders, information can be spread to the remaining nodes as much as possible in the network. The existing heuristic algorithms based on network topology usually only consider one single network centrality, failing to comprehensively combine node characteristics and network topology; thus, their performance is unstable and can be easily affected by the network structure. To solve the above problem, an influence maximization algorithm based on Node Coverage and Structural Hole (NCSH) was proposed. Firstly, the coverages and grid constraint coefficients of all nodes were calculated. Then the seed was selected according to the principle of maximum coverage gain. Secondly, if there were multiple nodes with the same gain, the seed was selected according to the principle of minimum grid constraint coefficient. Finally, the above steps were performed repeatedly until all seeds were selected. The proposed NCSH maintains good performance on six real networks under different numbers of seeds and different spreading probabilities. NCSH achieves 3.8% higher node coverage than to the similar NCA (Node Coverage Algorithm) on average, and 43% lower time consumption than the similar SHDD (maximization algorithm based on Structure Hole and DegreeDiscount). The experimental results show that the NCSH can effectively solve the problem of influence maximization.
Focusing on the issue that the single point of failure cannot be efficiently handled by streaming data processing system Flink, a new fault?tolerant system based on incremental state and backup, Flink+, was proposed. Firstly, backup operators and data paths were established in advance. Secondly, the output data in the data flow diagram was cached, and disks were used if necessary. Thirdly, task state synchronization was performed during system snapshots. Finally, backup tasks and cached data were used to recover calculation in case of system failure. In the system experiment and test, Flink+ dose not significantly increase the additional fault tolerance overhead during fault?free operation; when dealing with the single point of failure in both single?machine and distributed environments, compared with Flink system, the proposed system has the failure recovery time reduced by 96.98% in single?machine 8?task parallelism and by 88.75% in distributed 16?task parallelism. Experimental results show that using incremental state and backup method together can effectively reduce the recovery time of the single point of failure of the stream system and enhance the robustness of the system.
Text feature is the key part of natural language processing. Concerning the problems of high dimensionality and sparseness of text features, a text feature selection method based on Word2Vec word embedding and Genetic AlgoRithm for Biomarker selection in high-dimensional Omics (GARBO) was proposed, so as to facilitate the subsequent text classification tasks. Firstly, the data input form was optimized, and the Word2Vec word embedding method was used to transform the text into the word vectors similar to gene expression. Then, the gene expression simulated by the high-dimensional word vectors was iteratively evolved. Finally, the random forest classifier was used to classify the text after feature selection. The experiments were conducted on the Chinese comment dataset to verify the proposed method. The experimental results show that, the optimized GARBO feature selection method is effective in text feature selection, successfully reducing 300-dimensional features to 50-dimensional features with more value, and has the classification accuracy reached 88%. Compared with other filtering type text feature selection methods, the proposed method can effectively reduce the dimension of text features and improve the effect of text classification.
Aiming at the limitations of easily falling into local minimum and poor stability in simple Monkey-King Genetic Algorithm (MKGA), a MKGA by Immune Evolutionary Hybridized (MKGAIEH) was proposed. MKGAIEH divided the total population into several sub-groups. In order to make full use of the best individual (monkey-king) information of total population, the Immune Evolutionary Algorithm (IEA) was introduced to iterative calculation. In addition, for the other individuals in the sub-groups, the crossover and mutation operations were performed on the monkey-kings of sub-groups and total population. As local searches of all sub-groups were completed, the solutions of sub-groups were mixed again. As the iteration proceeds, this strategy combined the global information exchange with local search is not only to avoid the premature convergence, but also to approximate the global optimal solution with a higher accuracy. Comparison experiments on 6 test functions using MKGAIEH, MKGA, Improved MKGA (IMKGA), Bee Evolutionary Genetic Algorithm (BEGA), Algorithm of Shuffled Frog Leaping based on Immune Evolutionary Particle Swarm Optimization (IEPSOSFLA), and Common climbing Operator Genetic Algorithm (COGA) were given. The results show that the MKGAIEH can find the global optimal solutions for all 6 test functions, and the mean values and standard deviation accuracy of 5 test functions achieve the minimums with improving several orders of magnitude than those of the comparison algorithms. Therefore, MKGAIEH has the optimal searching ability and the stability all the better.
JOMP that is the OpenMP-like implementation in Java needs to be optimized, so a parallel framework, which can separate parallel logic and logic function, was proposed.The parallel framework was implemented by a parallel library named waxberry, and the parts which need to be processed parallelly were annotated and executed by using Aspect-Oriented Programming (AOP) and run-time reflection. AOP was used to separate parallel parts with core ones, and to weave them together. Run-time reflection was used to obtain the related information during the parallel execution. The library waxberry was evaluated using Java Grande Forum (JGF) benchmarks on a quad-core processor. The experimental results show that the waxberry can obtain good performance.