Project Articles

    The 18th China Conference on Machine Learning

    Default Latest Most Read
    Please wait a minute...
    For Selected: Toggle Thumbnails
    Deep distance metric learning method based on optimized triplet loss
    Zilong LI, Yong ZHOU, Rong BAO, Hongdong WANG
    Journal of Computer Applications    2021, 41 (12): 3480-3484.   DOI: 10.11772/j.issn.1001-9081.2021061107
    Abstract319)   HTML5)    PDF (581KB)(105)       Save

    Focused on the issues that the single deep distance metric based on triplet loss has poor adaptability to the diversified datasets and easily leads to overfitting, a deep distance metric learning method based on optimized triplet loss was proposed. Firstly, by thresholding the relative distance of triplet training samples mapped by neural network, and a piecewise linear function was used as the evaluation function of relative distance. Secondly, the evaluation function was added to the Boosting algorithm as a weak classifier to generate a strong classifier. Finally, an alternating optimization method was used to learn the parameters of the weak classifier and neural network. Through the evaluation of various deep distance metric learning methods in the image retrieval task, it can be seen that the Recall@1 of the proposed method is 4.2, 3.2 and 0.6 higher than that of the previous best score on CUB-200-2011, Cars-196 and SOP datasets respectively. Experimental results show that the proposed method outperforms the comparison methods, while avoiding overfitting to a certain extent.

    Table and Figures | Reference | Related Articles | Metrics
    Label noise filtering method based on dynamic probability sampling
    Zenghui ZHANG, Gaoxia JIANG, Wenjian WANG
    Journal of Computer Applications    2021, 41 (12): 3485-3491.   DOI: 10.11772/j.issn.1001-9081.2021061026
    Abstract264)   HTML13)    PDF (1379KB)(124)       Save

    In machine learning, data quality has a far-reaching impact on the accuracy of system prediction. Due to the difficulty of obtaining information and the subjective and limited cognition of human, experts cannot accurately mark all samples. And some probability sampling methods proposed in resent years fail to avoid the problem of unreasonable and subjective sample division by human. To solve this problem, a label noise filtering method based on Dynamic Probability Sampling (DPS) was proposed, which fully considered the differences between samples of each dataset. By counting the frequency of built-in confidence distribution in each interval and analyzing the trend of information entropy of built-in confidence distribution in each interval, the reasonable threshold was determined. Fourteen datasets were selected from UCI classic datasets, and the proposed algorithm was compared with Random Forest (RF), High Agreement Random Forest Filter (HARF), Majority Vote Filter (MVF) and Local Probability Sampling (LPS) methods. Experimental results show that the proposed method shows high ability on both label noise recognition and classification generalization.

    Table and Figures | Reference | Related Articles | Metrics
    Manifold regularized nonnegative matrix factorization based on clean data
    Hua LI, Guifu LU, Qinru YU
    Journal of Computer Applications    2021, 41 (12): 3492-3498.   DOI: 10.11772/j.issn.1001-9081.2021060962
    Abstract242)   HTML6)    PDF (663KB)(124)       Save

    The existing Nonnegative Matrix Factorization (NMF) algorithms are often designed based on Euclidean distance, which makes the algorithms sensitive to noise. In order to enhance the robustness of these algorithms, a Manifold Regularized Nonnegative Matrix Factorization based on Clean Data (MRNMF/CD) algorithm was proposed. In MRNMF/CD algorithm, the low-rank constraints, manifold regularization and NMF technologies were seamlessly integrated, which makes the algorithm perform relatively excellent. Firstly, by adding the low-rank constraints, MRNMF/CD can recover clean data from noisy data and obtain the global structure of the data. Secondly, in order to use the local geometric structure information of the data, manifold regularization was incorporated into the objective function by MRNMF/CD. In addition, an iterative algorithm for solving MRNMF/CD was proposed, and the convergence of this solution algorithm was analyzed theoretically. Experimental results on ORL, Yale and COIL20 datasets show that MRNMF/CD algorithm has better accuracy than the existing algorithms including k-means, Principal Component Analysis (PCA), NMF and Graph Regularized Nonnegative Matrix Factorization (GNMF).

    Table and Figures | Reference | Related Articles | Metrics
    Unbiased recommendation model based on improved propensity score estimation
    Jinwei LUO, Dugang LIU, Weike PAN, Zhong MING
    Journal of Computer Applications    2021, 41 (12): 3508-3514.   DOI: 10.11772/j.issn.1001-9081.2021060910
    Abstract411)   HTML9)    PDF (567KB)(147)       Save

    In reality, recommender systems usually suffer from various bias problems, such as exposure bias, position bias and selection bias. A recommendation model that ignores the bias problems cannot reflect the real performance of the recommender system, and may be untrustworthy for users. Previous works show that a recommendation model based on propensity score estimation can effectively alleviate the exposure bias problem of implicit feedback data in recommender systems, but only item information is usually considered to estimate propensity scores, which may lead to inaccurate estimation of propensity scores. To improve the accuracy of propensity score estimation, a Match Propensity Estimator (MPE) method was proposed. Specifically, a concept of users’ popularity preference was introduced at first, and then more accurate modeling of the sample exposure rate was achieved by calculating the matching degree of the user’s popularity preference and the item’s popularity. The proposed estimation method was integrated with a traditional recommendation model and an unbiased recommendation model, and the integrated models were compared to three baseline models including the above two models. Experimental results on a public dataset show that the models combining MPE method achieve significant improvement on three evaluation metrics such as recall, Discounted Cumulative Gain (DCG) and Mean Average Precision (MAP) compared with the corresponding baseline models respectively. In addition, experimental results demonstrate that a large part of the performance gain comes from long-tail items, showing that the proposed method is helpful to improve the diversity and coverage of recommended items.

    Table and Figures | Reference | Related Articles | Metrics
    Social collaborative ranking recommendation algorithm by exploiting both explicit and implicit feedback
    Gai LI, Lei LI, Jiaqiang ZHANG
    Journal of Computer Applications    2021, 41 (12): 3515-3520.   DOI: 10.11772/j.issn.1001-9081.2021060908
    Abstract269)   HTML8)    PDF (631KB)(99)       Save

    The traditional social collaborative filtering algorithms based on rating prediction have the inherent deficiency in which the prediction value does not match the real sort, and social collaborative ranking algorithms based on ranking prediction are more suitable to practical application scenarios. However, most existing social collaborative ranking algorithms focus on explicit feedback data only or implicit feedback data only, and not make full use of the information in the dataset. In order to fully exploit both the explicit and implicit scoring information of users’ social networks and recommendation objects, and to overcome the inherent deficiency of traditional social collaborative filtering algorithms based on rating prediction, a new social collaborative ranking model based on the newest xCLiMF model and TrustSVD model, namely SPR_SVD++, was proposed. In the algorithm, both the explicit and implicit information of user scoring matrix and social network matrix were exploited simultaneously and the learning to rank’s evaluation metric Expected Reciprocal Rank (ERR) was optimized. Experimental results on real datasets show that SPR_SVD++ algorithm outperforms the existing state-of-the-art algorithms TrustSVD, MERR_SVD++ and SVD++ over two different evaluation metrics Normalized Discounted Cumulative Gain (NDCG) and ERR. Due to its good performance and high expansibility, SPR_SVD++ algorithm has a good application prospect in the Internet information recommendation field.

    Table and Figures | Reference | Related Articles | Metrics
    Hybrid K-anonymous feature selection algorithm
    Liu YANG, Yun LI
    Journal of Computer Applications    2021, 41 (12): 3521-3526.   DOI: 10.11772/j.issn.1001-9081.2021060980
    Abstract295)   HTML8)    PDF (619KB)(180)       Save

    K-anonymous algorithm makes the data reached the condition of K-anonymity by generalizing and suppressing the data. It can be seen as a special feature selection method named K-anonymous feature selection which considers both data privacy and classification performance. In K-anonymous feature selection method, the characteristics of K-anonymity and feature selection are combined to use multiple evaluation criteria to select the subset of K-anonymous features. It is difficult for the filtered K-anonymous feature selection method to search all the candidate feature subsets satisfying the K-anonymous condition, and the classification performance of the obtained feature subset cannot be guaranteed to be optimal, and the wrapper feature selection method has very high-cost calculation. Therefore, a hybrid K-anonymous feature selection method was designed by combining the characteristics of filtered feature sorting and wrapper feature selection by improving the forward search strategy in the existing methods and thereby using classification performance as the evaluation criterion to select the K-anonymous feature subset with the best classification performance. Experiments were carried out on multiple public datasets, and the results show that the proposed algorithm can outperform the existing algorithms in classification performance and has less information loss.

    Table and Figures | Reference | Related Articles | Metrics
    Event detection without trigger words incorporating syntactic information
    Cui WANG, Yafei ZHANG, Junjun GUO, Shengxiang GAO, Zhengtao YU
    Journal of Computer Applications    2021, 41 (12): 3534-3539.   DOI: 10.11772/j.issn.1001-9081.2021060928
    Abstract240)   HTML6)    PDF (697KB)(94)       Save

    Event Detection (ED) is one of the most important tasks in the field of information extraction, aiming to identify instances of specific event types in text. Existing ED methods usually use adjacency matrix to express syntactic dependencies, however, the adjacency matrix often needs to be encoded with Graph Convolutional Network (GCN) to obtain syntactic information, which increases the complexity of the model. Therefore, an event detection method without trigger words incorporating syntactic information was proposed. After converting the dependent parent word and its context into a position marker vector, the word embedding of dependent sub-word was incorporated at the source end of the model in a parameter-free manner to strengthen the semantic representation of the context, without the need of GCN for encoding. In addition, for the time-consuming and laborious labeling of trigger words, a type perceptron based on the multi-head attention mechanism was designed, which was able to model the potential trigger words in the sentence to complete the event detection without trigger words. In order to verify the performance of the proposed method, experiments were conducted on the ACE2005 dataset and the low-resource Vietnamese dataset. Compared with the Event Detection Using Graph Transformer Network (GTN-ED) method, the F1-score of the proposed method was increased by 3.7% on the ACE2005 dataset; compared with the binary classification method Type-aware Bias Neural Network with Attention Mechanisms (TBNNAM), the F1-score of the proposed method was increased by 9% on the Vietnamese dataset. The results show that the integration of syntactic information into Transformer can effectively connect the scattered event information in the sentence to improve the accuracy of event detection.

    Table and Figures | Reference | Related Articles | Metrics
    Rumor detection model based on user propagation network and message content
    Haitao XUE, Li WANG, Yanjie YANG, Biao LIAN
    Journal of Computer Applications    2021, 41 (12): 3540-3545.   DOI: 10.11772/j.issn.1001-9081.2021060963
    Abstract302)   HTML14)    PDF (697KB)(212)       Save

    Under the constrains of very short message content on social media platforms, a large number of empty forwards in the transmission structure, and the mismatch between user roles and contents, a rumor detection model based on user attribute information and message content in the propagation network, namely GMB_GMU, was proposed. Firstly, user propagation network was constructed with user attributes as nodes and propagation chains as edges, and Graph Attention neTwork (GAT) was introduced to obtain an enhanced representation of user attributes; meanwhile, based on this user propagation network, the structural representation of users was obtained by using node2vec, and it was enhanced by using mutual attention mechanism. In addition, BERT (Bidirectional Encoder Representations from Transformers) was introduced to establish the source post content representation of the source post. Finally, to obtain the final message representation, Gated Multimodal Unit (GMU) was used to integrate the user attribute representation, structural representation and source post content representation. Experimental results show that the GMB_GMU model achieves an accuracy of 0.952 on publicly available Weibo data and can effectively identify rumor events, which is significantly better than the propagation algorithms based on Recurrent Neural Network (RNN) and other neural network benchmark models.

    Table and Figures | Reference | Related Articles | Metrics
    Microblog rumor detection model based on heterogeneous graph attention network
    Bei BI, Huiyao PAN, Feng CHEN, Jingyan SUI, Yang GAO, Yaojun WANG
    Journal of Computer Applications    2021, 41 (12): 3546-3550.   DOI: 10.11772/j.issn.1001-9081.2021060981
    Abstract607)   HTML13)    PDF (541KB)(217)       Save

    Social media highly facilitates people’s daily communication and disseminating information, but it is also a breeding ground for rumors. Therefore, how to automatically monitor rumor dissemination in the early stage is of great practical significance, but the existing detection methods fail to take full advantage of the semantic information of the microblog information propagation graph. To solve this problem, based on Heterogeneous graph Attention Network (HAN), a rumor monitoring model was built, namely MicroBlog-HAN. In the model, a hierarchical attention mechanism including node-level attention and semantic-level attention was adopted. First, the neighbors of microblog nodes were combined by the node-level attention to generate two groups of node embeddings with specific semantics. After that, different semantics were fused by the semantic-level attention to obtain the final node embeddings of microblog, which were then treated as the classifier’s input to perform the binary classification task. In the end, the classification result of whether the input microblog is rumor or not was given. Experimental results on two real-world microblog rumor datasets convincingly prove that MicroBlog-HAN model can accurately identify microblog rumors with an accuracy over 87%.

    Table and Figures | Reference | Related Articles | Metrics
    Vietnamese scene text detection based on modified Mask R-CNN
    Yate FENG, Yimin WEN
    Journal of Computer Applications    2021, 41 (12): 3551-3557.   DOI: 10.11772/j.issn.1001-9081.2021050821
    Abstract260)   HTML12)    PDF (1209KB)(90)       Save

    In view of the lack of training data for Vietnamese scene text detection and the incomplete detection of Vietnamese tone marks in the detection, a text detection algorithm for Vietnamese scenes based on a modified instance segmentation method Mask R-CNN was proposed. In order to segment Vietnamese scene text with tone marks accurately, only P2 feature layer was utilized to segment the text area, and the mask matrix size of the text area was adjusted from 14 × 14 to 14 × 28 to adapt the shape of most texts. Aiming at the problem that duplicate text detection boxes cannot be eliminated by the conventional Non-Maximum Suppression (NMS) algorithm, a filter module for the text areas named Text region filtering branch was designed and added after the detection module to effectively eliminate duplicate detection boxes. A model joint training method was used to train the network. The training process consists of two parts: the first part is the training of the Feature Pyramid Network (FPN) and the Region Proposal Network (RPN) of the model, which used large-scale open Latin text data for training to enhance the generalization ability of the model to detect text in different scenes; the second part is the training of the candidate box coordinate regression module and the segmentation module named Box branch and Mask branch, which used pixel-level labelled Vietnamese scene text data for training to enable the model to segment the Vietnamese text area including tone marks. Many cross-validation experiments and comparison experiments verify that the proposed algorithm has better precision and recall under different Intersection over Union (IoU) thresholds compared with Mask R-CNN.

    Table and Figures | Reference | Related Articles | Metrics
    Robot path planning based on B-spline curve and ant colony algorithm
    Erchao LI, Kuankuan QI
    Journal of Computer Applications    2021, 41 (12): 3558-3564.   DOI: 10.11772/j.issn.1001-9081.2021060888
    Abstract308)   HTML19)    PDF (1368KB)(109)       Save

    In view of the problems of ant colony algorithm in global path planning under static environment, such as being unable to find the shortest path, slow convergence speed, great blindness of path search and many inflection points, an improved ant colony algorithm was proposed. Taking the grid map as the running environment of the robot, the initial pheromones were distributed unevenly, so that the path search tended to be near the line between the starting point and the target point; the information of the current node, the next node and the target point was added into the heuristic function, and the dynamic adjustment factor was introduced at the same time, so as to achieve the purpose of strong guidance of the heuristic function in the early stage and strengthening the guidance of pheromone in the later stage; the pseudo-random transfer strategy was introduced to reduce the blindness of path selection and speed up finding the shortest path; the volatilization coefficient was adjusted dynamically to make the volatilization coefficient larger in the early stage and smaller in the later stage, avoiding premature convergence of the algorithm; based on the optimal solution, B-spline curve smoothing strategy was introduced to further optimize the optimal solution, resulting in shorter and smoother path. The sensitivity analysis of the main parameters of the improved algorithm was conducted, the feasibility and effectiveness of each improved step of the algorithm were tested, the simulations compared with the traditional ant colony algorithm and other improved ant colony algorithms under 20×20 and 50×50 environments were given, and the experimental results verified the feasibility, effectiveness and superiority of the improved algorithm.

    Table and Figures | Reference | Related Articles | Metrics
    Object tracking algorithm based on spatio-temporal context information enhancement
    Jing WEN, Qiang LI
    Journal of Computer Applications    2021, 41 (12): 3565-3570.   DOI: 10.11772/j.issn.1001-9081.2021061034
    Abstract283)   HTML11)    PDF (915KB)(121)       Save

    Making full use of the spatio-temporal context information in the video can significantly improve the performance of object tracking, but most of the current object tracking algorithms based on deep learning only use the feature information of the current frame to locate the object, without using the spatio-temporal context information of the same object in the video frames before and after the current frame, which leads to the tracking object being susceptible to the interference from the similar object nearby, so a potential cumulative error will be introduced during tracking and locating. In order to retain spatio-temporal context information, a short-term memory storage pool was introduced based on SiamMask algorithm to store features of the historical frames; meanwhile, an Appearance Saliency Boosting Module (ASBM) was proposed, which not only enhanced the saliency features of the tracking object, but also suppressed the interference from similar object around the tracking object. On the basis of the above, an object tracking algorithm based on spatio-temporal context information enhancement was proposed. To verify the performance of the proposed algorithm, experiments were carried out on four datasets, including VOT2016, VOT2018, DAVIS-2016 and DAVIS-2017. Experimental results show that compared with SiamMask algorithm, the proposed algorithm has the accuracy and Expected Average Overlap rate (EAO) increased by 4 percentage points and 2 percentage points respectively on VOT2016 dataset, and has the accuracy, robustness and EAO improved by 3.7 percentage points, 2.8 percentage points and 1 percentage point respectively on VOT2018 dataset, and has the decay of the regional similarity and contour accuracy indicators on DAVIS-2016 datasets both reduced by 0.2 percentage points, and has the decay of the regional similarity and contour progress indicators on DAVIS-2017 datasets reduced by 1.3 and 0.9 percentage points respectively.

    Table and Figures | Reference | Related Articles | Metrics
    Unsupervised salient object detection based on graph cut refinement and differentiable clustering
    Xiaoyu LI, Tiyu FANG, Yingjie XIA, Jinping LI
    Journal of Computer Applications    2021, 41 (12): 3571-3577.   DOI: 10.11772/j.issn.1001-9081.2021061054
    Abstract396)   HTML12)    PDF (1317KB)(130)       Save

    Concerning that the traditional saliency detection algorithm has low segmentation accuracy and the deep learning-based saliency detection algorithm has strong dependence on pixel-level manual annotation data, an unsupervised saliency object detection algorithm based on graph cut refinement and differentiable clustering was proposed. In the algorithm, the idea of “coarse” to “fine” was adopted to achieve accurate salient object detection by only using the characteristics of a single image. Firstly, Frequency-tuned algorithm was used to obtain the salient coarse image according to the color and brightness of the image itself. Then, the candidate regions of the salient object were obtained by binarization according to the image’s statistical characteristics and combination of the central priority hypothesis. After that, GrabCut algorithm based on single image for graph cut was used for segmenting the salient object finely. Finally, in order to overcome the difficulty of imprecise detection when the background was very similar to the object, the unsupervised differentiable clustering algorithm with good boundary segmentation effect was introduced to further optimize the saliency map. Experimental results show that compared with the existing seven algorithms, the optimized saliency map obtained by the proposed algorithm is closer to the ground truth, achieving an Mean Absolute Error (MAE) of 14.3% and 23.4% on ECSSD and SOD datasets, respectively.

    Table and Figures | Reference | Related Articles | Metrics
    Spatio-temporal hyper-relationship graph convolutional network for traffic flow forecasting
    Yongkai ZHANG, Zhihao WU, Youfang LIN, Yiji ZHAO
    Journal of Computer Applications    2021, 41 (12): 3578-3584.   DOI: 10.11772/j.issn.1001-9081.2021060956
    Abstract535)   HTML18)    PDF (1112KB)(186)       Save

    Traffic flow forecasting is an important research topic for the intelligent transportation system, however, this research is very challenging because of the complex local spatio-temporal relationships among traffic objects such as stations and sensors. Although some previous studies have made great progress by transforming the traffic flow forecasting problem into a spatio-temporal graph forecasting problem, in which the direct correlations across spatio-temporal dimensions among traffic objects are ignored. At present, there is still lack of a comprehensive modeling approach for the local spatio-temporal relationships. A novel spatio-temporal hypergraph modeling scheme was first proposed to address this problem by constructing a kind of spatio-temporal hyper-relationships to comprehensively model the complex local spatio-temporal relationships. Then, a Spatio-Temporal Hyper-Relationship Graph Convolutional Network (STHGCN) forecasting model was proposed to capture these relationships for traffic flow forecasting. Extensive comparative experiments were conducted on four public traffic datasets. Experimental results show that compared with the spatio-temporal forecasting models such as Attention based Spatial-Temporal Graph Convolutional Network (ASTGCN) and Spatial-Temporal Synchronous Graph Convolutional Network (STSGCN), STHGCN achieves better results in Root Mean Square Error (RMSE), Mean Absolute Error (MAE) and Mean Absolute Percentage Error (MAPE); and the comparison of the running time of different models also shows that STHGCN has higher inference speed.

    Table and Figures | Reference | Related Articles | Metrics
    Multi-track music generative adversarial network based on Transformer
    Tao WANG, Cong JIN, Xiaobing LI, Yun TIE, Lin QI
    Journal of Computer Applications    2021, 41 (12): 3585-3589.   DOI: 10.11772/j.issn.1001-9081.2021060909
    Abstract777)   HTML20)    PDF (639KB)(324)       Save

    Symbolic music generation is still an unsolved problem in the field of artificial intelligence and faces many challenges. It has been found that the existing methods for generating polyphonic music fail to meet the marke requirements in terms of melody, rhythm and harmony, and most of the generated music does not conform to basic music theory knowledge. In order to solve the above problems, a new Transformer-based multi-track music Generative Adversarial Network (Transformer-GAN) was proposed to generate music with high musicality under the guidance of music rules. Firstly, the decoding part of Transformer and the Cross-Track Transformer (CT-Transformer) adapted on the basis of Transformer were used to learn the information within a single track and between multiple tracks respectively. Then, a combination of music rules and cross-entropy loss was employed to guide the training of the generative network, and the well-designed objective loss function was optimized while training the discriminative network. Finally, multi-track music works with melody, rhythm and harmony were generated. Experimental results show that compared with other multi-instrument music generation models, for piano, guitar and bass tracks, Transformer-GAN improves Prediction Accuracy (PA) by a minimum of 12%, 11% and 22%, improves Sequence Similarity (SS) by a minimum of 13%, 6% and 10%, and improves the rest index by a minimum of 8%, 4% and 17%. It can be seen that Transformer -GAN can effectively improve the indicators including PA and SS of music after adding CT-Transformer and music rule reward module, which leads to a relatively high overall improvement of the generated music.

    Table and Figures | Reference | Related Articles | Metrics
    Person re-identification method based on grayscale feature enhancement
    Yunpeng GONG, Zhiyong ZENG, Feng YE
    Journal of Computer Applications    2021, 41 (12): 3590-3595.   DOI: 10.11772/j.issn.1001-9081.2021061011
    Abstract272)   HTML12)    PDF (932KB)(140)       Save

    Whether the learned features have better invariance in the significant intra-class changes will determine the upper limit of performance of the Person Re-identification (ReID) model. Environmental light, image resolution change, motion blur and other factors may cause color deviation of pedestrian images, and these problems will cause overfitting of the model to color information of the data, thus limiting the performance of the model. By simulating the color information loss of the data samples and highlighting the structural information of the samples, the model was helped to learn more robust features. Specifically, during model training, the training batch was randomly selected according to the set probability, and then a rectangular area of the image or the entire image was randomly selected for each RGB image sample in the selected batch, and the pixels of the selected area was replaced with the pixels of the same rectangular area in the corresponding grayscale image, thus generating a training image with different grayscale areas. Experimental results demonstrate that compared with the benchmark model, the proposed method achieves a significant performance improvement of 3.3 percentage points at most on the evaluation index mean Average Precision (mAP), and performs well on multiple datasets.

    Table and Figures | Reference | Related Articles | Metrics
    Cloth-changing person re-identification based on joint loss capsule network
    Qian LIU, Hongyuan WANG, Liang CAO, Boyan SUN, Yu XIAO, Ji ZHANG
    Journal of Computer Applications    2021, 41 (12): 3596-3601.   DOI: 10.11772/j.issn.1001-9081.2021061090
    Abstract307)   HTML14)    PDF (610KB)(145)       Save

    Current research on Person Re-Identification (Re-ID) mainly concentrates on short-term situations with person’s clothing usually unchanged. However, more common practical cases are long-term situations, in which a person has higher possibility to change his clothes, which should be considered by Re-ID models. Therefore, a method of person re-identification with cloth changing based on joint loss capsule network was proposed. The proposed method was based on ReIDCaps, a capsule network for cloth-changing person re-identification. In the method, vector-neuron capsules that contain more information than traditional scalar neurons were used. The length of the vector-neuron capsule was used to represent the identity information of the person, and the direction of the capsule was used to represent the clothing information of the person. Soft Embedding Attention (SEA) was used to avoid the model over-fitting. Feature Sparse Representation (FSR) mechanism was adopted to extract discriminative features. The joint loss of label smoothing regularization cross-entropy loss and Circle Loss was added to improve the generalization ability and robustness of the model. Experimental results on three datasets including Celeb-reID, Celeb-reID-light and NKUP prove that the proposed method has certain advantages compared with the existing person re-identification methods.

    Table and Figures | Reference | Related Articles | Metrics
    Smoking behavior detection algorithm based on human skeleton key points
    Wanqing XU, Baodong WANG, Yimei HUANG, Jinping LI
    Journal of Computer Applications    2021, 41 (12): 3602-3607.   DOI: 10.11772/j.issn.1001-9081.2021061063
    Abstract521)   HTML14)    PDF (1345KB)(269)       Save

    In view of the small target of cigarette butts in surveillance videos of public places and the easy divergence of smoke generated by smoking, it is difficult to determine the smoking behavior only by target detection algorithm. Considering that the algorithm of posture estimation using skeleton key points is becoming more and more mature, a smoking behavior detection algorithm was proposed by using the relationship between human skeleton key points and smoking behavior. Firstly, AlphaPose and RetinaFace were used to detect the key points of human skeleton and face respectively. According to the ratio of distance between wrist and middle point of two corners of mouth and between wrist and the eye on the same side, a method for calculating whether the Smoking Action Ratio (SAR) in humans falls within the Golden Ratio of Smoking Actions (GRSA) to distinguish smoking from non-smoking behaviors was proposed. Then, YOLOv4 was used to detect whether cigarette butts existed in the video. The results of GRSA determination and YOLOv4 were combined to determine the possibility of smoking behavior in the video and make a determination of whether smoking behavior was present. The self-recorded dataset test shows that the proposed algorithm can accurately detect smoking behavior with the accuracy reached 92%.

    Table and Figures | Reference | Related Articles | Metrics
    Prediction method of liver transplantation complications based on transfer component analysis and support vector machine
    Hongliang CAO, Ying ZHANG, Bin WU, Fanyu LI, Xubo NA
    Journal of Computer Applications    2021, 41 (12): 3608-3613.   DOI: 10.11772/j.issn.1001-9081.2021060886
    Abstract244)   HTML5)    PDF (699KB)(68)       Save

    Many machine learning algorithms can cope well with prediction and classification, but these methods suffer from poor prediction accuracy and F1 score when they are used on medical datasets with small samples and large feature spaces. To improve the accuracy and F1 score of liver transplantation complication prediction, a prediction and classification method of liver transplantation complications based on Transfer Component Analysis (TCA) and Support Vector Machine (SVM) was proposed. In this method, TCA was used for mapping and dimension reduction of the feature space, and the source domain and the target domain were mapped to the same reproducing kernel Hilbert space, thereby achieving the adaptivity of edge distribution. The SVM was trained in the source domain after transferring, and the complications were predicted in the target domain after training. In the liver transplantation complication prediction experiments for complication Ⅰ, complication Ⅱ, complication Ⅲa, complication Ⅲb, and complication Ⅳ, compared with the traditional machine learning and Heterogeneous Domain Adaptation (HDA), the accuracy of the proposed method was improved by 7.8% to 42.8%, and the F1 score reached 85.0% to 99.0%, while the traditional machine learning and HDA had high accuracy but low recall due to the imbalance of positive and negative samples. Experimental results show that TCA combined with SVM can effectively improve the accuracy and F1 score of liver transplantation complication prediction.

    Table and Figures | Reference | Related Articles | Metrics
    Prediction model of lncRNA-encoded short peptides based on representation learning and deep forest
    Tengqi JI, Jun MENG, Siyuan ZHAO, Hehuan HU
    Journal of Computer Applications    2021, 41 (12): 3614-3619.   DOI: 10.11772/j.issn.1001-9081.2021061082
    Abstract249)   HTML14)    PDF (891KB)(95)       Save

    Small Open Reading Frames (sORFs) in long non-coding RNA (lncRNA) can encode short peptides with length no more than 100 amino acids. Aiming at the problem that the features of sORFs in lncRNA are not distinct and the data with high reliability are not enough in short peptide prediction research, a Deep Forest (DF) model based on representation learning was proposed. Firstly, the conventional lncRNA feature extraction method was used to encode the sORFs. Secondly, the AutoEncoder (AE) was used to perform representation learning to obtain highly efficient representation of the input data. Finally, a DF model was trained to predict the short peptides encoded by lncRNA. Experimental results show that the accuracy of this model can achieve 92.08% on Arabidopsis thaliana dataset, which is higher than those of the traditional machine learning models , deep learning models and combined models, and this model has better stability. In addition, the prediction accuracy of this method can reach 78.16% and 74.92% on Glycine max and Zea mays datasets respectively, verifying the good generalization ability of the proposed model.

    Table and Figures | Reference | Related Articles | Metrics
    IIoT hidden anomaly detection based on locality sensitive Bloom filter
    Ruliang XIAO, Zhixia ZENG, Chenkai XIAO, Shi ZHANG
    Journal of Computer Applications    2021, 41 (12): 3620-3625.   DOI: 10.11772/j.issn.1001-9081.2021061115
    Abstract266)   HTML9)    PDF (580KB)(87)       Save

    Damage to sensors in Industrial Internet of Things (IIoT) system due to continuous use and normal wear leads to hidden anomalies in the collected and recorded sensing data. To solve this problem, an anomaly detection algorithm based on Local Sensitive Bloom Filter (LSBF) model was proposed, namely LSBFAD. Firstly, the Spatial Partition based Fast Johnson-Lindenstrauss Transform (SP-FJLT) was used to perform hash mapping to the data, then the Mutual Competition (MC) strategy was used to reduce noise, and finally the Bloom filter was constructed by 0-1 coding. In simulation experiments conducted on three benchmark datasets including SIFT, MNIST and FMA, the false detection rate of LSBFAD algorithm is less than 10%. Experimental results show that compared with the current mainstream anomaly detection algorithms, the proposed anomaly detection algorithm based on LSBF has higher Detection Rate (DR) and lower False Alarm Rate (FAR) and can be effectively applied to anomaly detection of IIoT data.

    Table and Figures | Reference | Related Articles | Metrics
    Transfer learning based on graph convolutional network in bearing service fault diagnosis
    Xueying PENG, Yongquan JIANG, Yan YANG
    Journal of Computer Applications    2021, 41 (12): 3626-3631.   DOI: 10.11772/j.issn.1001-9081.2021060974
    Abstract351)   HTML8)    PDF (561KB)(260)       Save

    Deep learning methods are widely used in bearing fault diagnosis, but in actual engineering applications, real service fault data during bearing service are not easily collected and lack of data labels, which is difficult to train adequately. Focused on the difficulty of bearing service fault diagnosis, a transfer learning model based on Graph Convolutional Network (GCN) in bearing service fault diagnosis was proposed. In the model, the fault knowledge was learned from artificially simulated damage fault data with sufficient data and transferred to real service faults, so as to improve the diagnostic accuracy of service faults. Specifically, the original vibration signals of artificially simulated damage fault data and service fault data were converted into the time-frequency maps with both time and frequency information through wavelet transform, and the obtained maps were input into graph convolutional layers for learning, so as to effectively extract the fault feature representations in the source and target domains. Then the Wasserstein distance between the data distributions of source domain and target domain was calculated to measure the difference between two data distributions, and a fault diagnosis model that can diagnose bearing service faults was constructed by minimizing the difference in data distribution. A variety of different tasks were designed for experiments with different bearing failure data sets and different operating conditions. Experimental results show that the proposed model has the ability to diagnose bearing service faults and also can be transferred from one working condition to another, and perform fault diagnosis between different component types and different working conditions.

    Table and Figures | Reference | Related Articles | Metrics
    Stock index forecasting method based on corporate financial statement data
    Jihou WANG, Peiguang LIN, Jiaqian ZHOU, Qingtao LI, Yan ZHANG, Muwei JIAN
    Journal of Computer Applications    2021, 41 (12): 3632-3636.   DOI: 10.11772/j.issn.1001-9081.2021061006
    Abstract346)   HTML7)    PDF (580KB)(113)       Save

    All market activities of stock market participants combine to affect stock market changes, making stock market volatility fraught with complexity and making accurate prediction of stock prices a challenge. Among these activities that affect stock market changes, financial disclosure is an attractive and potentially financially rewarding means of predicting stock indexe changes. In order to deal with the complex changes in the stock market, a method of stock index prediction was proposed that incorporates data from financial statements disclosed by corporates. Firstly, the stock index historical data and corporate financial statement data were preprocessed, and the main task is dimension reduction of the high-dimensional matrix generated from corporate financial statement data, and then the dual-channel Long Short-Term Memory (LSTM) network was used to forecast and research the normalized data. Experimental results on SSE 50 and CSI 300 Index datasets show that the prediction effect of the proposed method is better than that using only historical data of stock indexes.

    Table and Figures | Reference | Related Articles | Metrics
    Constrained multi-objective evolutionary algorithm based on space shrinking technique
    Erchao LI, Yuyan MAO
    Journal of Computer Applications    2021, 41 (12): 3419-3425.   DOI: 10.11772/j.issn.1001-9081.2021060887
    Abstract316)   HTML28)    PDF (979KB)(145)       Save

    The reasonable exploration of the infeasible region in constrained multi-objective evolutionary algorithms for solving optimization problems with large infeasible domains not only helps the population to converge quickly to the optimal solution in the feasible region, but also reduces the impact of unpromising infeasible region on the performance of the algorithm. Based on this, a Constrained Multi-Objective Evolutionary Algorithm based on Space Shrinking Technique (CMOEA-SST) was proposed. Firstly, an adaptive elite retention strategy was proposed to improve the initial population in the Pull phase of Push and Pull Search for solving constrained multi-objective optimization problems (PPS), so as to increase the diversity and feasibility of the initial population in the Pull phase. Then, the space shrinking technique was used to gradually reduce the search space during the evolution process, which reduced the impact of unpromising infeasible regions on the algorithm performance. Therefore, the algorithm was able to improve the convergence accuracy while taking account of both convergence and diversity. In order to verify the performance of the proposed algorithm, it was simulated and compared with four representative algorithms including C-MOEA/D (adaptive Constraint handling approach embedded MOEA/D), ToP (handling constrained multi-objective optimization problems with constraints in both the decision and objective spaces), C-TAEA (Two-Archive Evolutionary Algorithm for Constrained multi-objective optimization) and PPS on the test problems of LIRCMOP series. Experimental results show that CMOEA-SST has better convergence and diversity when dealing with constrained optimization problems with large infeasible regions.

    Table and Figures | Reference | Related Articles | Metrics
    Dynamic graph representation learning method based on deep neural network and gated recurrent unit
    Huibo LI, Yunxiao ZHAO, Liang BAI
    Journal of Computer Applications    2021, 41 (12): 3432-3437.   DOI: 10.11772/j.issn.1001-9081.2021060994
    Abstract316)   HTML15)    PDF (869KB)(126)       Save

    Learning the latent vector representations of nodes in the graph is an important and ubiquitous task, which aims to capture various attributes of the nodes in the graph. A lot of work demonstrates that static graph representation learning can learn part of the node information; however, real-world graphs evolve over time. In order to solve the problem that most dynamic network algorithms cannot effectively retain node neighborhood structure and temporal information, a dynamic network representation learning method based on Deep Neural Network (DNN) and Gated Recurrent Unit (GRU), namely DynAEGRU, was proposed. With Auto-Encoder (AE) as the framework of the DynAEGRU, the neighborhood information was aggregated by encoder with a DNN to obtain low-dimensional feature vectors, then the node temporal information was extracted by a GRU network,finally, the adjacency matrix was reconstructed by the decoder and compared with the real graph to construct the loss. Experimental results on three real-word datasets show that DynAEGRU method has better performance gain compared to several static and dynamic graph representation learning algorithms.

    Table and Figures | Reference | Related Articles | Metrics
    Robust multi-view subspace clustering based on consistency graph learning
    Zhenjun PAN, Cheng LIANG, Huaxiang ZHANG
    Journal of Computer Applications    2021, 41 (12): 3438-3446.   DOI: 10.11772/j.issn.1001-9081.2021061056
    Abstract359)   HTML14)    PDF (781KB)(139)       Save

    Concerning that the multi-view data analysis is susceptible to the noise of the original dataset and requires additional steps to calculate the clustering results, a Robust Multi-view subspace clustering based on Consistency Graph Learning (RMCGL) algorithm was proposed. Firstly, the potential robust representation of data in the subspace was learned in each view, and the similarity matrix of each view was obtained based on these representations. Then, a unified similarity graph was learned based on the obtained multiple similarity matrices. Finally, by adding rank constraints to the Laplacian matrix corresponding to the similarity graph, the obtained similarity graph had the optimal clustering structure, and the final clustering results were able to be obtained directly by using this similarity graph. The process was completed in a unified optimization framework, in which potential robust representations, similarity matrices and consistency graphs could be learned simultaneously. The clustering Accuracy (ACC) of RMCGL algorithm is 3.36 percentage points, 5.82 percentage points and 5.71 percentage points higher than that of Graph-based Multi-view Clustering (GMC) algorithm on BBC, 100leaves and MSRC datasets, respectively. Experimental results show that the proposed algorithm has a good clustering effect.

    Table and Figures | Reference | Related Articles | Metrics
    Directed graph clustering algorithm based on kernel nonnegative matrix factorization
    Xian CHEN, Liying HU, Xiaowei LIN, Lifei CHEN
    Journal of Computer Applications    2021, 41 (12): 3447-3454.   DOI: 10.11772/j.issn.1001-9081.2021061129
    Abstract312)   HTML9)    PDF (653KB)(88)       Save

    Most of the existing directed graph clustering algorithms are based on the assumption of approximate linear relationship between nodes in vector space, ignoring the existence of non-linear correlation between nodes. To address this problem, a directed graph clustering algorithm based on Kernel Nonnegative Matrix Factorization (KNMF) was proposed. First, the adjacency matrix of a directed graph was projected to the kernel space by using a kernel learning method, and the node similarity in both the original and kernel spaces was constrained by a specific regularization term. Second, the objective function of graph regularization kernel asymmetric NMF algorithm was proposed, and a clustering algorithm was derived by gradient descent method under the non-negative constraints. The algorithm accurately reveals the potential structural information in the directed graph by modeling the non-linear relationship between nodes using kernel learning method, as well as considering the directivity of the links of nodes. Finally, experimental results on the Patent Citation Network (PCN) dataset show that compared with the comparison algorithm, when the number of clusters is 2, the proposed algorithm improves the Davies-Bouldin (DB) and Distance-based Quality Function (DQF) by about 0.25 and 8% respectively, achieving better clustering quality.

    Table and Figures | Reference | Related Articles | Metrics
    Graph learning regularized discriminative non-negative matrix factorization based face recognition
    Han DU, Xianzhong LONG, Yun LI
    Journal of Computer Applications    2021, 41 (12): 3455-3461.   DOI: 10.11772/j.issn.1001-9081.2021060979
    Abstract365)   HTML16)    PDF (790KB)(131)       Save

    The Non-negative Matrix Factorization (NMF) algorithm based on graph regularization makes full use of the assumption that high-dimensional data are usually located in a low-dimensional manifold space to construct the Laplacian matrix. The disadvantage of this algorithm is that the constructed Laplacian matrix is calculated in advance and will not be iterated during the multiplicative update process. In order to solve this problem, the self-representation method in subspace learning was combined to generate the representation coefficient, and the similarity matrix was further calculated to obtain the Laplacian matrix, and the Laplacian matrix was iterated during the update process. In addition, the label information of the training set was used to construct the class indicator matrix, and two different regularization items were introduced to reconstruct the category indicator matrix respectively. This algorithm was called Graph Learning Regularized Discriminative Non-negative Matrix Factorization (GLDNMF), and the corresponding multiplicative update rules and the convergence proof of the objective function were given. Face recognition experimental results on two standard datasets show that the accuracy of the proposed algorithm for face recognition is increased by 1% - 5% compared to the existing classic algorithms, verifying the effectiveness of the proposed method.

    Table and Figures | Reference | Related Articles | Metrics
    Multi-kernel learning method based on neural tangent kernel
    Mei WANG, Chuanhai XU, Yong LIU
    Journal of Computer Applications    2021, 41 (12): 3462-3467.   DOI: 10.11772/j.issn.1001-9081.2021060998
    Abstract272)   HTML16)    PDF (510KB)(88)       Save

    Multi-kernel learning method is an important type of kernel learning method, but most of multi-kernel learning methods have the following problems: most of the basis kernel functions in multi-kernel learning methods are traditional kernel functions with shallow structure, which have weak representation ability when dealing with the problems of large data scale and uneven distribution; the generalization error convergence rates of the existing multi-kernel learning methods are mostly O 1 / n , and the convergence speeds are slow. Therefore, a multi-kernel learning method based on Neural Tangent Kernel (NTK) was proposed. Firstly, the NTK with deep structure was used as the basis kernel function of the multi-kernel learning method, so as to enhance the representation ability of the multi-kernel learning method. Then, a generalization error bound with a convergence rate of O 1 / n was proved based on the measure of principal eigenvalue ratio. On this basis, a new multi-kernel learning algorithm was designed in combination with the kernel alignment measure. Finally, experiments were carried out on several datasets. Experimental results show that compared with classification algorithms such as Adaboost and K-Nearest Neighbor (KNN), the newly proposed multi-kernel learning algorithm has higher accuracy and better representation ability, which also verifies the feasibility and effectiveness of the proposed method.

    Table and Figures | Reference | Related Articles | Metrics
    Multiple kernel clustering algorithm based on capped simplex projection graph tensor learning
    Haoyun LEI, Zenwen REN, Yanlong WANG, Shuang XUE, Haoran LI
    Journal of Computer Applications    2021, 41 (12): 3468-3474.   DOI: 10.11772/j.issn.1001-9081.2021061393
    Abstract369)   HTML7)    PDF (6316KB)(125)       Save

    Because multiple kernel learning can avoid selection of kernel functions and parameters effectively, and graph clustering can fully mine complex structural information between samples, Multiple Kernel Graph Clustering (MKGC) has received widespread attention in recent years. However, the existing MKGC methods suffer from the following problems: graph learning technique complicates the model, the high rank of graph Laplacian matrix cannot ensure the learned affinity graph to contain accurate c connected components (block diagonal property), and most of the methods ignore the high-order structural information among the candidate affinity graphs, making it difficult to fully utilize the multiple kernel information. To tackle these problems, a novel MKGC method was proposed. First, a new graph learning method based on capped simplex projection was proposed to directly project the kernel matrices onto graph simplex, which reduced the computational complexity. Meanwhile, a new block diagonal constraint was introduced to keep the accurate block diagonal property of the learned affinity graphs. Moreover, the low-rank tensor learning was introduced in capped simplex projection space to fully mine the high-order structural information of multiple candidate affinity graphs. Compared with the existing MKGC methods on multiple datasets, the proposed method has less computational cost and high stability, and has great advantages in Accuracy (ACC) and Normalized Mutual Information (NMI).

    Table and Figures | Reference | Related Articles | Metrics
2024 Vol.44 No.4

Current Issue
Archive
Honorary Editor-in-Chief: ZHANG Jingzhong
Editor-in-Chief: XU Zongben
Associate Editor: SHEN Hengtao XIA Zhaohui
Domestic Post Distribution Code: 62-110
Foreign Distribution Code: M4616
Address:
No. 9, 4th Section of South Renmin Road, Chengdu 610041, China
Tel: 028-85224283-803
  028-85222239-803
Website: www.joca.cn
E-mail: bjb@joca.cn
WeChat
Join CCF