Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Data augmentation technique incorporating label confusion for Chinese text classification
Haitao SUN, Jiayu LIN, Zuhong LIANG, Jie GUO
Journal of Computer Applications    2025, 45 (4): 1113-1119.   DOI: 10.11772/j.issn.1001-9081.2024040550
Abstract42)   HTML0)    PDF (863KB)(27)       Save

Traditional data augmentation techniques, such as synonym substitution, random insertion, and random deletion, may change the original semantics of text and even result in the loss of critical information. Moreover, data in text classification tasks typically have both textual and label parts. However, traditional data augmentation methods only focus on the textual part. To address these issues, a Label Confusion incorporated Data Augmentation (LCDA) technique was proposed for providing a comprehensive enhancement of data from both textual and label aspects. In terms of text, by enhancing the text through random insertion and replacement of punctuation marks and completing end-of-sentence punctuation marks, textual diversity was increased with all textual information and sequence preserved. In terms of labels, simulated label distribution was generated using a label confusion approach, and used to replace the traditional one-hot label distribution, so as to better reflect the relationships among instances and labels as well as between labels. In experiments conducted on few-shot datasets constructed from THUCNews (TsingHua University Chinese News) and Toutiao Chinese news datasets, the proposed technique was combined with TextCNN, TextRNN, BERT (Bidirectional Encoder Representations from Transformers), and RoBERTa-CNN (Robustly optimized BERT approach Convolutional Neural Network) text classification models. The experimental results indicate that compared to those before enhancement, all models demonstrate significant performance improvements. Specifically, on 50-THU, a dataset constructed on THUCNews dataset, the accuracies of four models combing LCDA technique are improved by 1.19, 6.87, 3.21, and 2.89 percentage points, respectively, compared to those before enhancement, and by 0.78, 7.62, 1.75, and 1.28 percentage points, respectively, compared to those of the four models combining softEDA (Easy Data Augmentation with soft labels) method. By both textual and label processing results, model accuracy is enhanced by LCDA technique significantly, particularly in application scenarios characterized by limited data availability.

Table and Figures | Reference | Related Articles | Metrics
Lightweight human pose estimation based on decoupled attention and ghost convolution
Junying CHEN, Shijie GUO, Lingling CHEN
Journal of Computer Applications    2025, 45 (1): 223-233.   DOI: 10.11772/j.issn.1001-9081.2024010099
Abstract137)   HTML3)    PDF (3442KB)(702)       Save

With the development of lightweight networks, human pose estimation tasks can be performed on devices with limited computational resources. However, improving accuracy has become more challenging. These challenges mainly led by the contradiction between network complexity and computational resources, resulting in the sacrifice of representation capabilities when simplifying the model. To address these issues, a Decoupled attention and Ghost convolution based Lightweight human pose estimation Network (DGLNet) was proposed. Specifically, in DGLNet, with Small High-Resolution Network (Small HRNet) model as basic architecture, by introducing a decoupled attention mechanism, DFDbottleneck module was constructed. The basic modules were redesigned with shuffleblock structure, in which computationally-intensive point convolutions were replaced with lightweight ghost convolutions, and the decoupled attention mechanism was utilized to enhance module performance, leading to the creation of DGBblock module. Additionally, the original transition layer modules were replaced with redesigned depthwise separable convolution modules that incorporated ghost convolution and decoupled attention, resulting in the construction of GSCtransition module. This modification further reduced computational complexity while enhancing feature interaction and performance. Experimental results on COCO validation set show that DGLNet outperforms the state-of-the-art Lite-High-Resolution Network (Lite-HRNet) model, achieving the maximum accuracy of 71.9% without increasing computational complexity or the number of parameters. Compared to common lightweight pose estimation networks such as MobileNetV2 and ShuffleNetV2, DGLNet achieves the precision improvement of 4.6 and 8.3 percentage points respectively, while only utilizing 21.2% and 25.0% of their computational resources. Furthermore, under the AP50 evaluation criterion, DGLNet surpasses the large High-Resolution Network (HRNet) while having significantly less computational and parameters.

Table and Figures | Reference | Related Articles | Metrics
Recommendation method based on knowledge‑awareness and cross-level contrastive learning
Jie GUO, Jiayu LIN, Zuhong LIANG, Xiaobo LUO, Haitao SUN
Journal of Computer Applications    2024, 44 (4): 1121-1127.   DOI: 10.11772/j.issn.1001-9081.2023050613
Abstract233)   HTML15)    PDF (968KB)(187)       Save

As a kind of side information, Knowledge Graph (KG) can effectively improve the recommendation quality of recommendation models, but the existing knowledge-awareness recommendation methods based on Graph Neural Network (GNN) suffer from unbalanced utilization of node information. To address the above problem, a new recommendation method based on Knowledge?awareness and Cross-level Contrastive Learning (KCCL) was proposed. To alleviate the problem of unbalanced node information utilization caused by the sparse interaction data and noisy knowledge graph that deviate from the true representation of inter-node dependencies during information aggregation, a contrastive learning paradigm was introduced into knowledge-awareness recommendation model of GNN. Firstly, the user-item interaction graph and the item knowledge graph were integrated into a heterogeneous graph, and the node representations of users and items were realized by a GNN based on the graph attention mechanism. Secondly, consistent noise was added to the information propagation aggregation layer for data augmentation to obtain node representations of different levels, and the obtained outermost node representation was compared with the innermost node representation for cross-level contrastive learning. Finally, the supervised recommendation task and the contrastive learning assistance task were jointly optimized to obtain the final representation of each node. Experimental results on DBbook2014 and MovieLens-1m datasets show that compared to the second prior contrastive method, the Recall@10 of KCCL is improved by 3.66% and 0.66%, respectively, and the NDCG@10 is improved by 3.57% and 3.29%, respectively, which verifies the effectiveness of KCCL.

Table and Figures | Reference | Related Articles | Metrics
Location control method for generated objects by diffusion model with exciting and pooling attention
Jinsong XU, Ming ZHU, Zhiqiang LI, Shijie GUO
Journal of Computer Applications    2024, 44 (4): 1093-1098.   DOI: 10.11772/j.issn.1001-9081.2023050634
Abstract221)   HTML14)    PDF (2886KB)(113)       Save

Due to the ambiguity of text and the lack of location information in training data, current state-of-the-art diffusion model cannot accurately control the locations of generated objects in the image under the condition of text prompts. To address this issue, a spatial condition of the object’s location range was introduced, and an attention-guided method was proposed based on the strong correlation between the cross-attention map in U-Net and the image spatial layout to control the generation of the attention map, thus controlling the locations of the generated objects. Specifically, based on the Stable Diffusion (SD) model, in the early stage of the generation of the cross-attention map in the U-Net layer, a loss was introduced to stimulate high attention values in the corresponding location range, and reduce the average attention value outside the range. The noise vector in the latent space was optimized step by step in each denoising step to control the generation of the attention map. Experimental results show that the proposed method can significantly control the locations of one or more objects in the generated image, and when generating multiple objects, it can reduce the phenomenon of object omission, redundant object generation, and object fusion.

Table and Figures | Reference | Related Articles | Metrics
Semi-supervised fake job advertisement detection model based on consistency training
Ruiqi WANG, Shujuan JI, Ning CAO, Yajie GUO
Journal of Computer Applications    2023, 43 (9): 2932-2939.   DOI: 10.11772/j.issn.1001-9081.2022081163
Abstract214)   HTML15)    PDF (2191KB)(248)       Save

The flood of fake job advertisements will not only damage the legitimate rights and interests of job seekers but also disrupt the normal employment order, which results in a poor user experience for job seekers. To effectively detect fake job advertisements, an SSC (Semi-Supervised fake job advertisements detection model based on Consistency training) was proposed. Firstly, the consistency regularization term was applied on all the data to improve the performance of the model. Then, supervised loss and unsupervised loss were integrated through joint training to obtain the semi-supervised loss. Finally, the semi-supervised loss was used to optimize the model. Experimental results on two real datasets EMSCAD (EMployment SCam Aegean Dataset) and IMDB (Internet Movie DataBase) show that SSC achieves the best detection performance when the labeled data are only 20, and the accuracy is increased by 2.2 and 2.8 percentage points compared with the existing advanced semi-supervised learning model UDA (Unsupervised Data Augmentation), and is increased by 3.4 and 11.7 percentage points compared with the deep learning model BERT (Bidirectional Encoder Representations from Transformers). At the same time, SSC has good scalability.

Table and Figures | Reference | Related Articles | Metrics
Attribute reduction algorithm based on cluster granulation and divergence among clusters
Yan LI, Bin FAN, Jie GUO
Journal of Computer Applications    2022, 42 (9): 2701-2712.   DOI: 10.11772/j.issn.1001-9081.2021081371
Abstract335)   HTML10)    PDF (3592KB)(82)    PDF(mobile) (654KB)(14)    Save

Attribute reduction is a hot research topic in rough set theory. Most of the algorithms of attribute reduction for continuous data are based on dominance relations or neighborhood relations. However, continuous datasets do not necessarily have dominance relations in attributes. And the attribute reduction algorithms based on neighborhood relations can adjust the granulation degree through neighborhood radius, but it is difficult to unify the radii due to the different dimensions of attributes and the continuous values of radius parameters, resulting in high computational cost of the whole parameter granulation process. To solve this problem, a multi-granularity attribute reduction strategy based on cluster granulation was proposed. Firstly, the similar samples were classified by the clustering method, and the concepts of approximate set, relative positive region and positive region reduction based on clustering were proposed. Secondly, according to JS (Jensen-Shannon) divergence theory, the difference of data distribution of each attribute among clusters was measured, and representative features were selected to distinguish different clusters. Finally, an attribute reduction algorithm was designed using a discernibility matrix. In the proposed algorithm, the attributes were not required to have ordered relations. Different from neighborhood radius, the clustering parameter was discrete, and the dataset was able to be divided into different granulation degrees by adjusting this parameter. Experimental results on UCI and Kent Ridge datasets show that this attribute reduction algorithm can directly deal with continuous data. At the same time, by using this algorithm, the redundant features in the datasets can be removed while maintaining or even improving the classification accuracy by discrete adjustment of the parameters in a small range.

Table and Figures | Reference | Related Articles | Metrics
Feature construction and preliminary analysis of uncertainty for meta-learning
Yan LI, Jie GUO, Bin FAN
Journal of Computer Applications    2022, 42 (2): 343-348.   DOI: 10.11772/j.issn.1001-9081.2021071198
Abstract542)   HTML68)    PDF (483KB)(196)       Save

Meta-learning is the learning process of applying machine learning methods (meta-algorithms) to seek the mapping between features of a problem (meta-features) and relative performance measures of the algorithm, thereby forming the learning process of meta-knowledge. How to construct and extract meta-features is an important research content. Concerning the problem that most of meta-features used in the existing related researches are statistical features of data, uncertainty modeling was proposed and the impact of uncertainty on learning system was studied. Based on inconsistency of data, complexity of boundary, uncertainty of model output, linear capability to be classified, degree of attribute overlap, and uncertainty of feature space, six kinds of uncertainty meta-features were established for data or models. At the same time,the uncertainty size of the learning problem itself was measured from different perspectives, and specific definitions were given. The correlations between these meta-features were analyzed on artificial datasets and real datasets of a large number of classification problems, and multiple classification algorithms such as K-Nearest Neighbor (KNN) were used to conduct a preliminary analysis of the correlation between meta-features and test accuracy. Results show that the average degree of correlation is about 0.8, indicating that these meta-features have a significant impact on learning performance.

Table and Figures | Reference | Related Articles | Metrics
Blind road recognition algorithm based on color and texture information
PENG Yuqing XUE Jie GUO Yongfang
Journal of Computer Applications    2014, 34 (12): 3585-3588.  
Abstract449)      PDF (738KB)(676)       Save

Concerning the problem that existing blind road recognition method has low recognition rate, simplistic handling, and is easily influenced by light, or shadow, an improved blind road recognition method was proposed. According to the color and texture features of blind road, the algorithm used two segmentation methods respectively including color histogram feature threshold segmentation combined with improved region growing segmentation and fuzzy C-means clustering segmentation for gray level co-occurrence matrix feature. And combined with Canny edge detection and Hough transform algorithm, the proposed algorithm made the blind area separated from the pedestrian area and determines the migration direction for the blind. The experimental results show that the proposed algorithm can segment several kinds of blind road more accurately, detect the boundary and direction of blind road and solve the light and shadow problem partly. It can choose the fastest and the most effective segmentation method adoptively, and can be used in a variety of devices, such as electronic guide ones.

Reference | Related Articles | Metrics
Estimation of RFID tags based on 0-1 distribution
QIAN Xiaojie GUO Hongyuan TIAN Yangguang
Journal of Computer Applications    2013, 33 (08): 2128-2131.  
Abstract734)      PDF (622KB)(606)       Save
In the large-scale Radio Frequency Identification (RFID) system, the estimated time of current tag estimation algorithms increase linearly with the increase of tags, and the deviation is large. Regarding these problems, a new estimation algorithm based on 0-1 distribution was proposed. By using the feature of 0-1 distribution, the algorithm set the specific frame length and selected flag to choose the collection of tags which responded to the command of query. In this way, estimation time was reduced to the logarithmic level of tag number and the deviation was reduced through picking numerous average values randomly. Compared with other algorithms, the simulation results show that the proposed algorithm drops deviation at least by 0.9%, and has less fluctuation.
Reference | Related Articles | Metrics
Fireworks simulation based on CUDA particle system
CHEN Xiuliang LIANG Yingjie GUO Fuliang
Journal of Computer Applications    2013, 33 (07): 2059-2062.   DOI: 10.11772/j.issn.1001-9081.2013.07.2059
Abstract1269)      PDF (603KB)(588)       Save
The elementary theory of the particle system coincides with the objective laws of the natural world. As a result, the particle system can be used for fireworks and other complex phenomena simulation. To solve the problem that the simulation of particle system is of huge computation and memory resources consumption, the paper built the basic particle system model based on Compute Unified Device Architecture (CUDA) framework. The storage and movement update of particles in the model were considered. Then the parallel KD-TRIE neighbor particle search algorithm based on CUDA was studied. Finally, the detailed implementation of the fireworks simulation was discussed based on the CUDA particle system. The results show that the model can simulate the rise and bloom of the fireworks realistically with the frame rate of up to 320 frames per second, enhancing the fidelity and real-time performance of the simulation.
Reference | Related Articles | Metrics
Survey on emerging pattern based contrast mining and applications
DUAN Lei TANG Chang-jie Guozhu DONG YANG Ning GOU Chi
Journal of Computer Applications    2012, 32 (02): 304-308.   DOI: 10.3724/SP.J.1087.2012.00304
Abstract1400)      PDF (945KB)(674)       Save
Contrast mining is one of fairly new hot data mining topics. Contrast mining focuses on knowledge that describes differences between classes and conditions, or describes changes over time. Contrast mining aims at developing techniques to discover patterns or models that contrast, and characterize multiple datasets associated with different classes or conditions. Contrast mining has wide applications in reality, due to its ability of simplifying problems and classifying accurately. Research on the mining and application of emerging patterns represents a major direction of contrast mining. This paper provided a survey of such issue. More specifically, after introducing the background, basic concepts and principles of emerging patterns, the paper analyzed the mining methods of emerging patterns, discussed extended definitions of emerging patterns and their mining, stated methods for constructing emerging pattern based classifiers, and illustrated applications of emerging pattern in several real-world fields. Finally, this paper gave out some topics for future research on emerging pattern based contrast mining.
Reference | Related Articles | Metrics
Filtering of ground point cloud based on scanning line and self-adaptive angle-limitation algorithm
Jie GUO Jian-yong LIU You-liang ZHANG Yu ZHU
Journal of Computer Applications    2011, 31 (08): 2243-2245.   DOI: 10.3724/SP.J.1087.2011.02243
Abstract1585)      PDF (451KB)(950)       Save
Concerning the filtering problem of trees, buildings or other ground objects in field terrain reverse engineering, the disadvantages of conventional angle-limitation algorithm were analyzed, which accumulated errors or used a single threshold and could not meet the requirement of wavy terrain. Therefore, a self-adaptive angle-limitation algorithm based on scanning line was put forward. This method worked through limiting the angle of scanning center, reference point (known ground point) and the point to be sorted, which was adaptive with the wavy terrain. Then the modified point cloud was optimized with a curve fitting method by moving window. The experimental results prove that, the proposed algorithm has a sound control of the macro-terrain, and it can filter the wavy terrain point cloud much better.
Reference | Related Articles | Metrics