Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
Coordinate enhancement and multi-source sampling for brain tumor image segmentation
Zhanjun JIANG, Yang LI, Jing LIAN, Xinfa MIAO
Journal of Computer Applications    2025, 45 (3): 996-1002.   DOI: 10.11772/j.issn.1001-9081.2024030359
Abstract31)   HTML2)    PDF (2626KB)(14)       Save

To address the issues of insufficient focus on tumor regions and the loss of spatial contextual information in brain tumor image segmentation models, which affect the accuracy of tumor segmentation, a TransUNet-based brain tumor segmentation network integrating Coordinate Enhanced Learning mechanism (CEL) and multi-source sampling was proposed. Firstly, a CEL was proposed, and ResNetv2 was combined as shallow feature extraction network of the model, so as to enhance attention to brain tumor regions. Secondly, a deep blended sampling feature extractor was designed, and deformable attention and self-attention mechanisms were used to perform multi-source sampling on both global and local information of brain tumors. Finally, an Interactive Level Fusion (ILF) module was designed between the encoder and the decoder, thereby realizing interaction between deep and shallow feature information while minimizing parameter computational cost. Experimental results on BraTS2018 and BraTS2019 datasets indicate that compared to the benchmark TransUNet, the proposed model has the mean Dice coefficient (mDice), the mean Intersection over Union (mIoU), the mean Average Precision (mAP) and the mean Recall (mRecall) improved by 4.84, 7.21, 3.83, 3.15 percentage points, respectively, and the model size reduced by 16.9 MB.

Table and Figures | Reference | Related Articles | Metrics
Data tamper-proof batch auditing scheme based on industrial cloud storage systems
Xiaojun ZHANG, Yunpu HAO, Lei LI, Chenyang LI, Ziyu ZHOU
Journal of Computer Applications    2025, 45 (3): 891-895.   DOI: 10.11772/j.issn.1001-9081.2024030349
Abstract41)   HTML1)    PDF (1386KB)(11)       Save

To address the issue of network active attacks such as tampering for industrial cloud storage system data, to achieve the goal of secure sharing of industrial data in cloud storage, and to ensure the confidentiality, integrity, and availability of industrial data transmission and storage processes, a data tamper-proof batch auditing scheme based on industrial cloud storage systems was proposed. In this scheme, a homomorphic digital signature algorithm based on bilinear pairing mapping was proposed, enabling a third-party auditor to achieve batch tamper-proof integrity detection of industrial cloud storage system data, and feedback the tamper-proof integrity auditing results to engineering service end users timely. Besides, the computational burden on engineering service end users was reduced by adding auditors, while ensuring the integrity of industrial encrypted data during transmission and storage processes. Security analysis and performance comparison results demonstrate that the proposed scheme reduces the third-party auditing computational cost significantly by reducing the third-party auditor’s computational cost from On) bilinear pairing operations to O(1) constant-level bilinear pairing operations through the design of tamper-proof detection vectors. It can be seen that the proposed scheme is suitable for lightweight batch auditing scenarios that require tamper-proof detection of a large number of core data files of industrial cloud storage systems.

Table and Figures | Reference | Related Articles | Metrics
Multivariate controllable text generation based on diffusion sequences
Chenyang LI, Long ZHANG, Qiusheng ZHENG, Shaohua QIAN
Journal of Computer Applications    2024, 44 (8): 2414-2420.   DOI: 10.11772/j.issn.1001-9081.2023081137
Abstract168)   HTML3)    PDF (1267KB)(19)       Save

With the emergence of large-scale pre-trained language models, text generation technology has made breakthrough progress. However, in the field of open text generation, the generated content lacks anthropomorphic emotional features, making it difficult for the generated text to resonate and connect emotionally. Controllable text generation is of great significance in compensating for the shortcomings of current text generation technology. Firstly, the extension of theme and emotional attributes was completed on the basis of the ChnSensiCorp dataset. At the same time, in order to construct a multivariate controllable text generation model that could generate smooth text with rich emotion, a diffusion sequence based controllable text generation model DiffuSeq-PT was proposed based on a diffusion model architecture. Theme emotion attributes and text data were used to perform the diffusion process on the sequences without the guidance of classifier. The encoding and decoding capabilities of the pre-trained model ERNIE 3.0(Large-scale Knowledge Enhanced Pre-training for Language Understanding and Generation) were used to fit the noising and denoising process of the diffusion model, and ultimately, target text that matched the relevant theme and multiple sentiment granularities were generated. Compared with the benchmark model DiffuSeq, the proposed model achieved an improvement of 0.13 and 0.01 in BERTScore on two publicly available real datasets (ChnSentiCorp and Debate dataset), and decreased the perplexity by 14.318 and 9.46.

Table and Figures | Reference | Related Articles | Metrics
Reliability enhancement algorithm for physical unclonable function based on non-orthogonal discrete transform
Shiyang LI, Shaojie NI, Ding DENG, Lei CHEN, Honglei LIN
Journal of Computer Applications    2024, 44 (7): 2116-2122.   DOI: 10.11772/j.issn.1001-9081.2023070936
Abstract187)   HTML8)    PDF (2608KB)(76)       Save

A reliability enhancement algorithm for Physical Unclonable Function (PUF) was proposed to address the instability of PUF’s response caused by external and internal factors. The proposed algorithm is based on the Non-Orthogonal Discrete (NOD) transform. Firstly, a reorder mixer was designed to iteratively process the random seed vector and PUF response, resulting in the inner product of the non-orthogonal confusion matrix and the response confusion matrix, upon which the NOD spectrum was established. The algorithm effectively solved the bias of key caused by insufficient uniformity of PUF. Then, the partition encoding and decoding strategy enabled the NOD spectrum to have the ability to tolerate certain errors, significantly improving the reliability of the final response by limiting the impact of unstable responses to a limited range. Compared to traditional error correcting code-based methods, the proposed algorithm requires fewer auxiliary data. Experimental results on SRAM-XMC dataset show that, during 101 repeated experiments with 2 949 120 sets of 64-bit responses, the average reliability of the proposed algorithm reaches 99.97%, the uniqueness achieves 49.92%, and the uniformity reaches 50.61%. The experimental results demonstrate that the proposed algorithm can effectively improve reliability while ensuring uniformity and uniqueness of PUF responses.

Table and Figures | Reference | Related Articles | Metrics
Rectified cross pseudo supervision method with attention mechanism for stroke lesion segmentation
Yan ZHOU, Yang LI
Journal of Computer Applications    2024, 44 (6): 1942-1948.   DOI: 10.11772/j.issn.1001-9081.2023060742
Abstract169)   HTML5)    PDF (1757KB)(283)       Save

The automatic segmentation of brain lesions provides a reliable basis for the timely diagnosis and treatment of stroke patients and the formulation of diagnosis and treatment plans, but obtaining large-scale labeled data is expensive and time-consuming. Semi-Supervised Learning (SSL) methods alleviate this problem by utilizing a large number of unlabeled images and a limited number of labeled images. Aiming at the two problems of pseudo-label noise in SSL and the lack of ability of existing Three-Dimensional (3D) networks to focus on smaller objects, a semi-supervised method was proposed, namely, a rectified cross pseudo supervised method with attention mechanism for stroke lesion segmentation RPE-CPS (Rectified Cross Pseudo Supervision with Project & Excite modules). First, the data was input into two 3D U-Net segmentation networks with the same structure but different initializations, and the obtained pseudo-segmentation graphs were used for cross-supervised training of the segmentation networks, making full use of the pseudo-label data to expand the training set, and encouraging a high similarity between the predictions of different initialized networks for the same input image. Second, a correction strategy about cross-pseudo-supervised approach based on uncertainty estimation was designed to reduce the impact of the noise in pseudo-labels. Finally, in the segmentation network of 3D U-Net, in order to improve the segmentation performance of small object classes, Project & Excite (PE) modules were added behind each encoder module, decoder module and bottleneck module. In order to verify the effectiveness of the proposed method, evaluation experiments were carried out on the Acute Ischemic Stroke (AIS) dataset of the cooperative hospital and the Ischemic Stroke Lesion Segmentation Challenge (ISLES2022) dataset. The experimental results showed that when only using 20% of the labeled data in the training set, the Dice Similarity Coefficient (DSC), 95% Hausdorff Distance (HD95), and Average Surface Distance (ASD) on the public dataset ISLES2022 reached 73.87%, 6.08 mm and 1.31 mm; on the AIS dataset, DSC, HD95, and ASD reached 67.74%, 15.38 mm and 1.05 mm, respectively. Compared with the state-of-the-art semi-supervised method Uncertainty Rectified Pyramid Consistency(URPC), DSC improved by 2.19 and 3.43 percentage points, respectively. The proposed method can effectively utilize unlabeled data to improve segmentation accuracy, outperforms other semi-supervised methods, and is robust.

Table and Figures | Reference | Related Articles | Metrics
Progressive enhancement algorithm for low-light images based on layer guidance
Mengyuan HUANG, Kan CHANG, Mingyang LING, Xinjie WEI, Tuanfa QIN
Journal of Computer Applications    2024, 44 (6): 1911-1919.   DOI: 10.11772/j.issn.1001-9081.2023060736
Abstract186)   HTML8)    PDF (6161KB)(165)       Save

The quality of low-light images is poor and Low-Light Image Enhancement (LLIE) aims to improve the visual quality. Most of LLIE algorithms focus on enhancing luminance and contrast, while neglecting details. To solve this issue, a Progressive Enhancement algorithm for low-light images based on Layer Guidance (PELG) was proposed, which enhanced algorithm images to a suitable illumination level and reconstructed clear details. First, to reduce the task complexity and improve the efficiency, the image was decomposed into several frequency components by Laplace Pyramid (LP) decomposition. Secondly, since different frequency components exhibit correlation, a Transformer-based fusion model and a lightweight fusion model were respectively proposed for layer guidance. The Transformer-based model was applied between the low-frequency and the lowest high-frequency components. The lightweight model was applied between two neighbouring high-frequency components. By doing so, components were enhanced in a coarse-to-fine manner. Finally, the LP was used to reconstruct the image with uniform brightness and clear details. The experimental results show that, the proposed algorithm achieves the Peak Signal-to-Noise Ratio (PSNR) 2.3 dB higher than DSLR (Deep Stacked Laplacian Restorer) on LOL(LOw-Light dataset)-v1 and 0.55 dB higher than UNIE (Unsupervised Night Image Enhancement) on LOL-v2. Compared with other state-of-the-art LLIE algorithms, the proposed algorithm has shorter runtime and achieves significant improvement in objective and subjective quality, which is more suitable for real scenes.

Table and Figures | Reference | Related Articles | Metrics
Multi-object cache side-channel attack detection model based on machine learning
Zihao YAO, Yuanming LI, Ziqiang MA, Yang LI, Lianggen WEI
Journal of Computer Applications    2024, 44 (6): 1862-1871.   DOI: 10.11772/j.issn.1001-9081.2023060787
Abstract256)   HTML9)    PDF (3387KB)(103)       Save

Current cache side-channel attack detection technology mainly aims at a single attack mode. The detection methods for two to three attacks are limited and cannot fully cover them. In addition, although the detection accuracy of a single attack is high, as the number of attacks increases, the accuracy decreases and false positives are easily generated. To effectively detect cache side-channel attacks, a multi-object cache side-channel attack detection model based on machine learning was proposed, which utilized Hardware Performance Counter (HPC) to collect various cache side-channel attack features. Firstly, relevant feature analysis was conducted on various cache side-channel attack modes, and key features were selected and data sets were collected. Then, independent training was carried out to establish a detection model for each attack mode. Finally, during detection, test data was input into multiple models in parallel. The detection results from multiple models were employed to ascertain the presence of any cache side-channel attack. Experimental results show that the proposed model reaches high accuracies of 99.91%, 98.69% and 99.54% respectively when detecting three cache side-channel attacks: Flush+Reload, Flush+Flush and Prime+Probe. Even when multiple attacks exist at the same time, various attack modes can be accurately identified.

Table and Figures | Reference | Related Articles | Metrics
3D-GA-Unet: MRI image segmentation algorithm for glioma based on 3D-Ghost CNN
Lijun XU, Hui LI, Zuyang LIU, Kansong CHEN, Weixuan MA
Journal of Computer Applications    2024, 44 (4): 1294-1302.   DOI: 10.11772/j.issn.1001-9081.2023050606
Abstract174)   HTML12)    PDF (3121KB)(499)       Save

Gliomas are the most common primary cranial tumors arising from cancerous changes in the glia of the brain and spinal cord, with a high proportion of malignant gliomas and a significant mortality rate. Quantitative segmentation and grading of gliomas based on Magnetic Resonance Imaging (MRI) images is the main method for diagnosis and treatment of gliomas. To improve the segmentation accuracy and speed of glioma, a 3D-Ghost Convolutional Neural Network (CNN) -based MRI image segmentation algorithm for glioma, called 3D-GA-Unet, was proposed. 3D-GA-Unet was built based on 3D U-Net (3D U-shaped Network). A 3D-Ghost CNN block was designed to increase the useful output and reduce the redundant features in traditional CNNs by using linear operation. Coordinate Attention (CA) block was added, which helped to obtain more image information that was favorable to the segmentation accuracy. The model was trained and validated on the publicly available glioma dataset BraTS2018. The experimental results show that 3D-GA-Unet achieves average Dice Similarity Coefficients (DSCs) of 0.863 2, 0.847 3 and 0.803 6 and average sensitivities of 0.867 6, 0.949 2 and 0.831 5 for Whole Tumor (WT), Tumour Core (TC), and Enhanced Tumour (ET) in glioma segmentation results. It is verified that 3D-GA-Unet can accurately segment glioma images and further improve the segmentation efficiency, which is of positive significance for the clinical diagnosis of gliomas.

Table and Figures | Reference | Related Articles | Metrics
Video super-resolution reconstruction network based on frame straddling optical flow
Yang LIU, Rong LIU, Ke FANG, Xinyue ZHANG, Guangxu WANG
Journal of Computer Applications    2024, 44 (4): 1277-1284.   DOI: 10.11772/j.issn.1001-9081.2023040523
Abstract242)   HTML7)    PDF (3588KB)(118)       Save

Current Video Super-Resolution (VSR) algorithms cannot fully utilize inter-frame information of different distances when processing complex scenes with large motion amplitude, resulting in difficulty in accurately recovering occlusion, boundaries, and multi-detail regions. A VSR model based on frame straddling optical flow was proposed to solve these problems. Firstly, shallow features of Low-Resolution frames (LR) were extracted through Residual Dense Blocks (RDBs). Then, motion estimation and compensation was performed on video frames using a Spatial Pyramid Network (SPyNet) with straddling optical flows of different time lengths, and deep feature extraction and correction was performed on inter-frame information through RDBs connected in multiple layers. Finally, the shallow and deep features were fused, and High-Resolution frames (HR) were obtained through up-sampling. The experimental results on the REDS4 public dataset show that compared with deep Video Super-Resolution network using Dynamic Upsampling Filters without explicit motion compensation (DUF-VSR), the proposed model improves Peak Signal-to-Noise Ratio (PSNR) and Structure Similarity Index Measure (SSIM) by 1.07 dB and 0.06, respectively. The experimental results show that the proposed model can effectively improve the quality of video image reconstruction.

Table and Figures | Reference | Related Articles | Metrics
Hyperparameter optimization for neural network based on improved real coding genetic algorithm
Wei SHE, Yang LI, Lihong ZHONG, Defeng KONG, Zhao TIAN
Journal of Computer Applications    2024, 44 (3): 671-676.   DOI: 10.11772/j.issn.1001-9081.2023040441
Abstract446)   HTML58)    PDF (1532KB)(548)       Save

To address the problems of poor effects, easily falling into suboptimal solutions, and inefficiency in neural network hyperparameter optimization, an Improved Real Coding Genetic Algorithm (IRCGA) based hyperparameter optimization algorithm for the neural network was proposed, which was named IRCGA-DNN (IRCGA for Deep Neural Network). Firstly, a real-coded form was used to represent the values of hyperparameters, which made the search space of hyperparameters more flexible. Then, a hierarchical proportional selection operator was introduced to enhance the diversity of the solution set. Finally, improved single-point crossover and variational operators were designed to explore the hyperparameter space more thoroughly and improve the efficiency and quality of the optimization algorithm, respectively. Two simulation datasets were used to show IRCGA’s performance in damage effectiveness prediction and convergence efficiency. The experimental results on two datasets indicate that, compared to GA-DNN(Genetic Algorithm for Deep Neural Network), the proposed algorithm reduces the convergence iterations by 8.7% and 13.6% individually, and the MSE (Mean Square Error) is not much different; compared to IGA-DNN(Improved Genetic Algorithm for Deep Neural Network), IRCGA-DNN achieves reductions of 22.2% and 13.6% in convergence iterations respectively. Experimental results show that the proposed algorithm is better in both convergence speed and prediction performance, and is suitable for hyperparametric optimization of neural networks.

Table and Figures | Reference | Related Articles | Metrics
High-efficiency dual-LAN Terahertz WLAN MAC protocol based on spontaneous data transmission
Zhi REN, Jindong GU, Yang LIU, Chunyu CHEN
Journal of Computer Applications    2024, 44 (2): 519-525.   DOI: 10.11772/j.issn.1001-9081.2023020250
Abstract180)   HTML5)    PDF (1941KB)(72)       Save

In the existing Dual LAN (Local Area Network) Terahertz Wireless LAN (Dual-LAN THz WLAN) related MAC (Medium Access Control) protocol, some nodes may repeatedly send the same Channel Time Request (CTRq) frame within multiple superframes to apply for time slot resources and idle time slots exist in some periods of network operation, therefore an efficient MAC protocol based on spontaneous data transmission SDTE-MAC (high-Efficiency MAC Protocol based on Spontaneous Data Transmission) was proposed. SDTE-MAC protocol enabled each node to maintain one or more time unit linked lists to synchronize with the rest of the nodes in the network running time, so as to know where each node started sending data frames at the channel idle time slot. The protocol optimized the traditional channel slot allocation and channel remaining slot reallocation processes, improved network throughput and channel slot utilization, reduced data delay, and could further improve the performance of Dual-LAN THz WLAN. The simulation results showed that when the network saturates, compared with the new N-CTAP (Normal Channel Time Allocation Period) slot resource allocation mechanism and adaptive shortening superframe period mechanism in the AHT-MAC (Adaptive High Throughout multi-pan MAC protocol), the MAC layer throughput of the SDTE-MAC protocol was increased by 9.2%, the channel slot utilization was increased by 10.9%, and the data delay was reduced by 22.2%.

Table and Figures | Reference | Related Articles | Metrics
Construction and benchmark detection of multimodal partial forgery dataset
Shengyou ZHENG, Yanxiang CHEN, Zuxing ZHAO, Haiyang LIU
Journal of Computer Applications    2024, 44 (10): 3134-3140.   DOI: 10.11772/j.issn.1001-9081.2023101506
Abstract91)   HTML3)    PDF (1323KB)(27)       Save

Aiming at the lack of multimodal forgery scenarios and partial forgery scenarios in existing video forgery datasets, a multimodal partial forgery dataset with adjustable forgery ratios — PartialFAVCeleb was constructed by using a wide varieties of audio and video forgery methods. The proposed dataset was based on the FakeAVCeleb multimodal forgery dataset and was with the real and forged data spliced, in which the forgery data were generated by four methods, that is, FaceSwap, FSGAN (Face Swapping Generative Adversarial Network), Wav2Lip (Wave to Lip), and SV2TTS (Speaker Verification to Text-To-Speech). In the splicing process, probabilistic methods were used to generate the locations of the forgery segments in the time domain and modality, then the boundary was randomized to fit the actual forged scenario. And, the phenomenon of background hopping was avoided through material screening. The finally obtained dataset contains forgery videos of different ratios, with one ratio corresponding to 3 970 video data. In the benchmark detection, several audio and video feature extractors were used. And the data was tested in strong supervised and weakly-supervised conditions respectively, and Hierarchical Multi-Instance Learning (HMIL) method was used to realize the latter condition. As the test results indicate, for each test model, the performance on data with low forgery ratio is significantly inferior to that on data with high forgery ratio, and the performance under weakly-supervised condition is significantly inferior to that under strong supervised condition. The difficulty of weakly-supervised detection of proposed partial forgery dataset is verified. Experimental results show that the multimodal partial forgery scenario represented by the proposed dataset has sufficient research value.

Table and Figures | Reference | Related Articles | Metrics
Image tampering forensics network based on residual feedback and self-attention
Guolong YUAN, Yujin ZHANG, Yang LIU
Journal of Computer Applications    2023, 43 (9): 2925-2931.   DOI: 10.11772/j.issn.1001-9081.2022081283
Abstract349)   HTML22)    PDF (1998KB)(187)       Save

The existing multi-tampering type image forgery detection algorithms using noise features often can not effectively detect the feature difference between tampered areas and non-tampered areas, especially for copy-move tampering type. To this end, a dual-stream image tampering forensics network fusing residual feedback and self-attention mechanism was proposed to detect tampering artifacts such as unnatural edges of RGB pixels and local noise inconsistence respectively through two streams. Firstly, in the encoder stage, multiple dual residual units integrating residual feedback were used to extract relevant tampering features to obtain coarse feature maps. Secondly, further feature reinforcement was performed on the coarse feature maps by the improved self-attention mechanism. Thirdly, the mutual corresponding shallow features of encoder and deep features of decoder were fused. Finally, the final features of tempering extracted by the two streams were fused in series, and then the pixel-level localization of the tampered area was realized through a special convolution operation. Experimental results show that the F1 score and Area Under Curve (AUC) value of the proposed network on COVERAGE dataset are better than those of the comparison networks. The F1 score of the proposed network is 9.8 and 7.7 percentage points higher than that of TED-Net (Two-stream Encoder-Decoder Network) on NIST16 and Columbia datasets, and the AUC increases by 1.1 and 6.5 percentage points, respectively. The proposed network achieves good results in copy-move tampering type detection, and is also suitable for other tampering type detection. At the same time, the proposed network can locate the tampered area at pixel level accurately, and its detection performance is superior to the comparison networks.

Table and Figures | Reference | Related Articles | Metrics
Research progress on motion segmentation of visual localization and mapping in dynamic environment
Dongying ZHU, Yong ZHONG, Guanci YANG, Yang LI
Journal of Computer Applications    2023, 43 (8): 2537-2545.   DOI: 10.11772/j.issn.1001-9081.2022070972
Abstract333)   HTML20)    PDF (2687KB)(247)       Save

Visual localization and mapping system is affected by dynamic objects in a dynamic environment, so that it has increase of localization and mapping errors and decrease of robustness. And motion segmentation of input images can significantly improve the performance of visual localization and mapping system in dynamic environment. Dynamic objects in dynamic environment can be divided into moving objects and potential moving objects. Current dynamic object recognition methods have problems of chaotic moving subjects and poor real-time performance. Therefore, motion segmentation strategies of visual localization and mapping system in dynamic environment were reviewed. Firstly, the strategies were divided into three types of methods according to preset conditions of the scene: methods based on static assumption of image subject, methods based on prior semantic knowledge and multi-sensor fusion methods without assumption. Then, these three types of methods were summarized, and their accuracy and real-time performance were analyzed. Finally, aiming at the difficulty of balancing accuracy and real-time performance of motion segmentation strategy of visual localization and mapping system in dynamic environment, development trends of the motion segmentation methods in dynamic environment were discussed and prospected.

Table and Figures | Reference | Related Articles | Metrics
Adaptive image deblurring generative adversarial network algorithm based on active discrimination mechanism
Anyang LIU, Huaici ZHAO, Wenlong CAI, Zechao XU, Ruideng XIE
Journal of Computer Applications    2023, 43 (7): 2288-2294.   DOI: 10.11772/j.issn.1001-9081.2022060840
Abstract267)   HTML9)    PDF (2675KB)(156)       Save

Aiming at the problems that existing image deblurring algorithms suffer from diffusion and artifacts when dealing with edge loss and the use of full-frame deblurring in video processing does not meet real-time requirements, an Adaptive DeBlurring Generative Adversarial Network (ADBGAN)algorithm based on active discrimination mechanism was proposed. Firstly, an adaptive fuzzy discrimination mechanism was proposed, and an adaptive fuzzy processing network module was developed to make a priori judgment of fuzziness on the input image. When collecting the input, the blurring degree of the input image was judged in advance, and the input frame which was clear enough was eliminated to improve the running efficiency of the algorithm. Then, the incentive link of the attention mechanism was introduced in the process of fine feature extraction, so that weight normalization was carried out in the forward flow of feature extraction to improve the performance of the network to recover fine-grained features. Finally, the feature pyramid fine feature recovery structure was improved in the generator architecture, and a more lightweight feature fusion process was adopted to improve the running efficiency. In order to verify the effectiveness of the algorithm, detailed comparison experiments were conducted on the open source datasets GoPro and Kohler. Experimental results on GoPro dataset show that the visual fidelity of ADBGAN is 2.1 times that of Scale-Recurrent Network (SRN) algorithm, the Peak Signal-to-Noise Ratio (PSNR) of ADBGAN is improved by 0.762 dB compared with that of SRN algorithm, and ADBGAN has good image information recovery ability; in terms of video processing time,the actual processing time is reduced by 85.9% compared to SRN.The proposed algorithm can generate deblurred images with higher information quality efficiently.

Table and Figures | Reference | Related Articles | Metrics
Pedestrian fall detection algorithm in complex scenes
Ke FANG, Rong LIU, Chiyu WEI, Xinyue ZHANG, Yang LIU
Journal of Computer Applications    2023, 43 (6): 1811-1817.   DOI: 10.11772/j.issn.1001-9081.2022050754
Abstract369)   HTML19)    PDF (2529KB)(202)       Save

With the deepening of population aging, fall detection has become a key issue in the medical and health field. Concerning the low accuracy of fall detection algorithms in complex scenes, an improved fall detection model PDD-FCOS (PVT DRFPN DIoU-Fully Convolutional One-Stage object detection) was proposed. Pyramid Vision Transformer (PVT) was introduced into the backbone network of baseline FCOS algorithm to extract richer semantic information without increasing the amount of computation. In the feature information fusion stage, Double Refinement Feature Pyramid Networks (DRFPN) were inserted to learn the positions and other information of sampling points between feature maps more accurately, and more accurate semantic relationship between feature channels was captured by context information to improve the detection performance. In the training stage, the bounding box regression was carried out by the Distance Intersection Over Union (DIoU) loss. By optimizing the distance between the prediction box and the center point of the object box, the regression box was made to converge faster and more accurately, which improved the accuracy of the fall detection algorithm effectively. Experimental results show that on the open-source dataset Fall detection Database, the mean Average Precision (mAP) of the proposed model reaches 82.2%, which is improved by 6.4 percentage points compared with that of the baseline FCOS algorithm, and the proposed algorithm has accuracy improvement and better generalization ability compared with other state-of-the-art fall detection algorithms.

Table and Figures | Reference | Related Articles | Metrics
Gibbs artifact removal algorithm for magnetic resonance imaging based on self-attention connection UNet
Yang LIU, Zhiyang LU, Jun WANG, Jun SHI
Journal of Computer Applications    2023, 43 (5): 1606-1611.   DOI: 10.11772/j.issn.1001-9081.2022040618
Abstract478)   HTML14)    PDF (1363KB)(194)       Save

To remove Gibbs artifacts in Magnetic Resonance Imaging (MRI), a Self-attention connection UNet based on Self-Distillation training (SD-SacUNet) algorithm was proposed. In order to reduce the semantic gap between the encoding and decoding features at both ends of the skip connection in the UNet framework and help to capture the location information of artifacts, the output features of each down-sampling layer at the UNet encoding end was input to the corresponding self-attention connection module for the calculation of the self-attention mechanism, then they were fused with the decoding features to participate in the reconstruction of the features. Self-distillation training was performed on the network decoding end, by establishing the loss function between the deep and shallow features, the feature information of the deep reconstruction network was used to guide the training of the shallow network, and at the same time, the entire network was optimized to improve the level of image reconstruction quality. The performance of SD-SacUNet algorithm was evaluated on the public MRI dataset CC359, with the Peak Signal-to-Noise Ratio (PSNR) of 30.261 dB and the Structure Similarity Index Measure (SSIM) of 0.917 9. Compared with GRACNN (Gibbs-Ringing Artifact reduction using Convolutional Neural Network), the proposed algorithm had the PSNR increased by 0.77 dB and SSIM increased by 0.018 3; compared with SwinIR (Image Restoration using Swin Transformer), the proposed algorithm had the PSNR increased by 0.14 dB and SSIM increased by 0.003 3. Experimental results show that SD-SacUNet algorithm improves the image reconstruction performance of MRI with Gibbs artifacts removal and has potential application values.

Table and Figures | Reference | Related Articles | Metrics
Multi-scale ship detection based on adaptive feature fusion in complex scenes
Fang LUO, Yang LIU, G. T. S HO
Journal of Computer Applications    2023, 43 (11): 3587-3593.   DOI: 10.11772/j.issn.1001-9081.2022101593
Abstract310)   HTML8)    PDF (1778KB)(181)       Save

Under the influence of complex weather such as typhoon, heavy fog, rain and snow, as well as occlusions and scale changes, the existing ship detection methods have the problems of false detection and missed detection. In order to solve the above complex scene problems, based on YOLOX-S model, a multi-scale ship detection method based on adaptive feature fusion was proposed. Firstly, a feature augmentation module was introduced into the backbone feature extraction network to suppress the interference of complex background noise on ship feature extraction. Then, considering the problem of deep and shallow feature fusion proportion, an adaptive feature fusion module was designed to make full use of deep and shallow features, thereby improving the multi-scale ship detection ability of the model. Finally, in the detection head network, the detection head was decoupled and an adaptive multi-task loss function was introduced to balance classification tasks and regression tasks, thereby improving the multi-scale ship detection robustness of the model. Experimental results show that the detection mean Average Precision (mAP) of the proposed method on the public ship detection datasets SeaShips and McShips is 97.43% and 96.10%, respectively. The detection speed of the proposed method reaches 189 frames per second, which meets the requirements of real-time detection, demonstrating that the proposed method achieves high-precision detection of multi-scale ship targets even in complex scenes.

Table and Figures | Reference | Related Articles | Metrics
Identifier obfuscation method based on low level virtual machine
Dajiang TIAN, Chengyang LI, Tianbo HUANG, Weiping WEN
Journal of Computer Applications    2022, 42 (8): 2540-2547.   DOI: 10.11772/j.issn.1001-9081.2021071166
Abstract330)   HTML7)    PDF (901KB)(135)       Save

Most of the existing code obfuscation solutions are limited to a specific programming language or a platform, which are not widespread and general. Moreover, control flow obfuscation and data obfuscation introduce additional overhead. Aiming at the above problems, an identifier obfuscation method was proposed based on Low Level Virtual Machine (LLVM). Four identifier obfuscation algorithms were implemented in the method, including random identifier algorithm, overload induction algorithm, abnormal identifier algorithm, and high-frequency word replacement algorithm. At the same time, a new hybrid obfuscation algorithm was designed by combining these algorithms. In the proposed method, firstly, in the intermediate files compiled by the front-ends, the function names, which met the obfuscation criteria, were selected. Secondly, these function names were processed by using specific obfuscation algorithms. Finally, the obfuscated files were transformed into binary files by using specific compilation back-ends. The identifier obfuscation method based on LLVM is suitable for the languages supported by LLVM and does not affect the normal functions of the program. For different programming languages, the time overhead is within 20% and the space overhead hardly increases. At the same time, the average confusion ratio of the program is 77.5%, and compared with the single replacement algorithm and overload algorithm, the proposed mixed identifier algorithm can provide stronger concealment in theoretical analysis. Experimental results show that the proposed method has the characteristics of low-performance overhead, strong concealment, and wide versatility.

Table and Figures | Reference | Related Articles | Metrics
Hierarchical resource allocation mechanism of cooperative mobile edge computing
Jieqin WANG, Shihyang LIN, Shiming PENG, Shuo JIA, Miaohui YANG
Journal of Computer Applications    2022, 42 (8): 2501-2510.   DOI: 10.11772/j.issn.1001-9081.2021060901
Abstract357)   HTML5)    PDF (1262KB)(105)       Save

Concerning the large number of computing needs of vehicle task offloading and the limited computing capacity of local edge servers in the Internet of Vehicles (IoV), a Hierarchical Resource Allocation Mechanism of cooperative mobile edge computing (HRAM) was proposed. In this algorithm, the computing resources of Mobile Edge Computing (MEC) servers were reasonably allocated and effectively utilized with a multi-layer architecture,so that the data multi-hop forwarding delay between different MEC servers was reduced, and the delay of task offloading requests was optimized. Firstly, the system model, communication model, decision model, and calculation model of the IoV edge computing were built. Next, the Analytic Hierarchy Process (AHP) was used to comprehensively consider multiple factors to determine the target server the offloaded task transferred to. Finally, a task routing strategy with dynamic weights was proposed to make use of communication capabilities of the overall network to shorten the request delay of task offloading. Simulation results show that compared with Resource Allocation of Task Offloading in Single-hop (RATOS) algorithm and Resource Allocation of Task Offloading in Multi-hop (RATOM) algorithm, HRAM algorithm reduces the request delay of task offloading by 40.16% and 19.01% respectively, and this algorithm can satisfy the computing needs of more offloaded tasks under the premise of meeting the maximum tolerable delay.

Table and Figures | Reference | Related Articles | Metrics
No-reference image quality assessment algorithm based on saliency deep features
Jia LI, Yuanlin ZHENG, Kaiyang LIAO, Haojie LOU, Shiyu LI, Zehao CHEN
Journal of Computer Applications    2022, 42 (6): 1957-1964.   DOI: 10.11772/j.issn.1001-9081.2021040597
Abstract431)   HTML15)    PDF (1551KB)(188)       Save

Aiming at the universal No-Reference Image Quality Assessment (NR-IQA) algorithms, a new NR-IQA algorithm based on the saliency deep features of the pseudo reference image was proposed. Firstly, based on the distorted image, the corresponding pseudo reference image of the distorted image generated by ConSinGAN model was used as compensation information of the distorted image, thereby making up for the weakness of NR-IQA methods: lacking real reference information. Secondly, the saliency information of the pseudo reference image was extracted, and the pseudo saliency map and the distorted image were input into VGG16 netwok to extract deep features. Finally, the obtained deep features were merged and mapped into the regression network composed of fully connected layers to obtain a quality prediction consistent with human vision.Experiments were conducted on four large public image datasets TID2013, TID2008, CSIQ and LIVE to prove the effectiveness of the proposed algorithm. The results show that the Spearman Rank-Order Correlation Coefficient (SROCC) of the proposed algorithm on the TID2013 dataset is 5 percentage points higher than that of H-IQA (Hallucinated-IQA) algorithm and 14 percentage points higher than that of RankIQA (learning from Rankings for no-reference IQA) algorithm. The proposed algorithm also has stable performance for the single distortion types. Experimental results indicate that the proposed algorithm is superior to the existing mainstream Full-Reference Image Quality Assessment (FR-IQA) and NR-IQA algorithms, and is consistent with human subjective perception performance.

Table and Figures | Reference | Related Articles | Metrics
Smart contract-based access control architecture and verification for internet of things
Yang LI, Long XU, Yanqiang LI, Shaopeng LI
Journal of Computer Applications    2022, 42 (6): 1922-1931.   DOI: 10.11772/j.issn.1001-9081.2021040553
Abstract445)   HTML25)    PDF (3580KB)(206)       Save

Concerning the problem that the traditional access control methods face single point of failure and fail to provide trusted, secure and dynamic access management, a new access control model based on blockchain and smart contract for Wireless Sensor Network (WSN) was proposed to solve the problems of access dynamics and low level of intelligence of existing blockchain-based access control methods. Firstly, a new access control architecture based on blockchain was proposed to reduce the network computing overhead. Secondly, a multi-level smart contract system including Agent Contract (AC), Authority Management Contract (AMC) and Access Control Contract (ACC) was built, thereby realizing the trusted and dynamic access management of WSN. Finally, the dynamic access generation algorithm based on Radial Basis Function (RBF) neural network was adopted, and access policy was combined to generate the credit score threshold of access node to realize the intelligent, dynamic access control management for the large number of sensors in WSN. Experimental results verify the availability, security and effectiveness of the proposed model in WSN secure access control applications.

Table and Figures | Reference | Related Articles | Metrics
Route discovery method based on trajectory point clustering
Haiyang LIU, Linghang MENG, Zhonghang LIN, Yuantao GU
Journal of Computer Applications    2022, 42 (3): 890-894.   DOI: 10.11772/j.issn.1001-9081.2021030425
Abstract375)   HTML9)    PDF (1771KB)(78)       Save

To strengthen the control and management of local airspace routes, a route discovery method based on trajectory point clustering was proposed. Firstly, for the simulation data generated according to the distribution characteristics of the real data, the pre-processing module was used to weaken and remove the noise of the trajectory data. Secondly, a route discovery method including outlier elimination, trajectory resampling, trajectory point clustering, clustering center correction, and connecting clustering centers was proposed to extract the routes. Finally, the result of route extraction was visualized and the proposed method was validated using civil aviation data. The experimental results on the simulated data show that the node coverage and the length coverage of the proposed method is 99% and 94% respectively, under the noise intensity of 0.1° and the buffer area of 30 km. Compared with the rasterization method, the proposed method has higher accuracy and can extract the routes more effectively, achieving the purpose of extracting the common routes of aircraft.

Table and Figures | Reference | Related Articles | Metrics
Multi-attention fusion network for medical image segmentation
Hong LI, Junying ZOU, Xicheng TAN, Guiyang LI
Journal of Computer Applications    2022, 42 (12): 3891-3899.   DOI: 10.11772/j.issn.1001-9081.2021101737
Abstract641)   HTML20)    PDF (1600KB)(347)       Save

In the field of deep medical image segmentation, TransUNet (merit both Transformers and U-Net) is one of the current advanced segmentation models. However, the local connection between adjacent blocks in its encoder is not considered, and the inter-channel information is not interactive during the upsampling process of the decoder. To address the above problems, a Multi-attention FUsion Network (MFUNet) model was proposed. Firstly, a Feature Fusion Module (FFM) was introduced in encoder part to enhance the local connections between adjacent blocks in the Transformer and maintain the spatial location relationships of the images themselves. Then, a Double Channel Attention (DCA) module was introduced in the decoder part to fuse the channel information of multi-level features, which enhanced the sensitivity of the model to the key information between channels. Finally, the model's constraints on the segmentation results was strengthened by combining cross-entropy loss and Dice loss. By conducting experiments on Synapse and ACDC public datasets, it can be seen that MFUNet achieves Dice Similarity Coefficient (DSC) of 81.06% and 90.91%, respectively. Compared with the baseline model TransUNet, MFUNet achieved an 11.5% reduction in Hausdorff Distance (HD) on the Synapse dataset, and improved segmentation accuracy by 1.43 and 3.48 percentage points on the ACDC dataset for both the right ventricular and myocardial components, respectively. The experimental results show that MFUNet can achieve better segmentation results in both internal filling and edge prediction of medical images, which can help improve the diagnostic efficiency of doctors in clinical practice.

Table and Figures | Reference | Related Articles | Metrics
Efficient failure recovery method for stream data processing system
Yang LIU, Yangyang ZHANG, Haoyi ZHOU
Journal of Computer Applications    2022, 42 (11): 3337-3345.   DOI: 10.11772/j.issn.1001-9081.2021122108
Abstract403)   HTML15)    PDF (2031KB)(155)       Save

Focusing on the issue that the single point of failure cannot be efficiently handled by streaming data processing system Flink, a new fault?tolerant system based on incremental state and backup, Flink+, was proposed. Firstly, backup operators and data paths were established in advance. Secondly, the output data in the data flow diagram was cached, and disks were used if necessary. Thirdly, task state synchronization was performed during system snapshots. Finally, backup tasks and cached data were used to recover calculation in case of system failure. In the system experiment and test, Flink+ dose not significantly increase the additional fault tolerance overhead during fault?free operation; when dealing with the single point of failure in both single?machine and distributed environments, compared with Flink system, the proposed system has the failure recovery time reduced by 96.98% in single?machine 8?task parallelism and by 88.75% in distributed 16?task parallelism. Experimental results show that using incremental state and backup method together can effectively reduce the recovery time of the single point of failure of the stream system and enhance the robustness of the system.

Table and Figures | Reference | Related Articles | Metrics
Color image demosaicking network based on inter-channel correlation and enhanced information distillation
Hengxin LI, Kan CHANG, Yufei TAN, Mingyang LING, Tuanfa QIN
Journal of Computer Applications    2022, 42 (1): 245-251.   DOI: 10.11772/j.issn.1001-9081.2021010127
Abstract345)   HTML7)    PDF (1841KB)(85)       Save

In commercial digital cameras, due to the limitation of Complementary Metal Oxide Semiconductor (CMOS) sensors, there is only one color channel information for each pixel in the sampled image. Therefore, the Color image DeMosaicking (CDM) algorithm is required to restore the full-color images. However, most of the existing Convolutional Neural Network (CNN)-based CDM algorithms cannot achieve satisfactory performance with relatively low computational complexity and small network parameter number. To solve this problem, a CDM network based on Inter-channel Correlation and Enhanced Information Distillation (ICEID) was proposed. Firstly, to fully utilize the inter-channel correlation of the color image, an inter-channel guided reconstruction structure was designed to obtain the initial CDM result. Secondly, an Enhanced Information Distillation Module (EIDM), which can effectively extract and refine features from image with relatively small parameter number, was presented to enhance the reconstructed full-color image in high efficiency. Experimental results demonstrate that compared with many state-of-the-art CDM methods, the proposed algorithm achieves significant improvement in both objective quality and subjective quality, and has relatively low computational complexity and small network parameter number.

Table and Figures | Reference | Related Articles | Metrics
Unsupervised attributed graph embedding model based on node similarity
Yang LI, Anbiao WU, Ye YUAN, Linlin ZHAO, Guoren WANG
Journal of Computer Applications    2022, 42 (1): 1-8.   DOI: 10.11772/j.issn.1001-9081.2021071221
Abstract923)   HTML132)    PDF (864KB)(552)       Save

Attributed graph embedding aims to represent the nodes in an attributed graph into low-dimensional vectors while preserving the topology information and attribute information of the nodes. There are lots of works related to attributed graph embedding. However, most of algorithms proposed in them are supervised or semi-supervised. In practical applications, the number of nodes that need to be labeled is large, which makes these algorithms difficult and consume huge manpower and material resources. Above problems were reanalyzed from an unsupervised perspective, and an unsupervised attributed graph embedding algorithm was proposed. Firstly, the topology information and attribute information of the nodes were calculated respectively by using the existing non-attributed graph embedding algorithm and attributes of the attributed graph. Then, the embedding vector of the nodes was obtained by using Graph Convolutional Network (GCN), and the difference between the embedding vector and the topology information and the difference between the embedding vector and attribute information were minimized. Finally, similar embeddings was obtained by the paired nodes with similar topological information and attribute information. Compared with Graph Auto-Encoder (GAE) method, the proposed method has the node classification accuracy improved by 1.2 percentage points and 2.4 percentage points on Cora and Citeseer datasets respectively. Experimental results show that the proposed method can effectively improve the quality of the generated embedding.

Table and Figures | Reference | Related Articles | Metrics
Strategy of energy-aware virtual machine migration based on three-way decision
YANG Ling, JIANG Chunmao
Journal of Computer Applications    2021, 41 (4): 990-998.   DOI: 10.11772/j.issn.1001-9081.2020081294
Abstract518)      PDF (1403KB)(593)       Save
As an important way to reduce energy consumption of the data center in cloud computing, virtual machine migration is widely used. By combing the trisecting-acting-outcome model of three-way decision, a Virtual Machine Migration scheduling strategy based on Three-Way Decision(TWD-VMM) was proposed. First, a hierarchical threshold tree was built to search all the possible threshold values to obtain the pair of thresholds with the lowest total energy consumption with the data center energy consumption as the optimization target. Thus, three regions were created:high-load region, medium-load region and low-load region. Second, different migration strategies were used for hosts with different loads. Specifically, for high-load hosts, the multidimensional resource balance and host load reduction after the pre-migration of hosts were adopted as targets; for low-load hosts, the host multidimensional resource balance after pre-placing hosts was mainly considered; for medium-load hosts, the virtual machines migrated from other regions would be accepted if they still met the medium-load characteristics. The experiments were conducted on CloudSim simulator, and TWD-VMM was compare with Threshold-based energy-efficient VM Scheduling in cloud datacenters(TVMS), Virtual machine migration Scheduling method optimising Energy-Efficiency of data center(EEVS) and Virtual Machine migration Scheduling to Reduce Energy consumption in datacenter(REVMS) algorithms respectively in the aspects including host load, balance of host multidimensional resource utilization and total data center energy consumption. The results show that TWD-VMM algorithm effectively improves host resource utilization and balances host load with an average energy consumption reduction of 27%.
Reference | Related Articles | Metrics
Point-of-interest recommendation algorithm combing dynamic and static preferences
YANG Li, WANG Shihui, ZHU Bo
Journal of Computer Applications    2021, 41 (2): 398-406.   DOI: 10.11772/j.issn.1001-9081.2020050677
Abstract596)      PDF (1355KB)(651)       Save
Since most existing Point-Of-Interest (POI) recommendation algorithms ignore the complexity of the modeling of the fusion of user dynamic and static preferences, a POI recommendation algorithm called CLSR (Combing Long Short Recommendation) was proposed that combined complex dynamic user preferences and general static user preferences. Firstly, in the process of modeling complex dynamic preferences, a hybrid neural network was designed based on the user's check-in behaviors and the skip behaviors in check-in behaviors to achieve the modeling of complex dynamic interests of the user. Secondly, in the process of general static preference modeling, a high-level attention network was used to learn the complex interactions between the user and POIs. Thirdly, a multi-layer neural network was used to further learn and express the above dynamic preferences and static preferences. Finally, a unified POI recommendation framework was used to integrate the preferences. Experimental results on real datasets show that, compared with FPMC-LR (Factorizing Personalized Markov Chain and Localized Region), PRME (Personalized Ranking Metric Embedding), Rank-GeoFM (Ranking based Geographical Factorization Method) and TMCA (Temporal and Multi-level Context Attention), CLSR has the performance greatly improved, and compared to the optimal TMCA among the comparison methods, the proposed algorithm has the precision, recall and normalized Discounted Cumulative Gain (nDCG) increased by 5.8%, 5.1%, and 7.2% on Foursquare dataset, and 7.3%, 10.2%, and 6.3% on Gowalla dataset. It can be seen that CLSR algorithm can effectively improve the results of POI recommendation.
Reference | Related Articles | Metrics
Real-time binocular foreground depth estimation algorithm based on sparse convolution
Zhehan QIU, Yang LI
Journal of Computer Applications    2021, 41 (12): 3680-3685.   DOI: 10.11772/j.issn.1001-9081.2021010076
Abstract414)   HTML6)    PDF (1709KB)(107)       Save

To improve the computational efficiency of stereo matching on foreground disparity estimation tasks, aiming at the disadvantage that the general networks use the complete binocular image as input and the input information redundancy is large due to the small proportion of the foreground space in the scene, a real-time target stereo matching algorithm based on sparse convolution was proposed. In order to realize and improve the sparse foreground disparity estimation of the algorithm, firstly, the sparse foreground mask and scene semantic features were obtained by the segmentation algorithm at the same time. Secondly, the sparse convolution was used to extract the spatial features of the foreground sparse region, and scene semantic features were fused with them. Then, the fused features were input into the decoding module for disparity regression. Finally, the foreground truth graph was used as the loss to generate the disparity graph. The test results on ApolloScape dataset show that the accuracy and real-time performance of the proposed algorithm are better than those of the state-of-the-art algorithms PSMNet (Pyramid Stereo Matching Network) and GANet (Guided Aggregation Network), and the single run time of the algorithm is as low as 60.5 ms. In addition, the proposed algorithm has certain robustness to the foreground occlusion, and can be used for the real-time depth estimation of targets.

Table and Figures | Reference | Related Articles | Metrics