Journals
  Publication Years
  Keywords
Search within results Open Search
Please wait a minute...
For Selected: Toggle Thumbnails
User data management and control in internet of behaviors: a review
Yi HE, Yinan XIAO, Yunkai WEI, Supeng LENG
Journal of Computer Applications    2025, 45 (5): 1535-1547.   DOI: 10.11772/j.issn.1001-9081.2024050599
Abstract44)   HTML0)    PDF (1223KB)(139)       Save

In recent years, the rapid development of Internet of Things (IoT) has spurred the emergence of the Internet of Behavior (IoB), which leverages IoT-derived data and information to achieve higher levels of knowledge and wisdom, rapidly evolving into a promising technology in various application potential. IoB involves extensive collection, processing, and utilization of user behavioral data, thereby exposing user data security and privacy to significant risks. Therefore, it is vital to protect the IoB user data with effective data management and control. After introducing the fundamental concepts and characteristics of IoB, its development trends and the security and privacy risks associated with user data were analyzed. Furthermore, the current situation of management and control of behavioral data was elaborated, the main problems and challenges existed in IoB were discussed, and the potential research directions to achieve user data management and control in IoB were proposed.

Table and Figures | Reference | Related Articles | Metrics
Bias challenges of large language models: identification, evaluation, and mitigation
Yuemei XU, Yuqi YE, Xueyi HE
Journal of Computer Applications    2025, 45 (3): 697-708.   DOI: 10.11772/j.issn.1001-9081.2024091350
Abstract126)   HTML14)    PDF (2112KB)(97)       Save

Aiming at the unsafety and being out of control problems caused by biases in the output of Large Language Model (LLM), research status, techniques, and limitations related to biases in the existing LLMs were sorted deeply and analyzed from three aspects: bias identification, evaluation, and mitigation. Firstly, three key techniques of LLM were summed up to study the basic reasons of LLMs’ inevitable intrinsic biases. Secondly, three types of biases in LLMs were categorized into linguistic bias, demographic bias, and evaluation bias, and characteristics and causes of the biases were explored. Thirdly, a systematic review of the existing LLM bias evaluation benchmarks was carried out, and the strengths and weaknesses of these general-purpose, language-specific, and task-specific benchmarks were discussed. Finally, current LLM bias mitigation techniques were analyzed in depth from both model bias mitigation and data bias mitigation perspectives, and directions for their future refinement were pointed out. At the same time, the research directions for biases in LLMs were indicated by analysis: multi-cultural attribute evaluation of bias, lightweight bias mitigation techniques, and enhancement of the interpretability of biases.

Table and Figures | Reference | Related Articles | Metrics
Multi-view clustering network with deep fusion
Ziyi HE, Yan YANG, Yiling ZHANG
Journal of Computer Applications    2023, 43 (9): 2651-2656.   DOI: 10.11772/j.issn.1001-9081.2022091394
Abstract640)   HTML58)    PDF (1074KB)(405)       Save

Current deep multi-view clustering methods have the following shortcomings: 1) When feature extraction is carried out for a single view, only attribute information or structural information of the samples is considered, and these two types of information are not integrated. Thus, the extracted features cannot fully represent latent structure of the original data. 2) Feature extraction and clustering were divided into two separated processes, without establishing the relationship between them, so that the feature extraction process cannot be optimized by the clustering process. To solve these problems, a Deep Fusion based Multi-view Clustering Network (DFMCN) was proposed. Firstly, the embedding space of each view was obtained by combining autoencoder and graph convolution autoencoder to fuse attribute information and structure information of samples. Then, the embedding space of the fusion view was obtained through weighted fusion, and clustering was carried out in this space. And in the process of clustering, the feature extraction process was optimized by a two-layer self-supervision mechanism. Experimental results on FM (Fashion-MNIST), HW (HandWritten numerals), and YTF (YouTube Face) datasets show that the accuracy of DFMCN is higher than those of all comparison methods; and DFMCN has the accuracy increased by 1.80 percentage points compared with the suboptimal CMSC-DCCA (Cross-Modal Subspace Clustering via Deep Canonical Correlation Analysis) method on FM dataset, the Normalized Mutual Information (NMI) of DFMCN is increased by 1.26 to 14.84 percentage points compared to all methods except for CMSC-DCCA and DMSC (Deep Multimodal Subspace Clustering networks). Experimental results verify the effectiveness of the proposed method.

Table and Figures | Reference | Related Articles | Metrics
Cloud migration performance of tsinghua cloud monitoring platform
MA Haifeng, YI Hebali, WANG Ye, YANG Jiahai, ZHANG Chao
Journal of Computer Applications    2015, 35 (11): 3026-3030.   DOI: 10.11772/j.issn.1001-9081.2015.11.3026
Abstract627)      PDF (919KB)(783)       Save
With the popularization of cloud computing technology, many enterprises have migrated or are planning to migrate their business and applications to the cloud. But it may face the problems of application performance degradation, and the key business and applications may suffer security threats. Therefore, migrating to cloud or deploying in independent server is a problem that needs to be further studied. In this paper, based on the Tsinghua cloud platform, Tsinghua cloud monitoring platform was set up based on Nagios. Firstly, Tsinghua cloud platform and its architecture were introduced, and then Nagios and the architecture of Tsinghua cloud monitoring platform were discussed. For cloud migration performance evaluation, Ubuntu and Windows were used as operating system platforms, CPU load and memory usage were used as evaluation indexes, two applications of CPU computing type and server load type respectively ran on the cloud server and the independent server. At last, the experimental results were analyzed and compared. The experimental result shows that some applications have better performance on independent servers that may not be suitable for migrating to cloud platform.
Reference | Related Articles | Metrics
Lexical active attack on chaotic text zero-watermarking
LI Jing FANG Ding-yi HE Lu
Journal of Computer Applications    2012, 32 (09): 2603-2605.   DOI: 10.3724/SP.J.1087.2012.02603
Abstract1048)      PDF (494KB)(523)       Save
The conventional steganalysis techniques does not work in zero-watermarking technology because it does not modify carrier text. However, active attack model on text zero-watermarking has not been well studied. In order to solve this problem, an active attack algorithm for text zero-watermarking was proposed. Making use of synonym substitution technology, sync-attack and birthday-attack were defined. Through combining the two active attack methods, a lexical active attack algorithm on text zero-watermarking was designed and implemented. The experimental result shows the proposed algorithm can effectively destroy zero-watermarking without modifying carrier text massively.
Reference | Related Articles | Metrics
Cross-lingual knowledge transfer method based on alignment of representational space structures
Siyuan REN, Cheng PENG, Ke CHEN, Zhiyi HE
Journal of Computer Applications    0, (): 18-23.   DOI: 10.11772/j.issn.1001-9081.2024030297
Abstract47)   HTML1)    PDF (737KB)(33)       Save

In the field of Natural Language Processing (NLP), as an efficient method for sentence representation learning, contrastive learning mitigates the anisotropy of Transformer-based pre-trained language models effectively and enhances the quality of sentence representations significantly. However, the existing research focuses on English conditions, especially under supervised settings. Due to the lack of labeled data, it is difficult to utilize contrastive learning effectively to obtain high-quality sentence representations in most non-English languages. To address this issue, a cross-lingual knowledge transfer method for contrastive learning models was proposed, transferring knowledge across languages by aligning the structures of different language representation spaces. Based on this, a simple and effective cross-lingual knowledge transfer framework, TransCSE, was developed to transfer the knowledge from supervised English contrastive learning models to non-English models. Through knowledge transfer experiments from English to six directions, including French, Arabic, Spanish, Turkish, and Chinese, knowledge was transferred successfully from the supervised contrastive learning model SimCSE (Simple contrastive learning of sentence embeddings) to the multilingual pre-trained language model mBERT (Multilingual Bidirectional Encoder Representations from Transformers) by TransCSE. Experimental results show that model trained using the TransCSE framework achieves accuracy improvements of 17.95 and 43.27 percentage points on XNLI (Cross-lingual Natural Language Inference) and STS (Semantic Textual Similarity) 2017 benchmark datasets, respectively, compared to the original mBERT, proving the effectiveness of TransCSE. Moreover, compared to cross-lingual knowledge transfer methods based on shared parameters and representation alignment, TransCSE has the best performance on both XNLI and STS 2017 benchmark datasets.

Table and Figures | Reference | Related Articles | Metrics