Traditional data augmentation techniques, such as synonym substitution, random insertion, and random deletion, may change the original semantics of text and even result in the loss of critical information. Moreover, data in text classification tasks typically have both textual and label parts. However, traditional data augmentation methods only focus on the textual part. To address these issues, a Label Confusion incorporated Data Augmentation (LCDA) technique was proposed for providing a comprehensive enhancement of data from both textual and label aspects. In terms of text, by enhancing the text through random insertion and replacement of punctuation marks and completing end-of-sentence punctuation marks, textual diversity was increased with all textual information and sequence preserved. In terms of labels, simulated label distribution was generated using a label confusion approach, and used to replace the traditional one-hot label distribution, so as to better reflect the relationships among instances and labels as well as between labels. In experiments conducted on few-shot datasets constructed from THUCNews (TsingHua University Chinese News) and Toutiao Chinese news datasets, the proposed technique was combined with TextCNN, TextRNN, BERT (Bidirectional Encoder Representations from Transformers), and RoBERTa-CNN (Robustly optimized BERT approach Convolutional Neural Network) text classification models. The experimental results indicate that compared to those before enhancement, all models demonstrate significant performance improvements. Specifically, on 50-THU, a dataset constructed on THUCNews dataset, the accuracies of four models combing LCDA technique are improved by 1.19, 6.87, 3.21, and 2.89 percentage points, respectively, compared to those before enhancement, and by 0.78, 7.62, 1.75, and 1.28 percentage points, respectively, compared to those of the four models combining softEDA (Easy Data Augmentation with soft labels) method. By both textual and label processing results, model accuracy is enhanced by LCDA technique significantly, particularly in application scenarios characterized by limited data availability.
To address the power resource limitations of wireless sensors in over-the-air computation networks and the spectrum competition with existing wireless information communication networks, a cognitive wireless network integrating information communication and over-the-air computation was studied, in which the primary network focused on wireless information communication, and the secondary network aimed to support over-the-air computation where the sensors utilized signals sent by the base station of the primary network for energy harvesting. Considering the constraints of the Mean Square Error (MSE) of over-the-air computation and the transmit power of each node in the network, base on the random channel uncertainty, a robust resource optimization problem was formulated, with the objective function of maximizing the sum rate of wireless information communication users. To solve the robust optimization problem effectively, an Alternating Optimization (AO)-Improved Constrained Stochastic Successive Convex Approximation (ICSSCA) algorithm called AO-ICSSCA,was proposed, by which the original robust optimization problem was transformed into deterministic optimization sub-problems, and the downlink beamforming vector of the base station in the primary network, the power factors of the sensors, and the fusion beamforming vector of the fusion center in the secondary network were alternately optimized. Simulation experimental results demonstrate that AO-ICSSCA algorithm achieves superior performance with less computing time compared to the Constrained Stochastic Successive Convex Approximation (CSSCA) algorithm before improvement.
As a kind of side information, Knowledge Graph (KG) can effectively improve the recommendation quality of recommendation models, but the existing knowledge-awareness recommendation methods based on Graph Neural Network (GNN) suffer from unbalanced utilization of node information. To address the above problem, a new recommendation method based on Knowledge?awareness and Cross-level Contrastive Learning (KCCL) was proposed. To alleviate the problem of unbalanced node information utilization caused by the sparse interaction data and noisy knowledge graph that deviate from the true representation of inter-node dependencies during information aggregation, a contrastive learning paradigm was introduced into knowledge-awareness recommendation model of GNN. Firstly, the user-item interaction graph and the item knowledge graph were integrated into a heterogeneous graph, and the node representations of users and items were realized by a GNN based on the graph attention mechanism. Secondly, consistent noise was added to the information propagation aggregation layer for data augmentation to obtain node representations of different levels, and the obtained outermost node representation was compared with the innermost node representation for cross-level contrastive learning. Finally, the supervised recommendation task and the contrastive learning assistance task were jointly optimized to obtain the final representation of each node. Experimental results on DBbook2014 and MovieLens-1m datasets show that compared to the second prior contrastive method, the Recall@10 of KCCL is improved by 3.66% and 0.66%, respectively, and the NDCG@10 is improved by 3.57% and 3.29%, respectively, which verifies the effectiveness of KCCL.
Existing multi-hop clustered routing protocols for Intelligent Road Cone Ad-hoc Network (IRCAN) suffer from redundancy in network control overhead and the average number of hops for data packet transmission is not guaranteed to be minimal. To solve the above problems, combined with the link characteristics of the network topology, an efficient clustered routing protocol based on non-random retroverted clustering, called Retroverted-Clustering-based Hierarchy Routing (RCHR), was proposed. Firstly, the retroverted clustering mechanism based on central extension and the cluster head selection algorithm based on overhearing, cross-layer sharing, and extending the adjacency matrix was proposed. Then, the proposed mechanism and the proposed algorithm were used to generate clusters with retroverted characteristics around sink nodes in sequence, and to select the optimal cluster heads for sink nodes at different directions without additional conditions. Thus, networking control overhead and time were decreased, and the formed network topology was profit for diminishing the average number of hops for data packet transmission. Theoretic analysis validated the effectiveness of the proposed protocol. The simulation experiment results show that compared with Ring-Based Multi-hop Clustering (RBMC) routing protocol and MODified Low Energy Adaptive Clustering Hierarchy (MOD-LEACH) protocol, the networking control overhead and the average number of hops for data packet transmission of the proposed protocol are reduced by 32.7% and 2.6% at least, respectively.
As a specific type of big data in biology, similarity of gene expression data is not based on Euclidean distance but on whether gene expression values show a trend of both rise and fall together, although they are all ordinary real values. The current gene Bayesian network uses gene expression level values as node random variables and does not reflect the similarity of this kind of subspace pattern. Therefore, a Bayesian network disease Classification algorithm based on Gene Association analysis (BCGA) was proposed to learn Bayesian networks from labeled disease sample-gene expression data and predict the classification of new disease samples. Firstly, disease samples were discretized and filtered to select genes, and the dimensionally reduced gene expression values were sorted and replaced with gene column subscripts. Secondly, the subscript sequence of gene column was decomposed into a set of atomic sequences with a length of 2, and the frequent atomic sequence of this set was corresponding to the association of a pair of genes. Finally, causal relationships were measured through gene association entropy for Bayesian network structure learning. Besides, the parameter learning of BCGA became easy, and the conditional probability distribution of a gene node was able to be obtained by counting the atomic sequence occurrence frequency of the gene and its parent node gene. Experimental results on multiple tumor and non-tumor gene expression datasets show that BCGA significantly improves disease classification accuracy and effectively reduces analysis time compared to the existing similar algorithms. In addition, BCGA uses gene association entropy instead of conditional independence, and gene atomic sequences instead of gene expression values, which can better fit gene expression data better.
Aiming at the classification problems of Three-Dimensional (3D) models, a method of few-shot recognition of 3D models based on Transformer was proposed. Firstly, the 3D point cloud models of the support and query samples were fed into the feature extraction module to obtain feature vectors. Then, the attention features of the support samples were calculated in the Transformer module. Finally, the cosine similarity network was used to calculate the relation scores between the query samples and the support samples. On ModelNet 40 dataset, compared with the Dual-Long Short-Term Memory (Dual-LSTM) method, the proposed method has the recognition accuracy of 5-way 1-shot and 5-way 5-shot increased by 34.54 and 21.00 percentage points, respectively. At the same time, the proposed method also obtains high accuracy on ShapeNet Core dataset. Experimental results show that the proposed method can recognize new categories of 3D models more accurately.
By using complex pre-training targets and a large number of model parameters, Pre-Training Model (PTM) can effectively obtain rich knowledge from unlabeled data. However, the development of the multimodal PTMs is still in its infancy. According to the difference between modals, most of the current multimodal PTMs were divided into the image-text PTMs and video-text PTMs. According to the different data fusion methods, the multimodal PTMs were divided into two types: single-stream models and two-stream models. Firstly, common pre-training tasks and downstream tasks used in validation experiments were summarized. Secondly, the common models in the area of multimodal pre-training were sorted out, and the downstream tasks of each model and the performance and experimental data of the models were listed in tables for comparison. Thirdly, the application scenarios of M6 (Multi-Modality to Multi-Modality Multitask Mega-transformer) model, Cross-modal Prompt Tuning (CPT) model, VideoBERT (Video Bidirectional Encoder Representations from Transformers) model, and AliceMind (Alibaba’s collection of encoder-decoders from Mind) model in specific downstream tasks were introduced. Finally, the challenges and future research directions faced by related multimodal PTM work were summed up.
To solve the long training time and slow convergence problems when applying SAC (Soft Actor-Critic) algorithm to the local path planning of mobile robots, a PER-SAC algorithm was proposed by introducing the Prioritized Experience Replay (PER) technique. Firstly, to improve the convergence speed and stability of the robot training process, a priority strategy was applied to extract samples from the experience pool instead of the traditional random sampling and the network prioritized the training of samples with larger errors. Then, the calculation of Temporal-Difference (TD) error was optimized, and the training deviation was reduced. Next, the transfer learning was used to train the robot from a simple environment to a complex one gradually in order to improve the training speed. In addition, an improved reward function was designed to increase the intrinsic reward of robots, and therefore, the sparsity problem of environmental reward was solved. Finally, the simulation was carried out on the ROS (Robot Operating System) platform, and the simulation results show that PER-SAC algorithm outperforms the original algorithm in terms of convergence speed and length of the planned path in different obstacle environments. Moreover, the PER-SAC algorithm can reduce the training time and is significantly better than the original algorithm on path planning performance.
Aiming at the problems of low efficiency, low accuracy, excessive occupancy of human resources and intelligent classification algorithm miniaturization deployment requirements in China Customs risk control methods at this stage, a customs risk control method based on an improved Butterfly Feedback neural Network Version 2 (BFNet-V2) was proposed. Firstly, the Filling in Code (FC) algorithm was used to realize the semantic replacement of the customs tabular data to the analog image. Then, the analog image data was trained by using the BFNet-V2. The regular neural network structure was composed of left and right links, different convolution kernels and blocks, and small block design, and the residual short path was added to improve the overfitting and gradient disappearance. Finally, a Historical momentum Adaptive moment estimation algorithm (H-Adam) was proposed to optimize the gradient descent process and achieve a better adaptive learning rate adjustment, and classify customs data. Xception (eXtreme inception), Mobile Network (MobileNet), Residual Network (ResNet), and Butterfly Feedback neural Network (BF-Net) were selected as the baseline network structures for comparison. The Receiver Operating Characteristic curve (ROC) and the Precision-Recall curve (PR) of the BFNet-V2 contain the curves of the baseline network structures. Taking Transfer Learning (TL) as an example, compared with the four baseline network structures, the classification accuracy of BFNet-V2 increases by 4.30%,4.34%,4.10% and 0.37% respectively. In the process of classifying real-label data, the misjudgment rate of BFNet-V2 reduces by 70.09%,57.98%,58.36% and 10.70%, respectively. The proposed method was compared with eight classification methods including shallow and deep learning methods, and the accuracies on three datasets increase by more than 1.33%. The proposed method can realize automatic classification of tabular data and improve the efficiency and accuracy of customs risk control.
Aiming at the shortcomings of traditional retinal vessel segmentation algorithm such as low accuracy of vessel segmentation and mis-segmentation of focal areas, a Multi-scale Feature Enhanced retinal vessel segmentation algorithm based on U-Net (MFEU-Net) was proposed. Firstly, in order to solve the vanishing gradient problem, an improved Feature Information Enhancement Residual Module (FIE-RM) was designed to replace the convolution block of U-Net. Secondly, in order to enlarge the receptive field and improve the extraction ability of vascular information features, a multi-scale dense atrous convolution module was introduced at the bottom of U-Net. Finally, in order to reduce the information loss in the process of encoding and decoding, a multi-scale channel enhancement module was constructed at the skip connection of U-Net. Experimental results on Digital Retinal Images for Vessel Extraction (DRIVE) and CHASE_DB1 datasets show that compared with CS-Net (Channel and Spatial attention Network), the suboptimal algorithm in retinal vessel segmentation, MFEU-Net has the F1 score improved by 0.35 and 1.55 percentage points respectively, and the Area Under Curve (AUC) improved by 0.34 and 1.50 percentage points respectively. It is verified that MFEU-Net can improve the accuracy and robustness of retinal vessel segmentation effectively.
Focusing on the problems of disturbed car-following behavior and instability of traffic flow caused by the uncertainty of the driver’s acquisition of road velocity limit and time delay information, a car-following model TD-VDVL (Time-Delayed Velocity Difference and Velocity limit) was proposed with the consideration of the time-delayed velocity difference and the velocity limit information in the Internet of Vehicles (IoV) environment. Firstly, the speed change caused by time delay and road velocity limit information were introduced to improve the Full Velocity Difference (FVD) model. Then, the linear spectrum wave perturbation method was used to derive the traffic flow stability judgment basis of TD-VDVL model, and the influence of each parameter in the model on the stability of the system was analyzed separately. Finally, the numerical simulation experiments and comparative analysis were carried out using Matlab. In the simulation experiments, straight roads and circular roads were selected, and slight disturbance was imposed on the fleet during driving. When conditions were the same, TD-VDVL model had the smallest velocity fluctuation rate and the fluctuation of fleet headway compared to the Optimal Velocity (OV) and FVD models. Especially when the sensitivity coefficient of the velocity limit information was 0.3, and the sensitivity coefficient of the time-delayed speed difference was 0.3, the proposed model had the average fluctuation rate of the fleet velocity reached 2.35% at time of 500 s, and the peak and valley difference of fleet headway of only 0.019 4 m. Experimental results show that TD-VDVL model has a better stable area after introducing time-delayed velocity difference and velocity limit information, and can significantly enhance the ability of car-following fleet to absorb disturbance.
The clinicians diagnose Attention Deficit/Hyperactivity Disorder (ADHD) mainly based on on their subjective assessment, which lacks objective criteria to assist. To solve this problem, an automatic detection algorithm for ADHD based on speech pause and flatness was proposed. Firstly, the Frequency band Difference Energy Entropy Product (FDEEP) parameter was used to automatically locate the segment with voice from the speech and extract the speech pause features. Then, Transform Average Amplitude Squared Difference (TAASD) parameter was presented to calculate the voice multi-frequency and extract the flatness features. Finally, fusion features and the Support Vector Machine (SVM) classifier were combined to realize the automatic recognition of ADHD. The speech samples of the experiment were collected from 17 normal control children and 37 children with ADHD. Experimental results show that the proposed algorithm can effectively discriminate the normal children and children with ADHD, with an accuracy of 91.38%.
Cryptographic S-boxes (or black boxes) are nonlinear components in symmetric encryption algorithms, and their algebraic properties usually determine the security performance of these encryption algorithms. Differential uniformity, nonlinearity and revised transparency order are three basic indicators to evaluate the security properties of cryptographic S-boxes. They describe the S-box’s ability against differential cryptanalysis, linear cryptanalysis and differential power attack respectively. When the input size of the cryptographic S-box is large (for example, the input length of the S-box is larger than 15 bits), the needed solving time in Central Processing Unit (CPU) is still too long, or even the solution is impracticable. How to evaluate the algebraic properties of the large-size S-box quickly is currently a research hot point in the field. Therefore, a method to evaluate the algebraic properties of cryptographic S-boxes quickly was proposed on the basis of Graphics Processing Unit (GPU). In this method, the kernel functions were split into multiple threads by slicing technique, and an optimization scheme was proposed by combining the characteristics of solving differential uniformity, nonlinearity and revised transparency order to realize parallel computing. Experimental results show that compared with CPU-based implementation environment, single GPU based environment has the implementation efficiency significantly improved. Specifically, the time spent on calculating differential uniformity, nonlinearity, and revised transparency order is saved by 90.28%, 80%, and 66.67% respectively, which verifies the effectiveness of this method.
Aiming at the problems of slow detection and low recognition accuracy of road traffic signs in Chinese intelligent driving assistance system, an improved road traffic sign detection algorithm based on YOLOv3 (You Only Look Once version 3) was proposed. Firstly, MobileNetv2 was introduced into YOLOv3 as the basic feature extraction network to construct an object detection network module MN-YOLOv3 (MobileNetv2-YOLOv3). And two Down-up links were added to the backbone network of MN-YOLOv3 for feature fusion, thereby reducing the model parameters, and improving the running speed of the detection module as well as information fusion performance of the multi-scale feature maps. Then, according to the shape characteristics of traffic sign objects, K-Means++ algorithm was used to generate the initial cluster center of the anchor, and the DIOU (Distance Intersection Over Union) loss function was introduced to combine DIOU and Non-Maximum Suppression (NMS) for the bounding box regression. Finally, the Region Of Interest (ROI) and the context information were unified by ROI Align and merged to enhance the object feature expression. Experimental results show that the proposed algorithm has better performance, and the mean Average Precision (mAP) of the algorithm on the dataset CSUST (ChangSha University of Science and Technology) Chinese Traffic Sign Detection Benchmark (CCTSDB) can reach 96.20%. Compared with Faster R-CNN (Region Convolutional Neural Network), YOLOv3 and Cascaded R-CNN detection algorithms, the proposed algorithm has better real-time performance, higher detection accuracy, and is more robustness to various environmental changes.
For addressing real root isolation problem of transcendental function polynomials, an interval isolation algorithm for exponential function polynomials named exRoot was proposed. In the algorithm, the real root isolation problem of non-polynomial real functions was transformed into sign determination problem of polynomial, then was solved. Firstly, the Taylor substitution method was used to construct the polynomial nested interval of the objective function. Then, the problem of finding the root of the exponential function was transformed into the problem of determining the positivity and negativity of the polynomial in the intervals. Finally, a comprehensive algorithm was given and applied to determine the reachability of rational eigenvalue linear system tentatively. The proposed algorithm was implemented in Maple efficiently and easily with readable output results. Different from HSOLVER and numerical calculation method fsolve, exRoot avoids discussing the existence of roots directly, and theoretically has termination and completeness. It can reach any precision and can avoid the systematic error brought by numerical solution when being applied into the optimization problem.
A Feature Fusion Network (FFN) was proposed to judge the quality of 3D face point cloud acquired by binocular structured light scanner. Firstly, the 3D point cloud was preprocessed to cut out the face area, and the image obtained from the point cloud and the corresponding 2D plane projection was used as the input. Secondly, Dynamic Graph Convolutional Neural Network (DGCNN) and ShuffleNet were trained for point cloud learning. Then, the middle layer features of the two network modules were extracted and fused to fine-tune the whole network. Finally, three full connected layers were used to realize the five-class classification of 3D face point cloud (excellent, ordinary, stripe, burr, deformation). The proposed FFN achieved the classification accuracy of 83.7%, which was 5.8% higher than that of ShufflNet and 2.2% higher than that of DGCNN. The experimental results show that the weighted fusion of two-dimensional image features and point cloud features can achieve the complementary effect between different features.
In online kernel regression learning, the inverse matrix of the kernel matrix needs to be calculated when a new sample arrives, and the computational complexity is at least the square of the number of rounds. The idea of applying sketching method to hypothesis updating was introduced, and a more efficient online kernel regression algorithm via sketching method was proposed. Firstly, The loss function was set as the square loss, a new gradient descent algorithm, called FTL-Online Kernel Regression (F-OKR) was proposed, using the Nystr?m approximation method to approximate the Kernel, and applying the idea of Follow-The-Leader (FTL). Then, sketching method was used to accelerate F-OKR so that the computational complexity of F-OKR was reduced to the level of linearity with the number of rounds and sketch scale, and square with the data dimension. Finally, an efficient online kernel regression algorithm called Sketched Online Kernel Regression (SOKR) was designed. Compared to F-OKR, SOKR had no change in accuracy and reduced the runtime by about 16.7% on some datasets. The sub-linear regret bounds of these two algorithms were proved, and experimental results on standard regression datasets also verify that the algorithms have better performance than NOGD (Nystr?m Online Gradient Descent) algorithm, the average loss of all the datasets was reduced by about 64%.
In the field of deep medical image segmentation, TransUNet (merit both Transformers and U-Net) is one of the current advanced segmentation models. However, the local connection between adjacent blocks in its encoder is not considered, and the inter-channel information is not interactive during the upsampling process of the decoder. To address the above problems, a Multi-attention FUsion Network (MFUNet) model was proposed. Firstly, a Feature Fusion Module (FFM) was introduced in encoder part to enhance the local connections between adjacent blocks in the Transformer and maintain the spatial location relationships of the images themselves. Then, a Double Channel Attention (DCA) module was introduced in the decoder part to fuse the channel information of multi-level features, which enhanced the sensitivity of the model to the key information between channels. Finally, the model's constraints on the segmentation results was strengthened by combining cross-entropy loss and Dice loss. By conducting experiments on Synapse and ACDC public datasets, it can be seen that MFUNet achieves Dice Similarity Coefficient (DSC) of 81.06% and 90.91%, respectively. Compared with the baseline model TransUNet, MFUNet achieved an 11.5% reduction in Hausdorff Distance (HD) on the Synapse dataset, and improved segmentation accuracy by 1.43 and 3.48 percentage points on the ACDC dataset for both the right ventricular and myocardial components, respectively. The experimental results show that MFUNet can achieve better segmentation results in both internal filling and edge prediction of medical images, which can help improve the diagnostic efficiency of doctors in clinical practice.
With the development of communication technology, communication terminals gradually adopt software to be compatible with multiple communication modes and protocols. As in the traditional software radio architecture with a Central Processing Unit (CPU) of computer as an arithmetic unit, the wideband data throughput of high-speed wireless communication systems such as Multiple-Input Multiple-Output (MIMO) is not be satisfied, an acceleration method of Low Density Parity Check (LDPC) code decoder based on Graphics Processing Unit (GPU) was proposed. Firstly, according to the theoretical analysis of the acceleration performance of GPU parallelly accelerated heterogeneous computing in GNU Radio 4G/5G physical layer signal processing module, a more parallelly efficient Layered Normalized Min-Sum (LNMS) algorithm was adopted. Then, the decoding delay of the decoder was reduced by using the methods such as global synchronization strategy, reasonably allocation of GPU memory space and stream parallelism mechanism. At the same time, the LDPC code decoding process was optimized in parallel with the multi-threaded parallel technology in GPU. Finally, the GPU accelerated decoder was implemented and verified on the software radio platform, and the bit error rate performance and acceleration performance bottlenecks of the parallel decoder were analyzed. Experimental results show that compared with the traditional CPU serial code processing method, CPU+GPU heterogeneous platform has the decoding rate for LDPC codes increased to about 200 times, and the throughput of decoder can reach more than 1 Gb/s, especially in the case of large-scale data, the decoding performance is greatly improved compared with traditional decoder.
Most of the existing target detection algorithms rely on large?scale annotation datasets to ensure the accuracy of detection, however, it is difficult for some scenes to obtain a large number of annotation data and it consums a lot of human and material resources. In order to resolve this problem, a Few?Shot Target Detection method based on Negative Margin loss (NM?FSTD) was proposed. The negative margin loss method belonging to metric learning in Few?Shot Learning (FSL) was introduced into target detection, which could avoid mistakenly mapping the samples of the same novel classes to multiple peaks or clusters and helping to the classification of novel classes in few?shot target detection. Firstly, a large number of training samples and the target detection framework based on negative margin loss were used to train the model with good generalization performance. Then, the model was finetuned through a small number of labeled target category samples. Finally, the finetuned model was used to detect the new sample of target category. To verify the detection effect of NM?FSTD, MS COCO was used for training and evaluation. Experimental results show that the AP50 of NM?FSTD reaches 22.8%; compared with Meta R?CNN (Meta Regions with CNN features) and MPSR (Multi?Scale Positive Sample Refinement), the accuracies are improved by 3.7 and 4.9 percentage points, respectively. NM?FSTD can effectively improve the detection performance of target categories in the case of few?shot, and solve the problem of insufficient data in the field of target detection.
The key of cross-modal image-text retrieval is how to capture the semantic correlation between images and text effectively. Most of the existing methods learn the global semantic correlation between image region features and text features or local semantic correlation between inter-modality objects, and ignore the correlation between the intra-modality object relationships and inter-modality object relationships. To solve this problem, a method of Cross-Modal Tensor Fusion Network based on Semantic Relation Graph (CMTFN-SRG) for image-text retrieval was proposed. Firstly, the relationships of image regions and text words were generated by Graph Convolutional Network (GCN) and Bidirectional Gated Recurrent Unit (Bi-GRU) respectively. Then, the fine-grained semantic correlation between the data of two modals was learned by using the tensor fusion network to match the learned semantic relation graph of image regions and the graph of text words. At the same time, Gated Recurrent Unit (GRU) was used to learn global features of the image, and the global features of the image and the text were matched to capture the inter-modality global semantic correlation. The proposed method was compared with the Multi-Modality Cross Attention (MMCA) method on the benchmark datasets Flickr30K and MS-COCO. Experimental results show that the proposed method improves the Recall@1 of text-to-image retrieval task by 2.6%, 9.0% and 4.1% respectively on the test datasets Flickr30K, MS-COCO1K and MS-COCO5K.And mean Recall (mR) improves by 0.4, 1.3 and 0.1 percentage points respectively. It can be seen that the proposed method can effectively improve the precision of image-text retrieval.
The event that the user is interested in is extracted from the unstructured information, and then displayed to the user in a structured way, that is event extraction. Event extraction has a wide range of applications in information collection, information retrieval, document synthesis, and information questioning and answering. From the overall perspective, event extraction algorithms can be divided into four categories: pattern matching algorithms, trigger lexical methods, ontology-based algorithms, and cutting-edge joint model methods. In the research process, different evaluation methods and datasets can be used according to the related needs, and different event representation methods are also related to event extraction research. Distinguished by task type, meta-event extraction and subject event extraction are the two basic tasks of event extraction. Among them, meta-event extraction has three methods based on pattern matching, machine learning and neural network respectively, while there are two ways to extract subjective events: based on the event framework and based on ontology respectively. Event extraction research has achieved excellent results in single languages such as Chinese and English, but cross-language event extraction still faces many problems. Finally, the related works of event extraction were summarized and the future research directions were prospected in order to provide guidelines for subsequent research.
In view of the disadvantages of the standard Artificial Bee Colony (ABC) algorithm such as weak development ability and slow convergence, a new ABC algorithm based on multi-population combination strategy was proposed. Firstly, the different-dimensional coordination and multi-dimensional matching update mechanisms were introduced into the search equation. Then, two combination strategies were designed for the hire bee and the follow bee respectively. The combination strategy was composed of two sub-strategies focusing on breadth exploration and depth development respectively. In the follow bee stage, the population was divided into free subset and non-free subset, and different sub-strategies were adopted by the individuals belonging to different subsets to balance the exploration and development ability of algorithm. The 15 benchmark functions were used to compare the proposed improved ABC algorithm with the standard ABC algorithm and other three improved ABC algorithms. The results show that the proposed algorithm has better optimization performance in both low-dimensional and high-dimensional problems.
Concerning the problem that the development cycle of existing elliptic curve algorithm system level design is long and the performance-overhead indicators are not clear, a method of Hardware/Software (HW/SW) co-design based on Electronic System Level (ESL) was proposed. This method presented several HW/SW partitions by analyzing the theories and implementations of SM2 algorithm, and generated cycle-accurate models for HW modules with SystemC. Module and system verification were proposed to compare the executing cycle counts of HW/SW modules to obtain the best partition. Finally, the ESL models were converted to Rigister Transfer Level (RTL) models according to the CFG (Control Flow Graph) and DFG (Data Flow Graph) to perform logic synthesis and comparison. In the condition of 50 MHz,180 nm CMOS technology, when getting best performance,the execute time of point-multiply was 20 ms, with 83 000 gates and the power consuption was 2.23 mW. The experimental result shows that the system analysis is conducive to performance and resources evaluation, and has high applicability in encryption chip based on elliptic curve algorithm. The embedded SoC (System on Chip) based on this algorithm can choose appropriate architecture based on performance and resource constraints.
In Positron Emission Tomography (PET) computed imaging, traditional iterative algorithms have the problem of details loss and fuzzy object edges. A high quality Median Prior (MP) reconstruction algorithm based on correlation coefficient and Forward-And-Backward (FAB) diffusion was proposed to solve the problem in this paper. Firstly, a characteristic factor called correlation coefficient was introduced to represent the image local gray information. Then through combining the correlation coefficient and forward-and-backward diffusion model, a new model was made up. Secondly, considering that the forward-and-backward diffusion model has the advantages of dealing with background and edge separately, the proposed model was applied to Maximum A Posterior (MAP) reconstruction algorithm of the median prior distribution, thus a median prior reconstruction algorithm based on forward-and-backward diffusion was obtained. The simulation results show that, the new algorithm can remove the image noise while preserving object edges well. The Signal-to-Noise Ratio (SNR) and Root Mean Squared Error (RMSE) also show visually the improvement of the reconstructed image quality.
In visual detection of subminiature accessory, the extracted target contour will be affected by the existence of foreign matter in the field like dust and hair crumbs. In order to avoid the impact for measurement brought by foreign matter, a method of culling foreign matter fake information based on prior knowledge was put forward. Firstly, the corners of component image with foreign matter were detected. Secondly, the corner-distribution features of standard component were obtained by statistics. Finally, the judgment condition of foreign matter fake imformation was derived from the corner-distribution features of standard component to cull the foreign matter fake information. Through successful application in an actual engineering project, the processing experiments on three typical images with foreign matter prove that the proposed algorithm ensures the accuracy of the measurement, while effectively culling the foreign matter fake information in the images.
A lot of side channel attacks and cold boot attacks can leak secret information of cryptographic systems and destroy the security of traditional cryptographic schemes. This paper presented a certificateless encryption scheme which can resist the private key leakage. Based on the q-ABDHE (Augmented Bilinear Diffie-Hellman Exponent) hypothesis, the security of the scheme was proved. The leakage-resilient property was obtained via extractor. The leakage-resilient performance was analyzed. The theoretical analyses show that the relative leakage rate of private key can reach 1/8.
In order to apply reasoning rules of the description logic to analyze and solve the simple contradiction problem, the extension set was introduced to be the set theory foundation of the description logic SHOQ, and a new description logic named D-SHOQES (Dynamic Description Logic SHOQ Based on Extension Set) was proposed. The cut sets of extension concepts and extension roles were defined as atomic concepts and atomic roles, and the action theory was injected to obtain the qualitative change domain and the quantitative change domain of the concepts and roles. The semantics of concepts, roles and actions in D-SHOQES were given, as well as the Tableau-algorithm reasoning rules. Finally, the method of solving contradiction problem was researched, which offered a strategy for the solution to contradiction problem.