The traditional age estimation methods based on ranking and regression cannot effectively utilize the evolutionary characteristics of human faces and build correlation between different ranking labels. Moreover, using binary classification methods for age estimation may result in inconsistent ranking issues. To solve above problems, an age estimation method based on integrated ranking matrix encoding and consistency preserving was proposed to fully utilize the correlation between age and ranking value and suppress the problem of inconsistent ranking. A new indicator, the proportion of samples with inconsistent ranking, was proposed to evaluate the problem of inconsistent rankings in the two-class ranking method. First, age categories were converted into a ranking matrix form through a designed coding method. Then, the ResNet34 (Residual Network) feature extraction network was used to extract facial features, which were then learned through the proposed encoding learning module. Finally, the network prediction results were decoded into the predicted age of the image through a ranking decoder based on a metric method. The experimental results show that: the proposed method achieves a Mean Absolute Error (MAE) of 2.18 on MORPH Ⅱ dataset, and has better results on other publicly available datasets compared to methods also based on ranking and ordinal regression, such as OR-CNN (Ordinal Regression with CNN) and CORAL (COnsistent RAnk Logits); at the same time, the proposed method decreases the proportion of samples with inconsistent ranking, and improves the measurement performance of ranking inconsistency by about 65% compared to the OR-CNN method.
In the era of big data and cloud computing, querying and analyzing temporal big data faces many important challenges. Focused on the issues such as poor query performance and ineffective utilization of indexes for temporal aggregation range query, a Distributed Temporal Index (DTI) for temporal aggregation range query was proposed. Firstly, random or round-robin strategy was used to partition the temporal data. Secondly, intra-partition index construction algorithm based on timestamp’s bit array prefix was used to build intra-partition index, and partition statistics including time span were recorded. Thirdly, the data partitions whose time span overlapped with the query time interval were selected by predicate pushdown operation, and were pre-aggregated by index scan. Finally, all pre-aggregated values obtained from each partition were merged and aggregated by time. The experimental results show that the execution time of intra-partition index construction algorithm of the index for processing data with density of 2 400 entries per unit of time is similar to the execution time for processing data with density of 0.001 entries per unit of time. Compared to ParTime, the temporal aggregation range query algorithm with index takes at least 22% less time for each step when querying the data in the first 75% of timeline and at least 11% less time for each step when executing selective aggregation. Therefore, the algorithm with index is faster in most temporal aggregate range query tasks and its intra-partition index construction algorithm is capable to solve data sparsity problem with high efficiency.
In order to solve the problems of low accuracy and poor generalization of the traditional sensitive information detection methods such as keyword character matching-based method and phrase-level sentiment analysis-based method, a sensitive information detection method based on Attention mechanism-based Embedding from Language Model (A-ELMo) was proposed. Firstly, the quick matched of trie tree was performed to reduce the comparison of useless words significantly, thereby improving the query efficiency greatly. Secondly, an Embedding from Language Model (ELMo) was constructed for context analysis, and the dynamic word vectors were used to fully represent the context characteristics to achieve high scalability. Finally, the attention mechanism was combined to enhance the identification ability of the model for sensitive features, and further improve the detection rate of sensitive information. Experiments were carried out on real datasets composed of multiple network data sources. The results show that the accuracy of the proposed sensitive information detection method is improved by 13.3 percentage points compared with that of the phrase-level sentiment analysis-based method, and the accuracy of the proposed method is improved by 43.5 percentage points compared with that of the keyword matching-based method, verifying that the proposed method has advantages in terms of enhancing identification ability of sensitive features and improving the detection rate of sensitive information.
The main task of text segmentation is to divide the text into several relatively independent text blocks according to the topic relevance. Aiming at the shortcomings of the existing text segmentation models in extracting fine-grained features such as text paragraph structural information, semantic correlation and context interaction, a text segmentation model TS-GCN (Text Segmentation-Graph Convolutional Network) based on Graph Convolutional Network (GCN) was proposed. Firstly, a text graph based on the structural information and semantic logic of text paragraphs was constructed. Then, the semantic similarity attention was introduced to capture the fine-grained correlation between text paragraph nodes, and the information transmission between high-order neighborhoods of text paragraph nodes was realized with the help of GCN, so that the model ability of multi-granularity extraction of text paragraph topic feature representations was enhanced. The proposed model was compared with the representative model CATS (Coherence-Aware Text Segmentation), and its basic model TLT-TS (Two-Level Transformer model for Text Segmentation), which were commonly used as benchmarks for text segmentation task. Experimental results show that TS-GCN’s evaluation index Pk is 0.08 percentage points lower than that of TLT-TS without any auxiliary module on Wikicities dataset. And the proposed model has the Pk value decreased by 0.38 percentage points and 2.30 percentage points respectively on Wikielements dataset compared with CATS and TLT-TS. It can be seen that TS-GCN achieves good segmentation effect.
The carrier-based aircrafts on the carrier deck are dense and occluded, so that the carrier-based aircraft targets are difficult to detect, and the detection effect is easily affected by the lighting condition and target size. Therefore, an improved Faster R-CNN (Faster Region with Convolutional Neural Network) carrier-based aircraft target detection method was proposed. In this method, a loss function with a repulsion loss strategy was designed, and combined with multi-scale training, pictures collected under laboratory condition were used to train and test the deep convolutional neural network. Test experiments show that compared with the original Faster R-CNN detection model, the improved model has a better detection effect on occluded aircraft targets, the recall increased by 7 percentage points, and the precision increased by 6 percentage points. The experimental results show that the proposed improved method can automatically and comprehensively extract the characteristics of carrier-based aircraft targets, solve the detection problem of occluded carrier-based aircraft targets, has the detection accuracy and speed which can meet the actual needs, and has strong adaptability and high robustness under different lighting conditions and target sizes.