Frontiers of Computer Science

ISSN 2095-2228

ISSN 2095-2236(Online)

CN 10-1014/TP

Postal Subscription Code 80-970

2018 Impact Factor: 1.129

   Online First

Administered by

, Volume 18 Issue 5

For Selected: View Abstracts Toggle Thumbnails
Software
Minimizing the cost of periodically replicated systems via model and quantitative analysis
Chenhao ZHANG, Liang WANG, Limin XIAO, Shixuan JIANG, Meng HAN, Jinquan WANG, Bing WEI, Guangjun QIN
Front. Comput. Sci.. 2024, 18 (5): 185206-.  
https://doi.org/10.1007/s11704-023-2625-8

Abstract   HTML   PDF (11214KB)

Geographically replicating objects across multiple data centers improves the performance and reliability of cloud storage systems. Maintaining consistent replicas comes with high synchronization costs, as it faces more expensive WAN transport prices and increased latency. Periodic replication is the widely used technique to reduce the synchronization costs. Periodic replication strategies in existing cloud storage systems are too static to handle traffic changes, which indicates that they are inflexible in the face of unforeseen loads, resulting in additional synchronization cost. We propose quantitative analysis models to quantify consistency and synchronization cost for periodically replicated systems, and derive the optimal synchronization period to achieve the best tradeoff between consistency and synchronization cost. Based on this, we propose a dynamic periodic synchronization method, Sync-Opt, which allows systems to set the optimal synchronization period according to the variable load in clouds to minimize the synchronization cost. Simulation results demonstrate the effectiveness of our models. Compared with the policies widely used in modern cloud storage systems, the Sync-Opt strategy significantly reduces the synchronization cost.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Empirically revisiting and enhancing automatic classification of bug and non-bug issues
Zhong LI, Minxue PAN, Yu PEI, Tian ZHANG, Linzhang WANG, Xuandong LI
Front. Comput. Sci.. 2024, 18 (5): 185207-.  
https://doi.org/10.1007/s11704-023-2771-z

Abstract   HTML   PDF (4767KB)

A large body of research effort has been dedicated to automated issue classification for Issue Tracking Systems (ITSs). Although the existing approaches have shown promising performance, the different design choices, including the different textual fields, feature representation methods and machine learning algorithms adopted by existing approaches, have not been comprehensively compared and analyzed. To fill this gap, we perform the first extensive study of automated issue classification on 9 state-of-the-art issue classification approaches. Our experimental results on the widely studied dataset reveal multiple practical guidelines for automated issue classification, including: (1) Training separate models for the issue titles and descriptions and then combining these two models tend to achieve better performance for issue classification; (2) Word embedding with Long Short-Term Memory (LSTM) can better extract features from the textual fields in the issues, and hence, lead to better issue classification models; (3) There exist certain terms in the textual fields that are helpful for building more discriminating classifiers between bug and non-bug issues; (4) The performance of the issue classification model is not sensitive to the choices of ML algorithms. Based on our study outcomes, we further propose an advanced issue classification approach, DEEPLABEL, which can achieve better performance compared with the existing issue classification approaches.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Artificial Intelligence
HACAN: a hierarchical answer-aware and context-aware network for question generation
Ruijun SUN, Hanqin TAO, Yanmin CHEN, Qi LIU
Front. Comput. Sci.. 2024, 18 (5): 185321-.  
https://doi.org/10.1007/s11704-023-2246-2

Abstract   HTML   PDF (9533KB)

Question Generation (QG) is the task of generating questions according to the given contexts. Most of the existing methods are based on Recurrent Neural Networks (RNNs) for generating questions with passage-level input for providing more details, which seriously suffer from such problems as gradient vanishing and ineffective information utilization. In fact, reasonably extracting useful information from a given context is more in line with our actual needs during questioning especially in the education scenario. To that end, in this paper, we propose a novel Hierarchical Answer-Aware and Context-Aware Network (HACAN) to construct a high-quality passage representation and judge the balance between the sentences and the whole passage. Specifically, a Hierarchical Passage Encoder (HPE) is proposed to construct an answer-aware and context-aware passage representation, with a strategy of utilizing multi-hop reasoning. Then, we draw inspiration from the actual human questioning process and design a Hierarchical Passage-aware Decoder (HPD) which determines when to utilize the passage information. We conduct extensive experiments on the SQuAD dataset, where the results verify the effectiveness of our model in comparison with several baselines.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Density estimation-based method to determine sample size for random sample partition of big data
Yulin HE, Jiaqi CHEN, Jiaxing SHEN, Philippe FOURNIER-VIGER, Joshua Zhexue HUANG
Front. Comput. Sci.. 2024, 18 (5): 185322-.  
https://doi.org/10.1007/s11704-023-2356-x

Abstract   HTML   PDF (12583KB)

Random sample partition (RSP) is a newly developed big data representation and management model to deal with big data approximate computation problems. Academic research and practical applications have confirmed that RSP is an efficient solution for big data processing and analysis. However, a challenge for implementing RSP is determining an appropriate sample size for RSP data blocks. While a large sample size increases the burden of big data computation, a small size will lead to insufficient distribution information for RSP data blocks. To address this problem, this paper presents a novel density estimation-based method (DEM) to determine the optimal sample size for RSP data blocks. First, a theoretical sample size is calculated based on the multivariate Dvoretzky-Kiefer-Wolfowitz (DKW) inequality by using the fixed-point iteration (FPI) method. Second, a practical sample size is determined by minimizing the validation error of a kernel density estimator (KDE) constructed on RSP data blocks for an increasing sample size. Finally, a series of persuasive experiments are conducted to validate the feasibility, rationality, and effectiveness of DEM. Experimental results show that (1) the iteration function of the FPI method is convergent for calculating the theoretical sample size from the multivariate DKW inequality; (2) the KDE constructed on RSP data blocks with sample size determined by DEM can yield a good approximation of the probability density function (p.d.f.); and (3) DEM provides more accurate sample sizes than the existing sample size determination methods from the perspective of p.d.f. estimation. This demonstrates that DEM is a viable approach to deal with the sample size determination problem for big data RSP implementation.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Label distribution similarity-based noise correction for crowdsourcing
Lijuan REN, Liangxiao JIANG, Wenjun ZHANG, Chaoqun LI
Front. Comput. Sci.. 2024, 18 (5): 185323-.  
https://doi.org/10.1007/s11704-023-2751-3

Abstract   HTML   PDF (6069KB)

In crowdsourcing scenarios, we can obtain each instance’s multiple noisy labels from different crowd workers and then infer its integrated label via label aggregation. In spite of the effectiveness of label aggregation methods, there still remains a certain level of noise in the integrated labels. Thus, some noise correction methods have been proposed to reduce the impact of noise in recent years. However, to the best of our knowledge, existing methods rarely consider an instance’s information from both its features and multiple noisy labels simultaneously when identifying a noise instance. In this study, we argue that the more distinguishable an instance’s features but the noisier its multiple noisy labels, the more likely it is a noise instance. Based on this premise, we propose a label distribution similarity-based noise correction (LDSNC) method. To measure whether an instance’s features are distinguishable, we obtain each instance’s predicted label distribution by building multiple classifiers using instances’ features and their integrated labels. To measure whether an instance’s multiple noisy labels are noisy, we obtain each instance’s multiple noisy label distribution using its multiple noisy labels. Then, we use the Kullback-Leibler (KL) divergence to calculate the similarity between the predicted label distribution and multiple noisy label distribution and define the instance with the lower similarity as a noise instance. The extensive experimental results on 34 simulated and four real-world crowdsourced datasets validate the effectiveness of our method.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Contactless interaction recognition and interactor detection in multi-person scenes
Jiacheng LI, Ruize HAN, Wei FENG, Haomin YAN, Song WANG
Front. Comput. Sci.. 2024, 18 (5): 185325-.  
https://doi.org/10.1007/s11704-023-2418-0

Abstract   HTML   PDF (2624KB)

Human interaction recognition is an essential task in video surveillance. The current works on human interaction recognition mainly focus on the scenarios only containing the close-contact interactive subjects without other people. In this paper, we handle more practical but more challenging scenarios where interactive subjects are contactless and other subjects not involved in the interactions of interest are also present in the scene. To address this problem, we propose an Interactive Relation Embedding Network (IRE-Net) to simultaneously identify the subjects involved in the interaction and recognize their interaction category. As a new problem, we also build a new dataset with annotations and metrics for performance evaluation. Experimental results on this dataset show significant improvements of the proposed method when compared with current methods developed for human interaction recognition and group activity recognition.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Graph-Segmenter: graph transformer with boundary-aware attention for semantic segmentation
Zizhang WU, Yuanzhu GAN, Tianhao XU, Fan WANG
Front. Comput. Sci.. 2024, 18 (5): 185327-.  
https://doi.org/10.1007/s11704-023-2563-5

Abstract   HTML   PDF (7855KB)

The transformer-based semantic segmentation approaches, which divide the image into different regions by sliding windows and model the relation inside each window, have achieved outstanding success. However, since the relation modeling between windows was not the primary emphasis of previous work, it was not fully utilized. To address this issue, we propose a Graph-Segmenter, including a graph transformer and a boundary-aware attention module, which is an effective network for simultaneously modeling the more profound relation between windows in a global view and various pixels inside each window as a local one, and for substantial low-cost boundary adjustment. Specifically, we treat every window and pixel inside the window as nodes to construct graphs for both views and devise the graph transformer. The introduced boundary-aware attention module optimizes the edge information of the target objects by modeling the relationship between the pixel on the object’s edge. Extensive experiments on three widely used semantic segmentation datasets (Cityscapes, ADE-20k and PASCAL Context) demonstrate that our proposed network, a Graph Transformer with Boundary-aware Attention, can achieve state-of-the-art segmentation performance.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Blockchain based federated learning for intrusion detection for Internet of Things
Nan SUN, Wei WANG, Yongxin TONG, Kexin LIU
Front. Comput. Sci.. 2024, 18 (5): 185328-.  
https://doi.org/10.1007/s11704-023-3026-8

Abstract   HTML   PDF (13066KB)

In Internet of Things (IoT), data sharing among different devices can improve manufacture efficiency and reduce workload, and yet make the network systems be more vulnerable to various intrusion attacks. There has been realistic demand to develop an efficient intrusion detection algorithm for connected devices. Most of existing intrusion detection methods are trained in a centralized manner and are incapable to identify new unlabeled attack types. In this paper, a distributed federated intrusion detection method is proposed, utilizing the information contained in the labeled data as the prior knowledge to discover new unlabeled attack types. Besides, the blockchain technique is introduced in the federated learning process for the consensus of the entire framework. Experimental results are provided to show that our approach can identify the malicious entities, while outperforming the existing methods in discovering new intrusion attack types.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
A MLP-Mixer and mixture of expert model for remaining useful life prediction of lithium-ion batteries
Lingling ZHAO, Shitao SONG, Pengyan WANG, Chunyu WANG, Junjie WANG, Maozu GUO
Front. Comput. Sci.. 2024, 18 (5): 185329-.  
https://doi.org/10.1007/s11704-023-3277-4

Abstract   HTML   PDF (13474KB)

Accurately predicting the Remaining Useful Life (RUL) of lithium-ion batteries is crucial for battery management systems. Deep learning-based methods have been shown to be effective in predicting RUL by leveraging battery capacity time series data. However, the representation learning of features such as long-distance sequence dependencies and mutations in capacity time series still needs to be improved. To address this challenge, this paper proposes a novel deep learning model, the MLP-Mixer and Mixture of Expert (MMMe) model, for RUL prediction. The MMMe model leverages the Gated Recurrent Unit and Multi-Head Attention mechanism to encode the sequential data of battery capacity to capture the temporal features and a re-zero MLP-Mixer model to capture the high-level features. Additionally, we devise an ensemble predictor based on a Mixture-of-Experts (MoE) architecture to generate reliable RUL predictions. The experimental results on public datasets demonstrate that our proposed model significantly outperforms other existing methods, providing more reliable and precise RUL predictions while also accurately tracking the capacity degradation process. Our code and dataset are available at the website of github.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Information Systems
How graph convolutions amplify popularity bias for recommendation?
Jiajia CHEN, Jiancan WU, Jiawei CHEN, Xin XIN, Yong LI, Xiangnan HE
Front. Comput. Sci.. 2024, 18 (5): 185603-.  
https://doi.org/10.1007/s11704-023-2655-2

Abstract   HTML   PDF (4510KB)

Graph convolutional networks (GCNs) have become prevalent in recommender system (RS) due to their superiority in modeling collaborative patterns. Although improving the overall accuracy, GCNs unfortunately amplify popularity bias — tail items are less likely to be recommended. This effect prevents the GCN-based RS from making precise and fair recommendations, decreasing the effectiveness of recommender systems in the long run.

In this paper, we investigate how graph convolutions amplify the popularity bias in RS. Through theoretical analyses, we identify two fundamental factors: (1) with graph convolution (i.e., neighborhood aggregation), popular items exert larger influence than tail items on neighbor users, making the users move towards popular items in the representation space; (2) after multiple times of graph convolution, popular items would affect more high-order neighbors and become more influential. The two points make popular items get closer to almost users and thus being recommended more frequently. To rectify this, we propose to estimate the amplified effect of popular nodes on each node’s representation, and intervene the effect after each graph convolution. Specifically, we adopt clustering to discover highly-influential nodes and estimate the amplification effect of each node, then remove the effect from the node embeddings at each graph convolution layer. Our method is simple and generic — it can be used in the inference stage to correct existing models rather than training a new model from scratch, and can be applied to various GCN models. We demonstrate our method on two representative GCN backbones LightGCN and UltraGCN, verifying its ability in improving the recommendations of tail items without sacrificing the performance of popular items. Codes are open-sourced

See github.com/MEICRS/DAP website.

.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Optimizing B+-tree for hybrid memory with in-node hotspot cache and eADR awareness
Peiquan JIN, Zhaole CHU, Gaocong LIU, Yongping LUO, Shouhong WAN
Front. Comput. Sci.. 2024, 18 (5): 185606-.  
https://doi.org/10.1007/s11704-023-3344-x

Abstract   HTML   PDF (1420KB)

The advance in Non-Volatile Memory (NVM) has changed the traditional DRAM-only memory system. Compared to DRAM, NVM has the advantages of non-volatility and large capacity. However, as the read/write speed of NVM is still lower than that of DRAM, building DRAM/NVM-based hybrid memory systems is a feasible way of adding NVM into the current computer architecture. This paper aims to optimize the well-known B+-tree for hybrid memory. The novelty of this study is two-fold. First, we observed that the space utilization of internal nodes in B+-tree is generally below 70%. Inspired by this observation, we propose to maintain hot keys in the free space within internal nodes, yielding a new index named HATree (Hotness-Aware Tree). The new idea of HATree is to use the unused space of the parent of leaf nodes (PLNs) as the hotspot data cache. Thus, no extra space is needed, and the in-node hotspot cache can efficiently improve query performance. Second, to further improve the update performance of HATree, we propose to utilize the eADR technology supported by the third-generation Intel Xeon Scalable Processors to enhance HATree with instant log persistence, which results in the new HATree-Log structure. We conduct extensive experiments on real hybrid memory architecture involving DRAM and Intel Optane Persistent Memory to evaluate the performance of HATree and HATree-Log. Three state-of-the-art indices for hybrid memory, namely NBTree, LBTree, and FPTree, are included in the experiments, and the results suggest the efficiency of HATree and HATree-Log.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
ARCHER: a ReRAM-based accelerator for compressed recommendation systems
Xinyang SHEN, Xiaofei LIAO, Long ZHENG, Yu HUANG, Dan CHEN, Hai JIN
Front. Comput. Sci.. 2024, 18 (5): 185607-.  
https://doi.org/10.1007/s11704-023-3397-x

Abstract   HTML   PDF (6980KB)

Modern recommendation systems are widely used in modern data centers. The random and sparse embedding lookup operations are the main performance bottleneck for processing recommendation systems on traditional platforms as they induce abundant data movements between computing units and memory. ReRAM-based processing-in-memory (PIM) can resolve this problem by processing embedding vectors where they are stored. However, the embedding table can easily exceed the capacity limit of a monolithic ReRAM-based PIM chip, which induces off-chip accesses that may offset the PIM profits. Therefore, we deploy the decomposed model on-chip and leverage the high computing efficiency of ReRAM to compensate for the decompression performance loss. In this paper, we propose ARCHER, a ReRAM-based PIM architecture that implements fully on-chip recommendations under resource constraints. First, we make a full analysis of the computation pattern and access pattern on the decomposed table. Based on the computation pattern, we unify the operations of each layer of the decomposed model in multiply-and-accumulate operations. Based on the access observation, we propose a hierarchical mapping schema and a specialized hardware design to maximize resource utilization. Under the unified computation and mapping strategy, we can coordinate the inter-processing elements pipeline. The evaluation shows that ARCHER outperforms the state-of-the-art GPU-based DLRM system, the state-of-the-art near-memory processing recommendation system RecNMP, and the ReRAM-based recommendation accelerator REREC by 15.79×, 2.21×, and 1.21 × in terms of performance and 56.06 ×, 6.45×, and 1.71 × in terms of energy savings, respectively.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Image and Graphics
GRAMO: geometric resampling augmentation for monocular 3D object detection
He GUAN, Chunfeng SONG, Zhaoxiang ZHANG
Front. Comput. Sci.. 2024, 18 (5): 185706-.  
https://doi.org/10.1007/s11704-023-3242-2

Abstract   HTML   PDF (8573KB)

Data augmentation is widely recognized as an effective means of bolstering model robustness. However, when applied to monocular 3D object detection, non-geometric image augmentation neglects the critical link between the image and physical space, resulting in the semantic collapse of the extended scene. To address this issue, we propose two geometric-level data augmentation operators named Geometric-Copy-Paste (Geo-CP) and Geometric-Crop-Shrink (Geo-CS). Both operators introduce geometric consistency based on the principle of perspective projection, complementing the options available for data augmentation in monocular 3D. Specifically, Geo-CP replicates local patches by reordering object depths to mitigate perspective occlusion conflicts, and Geo-CS re-crops local patches for simultaneous scaling of distance and scale to unify appearance and annotation. These operations ameliorate the problem of class imbalance in the monocular paradigm by increasing the quantity and distribution of geometrically consistent samples. Experiments demonstrate that our geometric-level augmentation operators effectively improve robustness and performance in the KITTI and Waymo monocular 3D detection benchmarks.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Information Security
BVDFed: Byzantine-resilient and verifiable aggregation for differentially private federated learning
Xinwen GAO, Shaojing FU, Lin LIU, Yuchuan LUO
Front. Comput. Sci.. 2024, 18 (5): 185810-.  
https://doi.org/10.1007/s11704-023-3142-5

Abstract   HTML   PDF (8813KB)

Federated Learning (FL) has emerged as a powerful technology designed for collaborative training between multiple clients and a server while maintaining data privacy of clients. To enhance the privacy in FL, Differentially Private Federated Learning (DPFL) has gradually become one of the most effective approaches. As DPFL operates in the distributed settings, there exist potential malicious adversaries who manipulate some clients and the aggregation server to produce malicious parameters and disturb the learning model. However, existing aggregation protocols for DPFL concern either the existence of some corrupted clients (Byzantines) or the corrupted server. Such protocols are limited to eliminate the effects of corrupted clients and server when both are in existence simultaneously due to the complicated threat model. In this paper, we elaborate such adversarial threat model and propose BVDFed. To our best knowledge, it is the first Byzantine-resilient and Verifiable aggregation for Differentially private FEDerated learning. In specific, we propose Differentially Private Federated Averaging algorithm (DPFA) as our primary workflow of BVDFed, which is more lightweight and easily portable than traditional DPFL algorithm. We then introduce Loss Score to indicate the trustworthiness of disguised gradients in DPFL. Based on Loss Score, we propose an aggregation rule DPLoss to eliminate faulty gradients from Byzantine clients during server aggregation while preserving the privacy of clients’ data. Additionally, we design a secure verification scheme DPVeri that are compatible with DPFA and DPLoss to support the honest clients in verifying the integrity of received aggregated results. And DPVeri also provides resistance to collusion attacks with no more than t participants for our aggregation. Theoretical analysis and experimental results demonstrate our aggregation to be feasible and effective in practice.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Provable secure authentication key agreement for wireless body area networks
Yuqian MA, Wenbo SHI, Xinghua LI, Qingfeng CHENG
Front. Comput. Sci.. 2024, 18 (5): 185811-.  
https://doi.org/10.1007/s11704-023-2548-4

Abstract   HTML   PDF (4080KB)

Wireless body area networks (WBANs) guarantee timely data processing and secure information preservation within the range of the wireless access network, which is in urgent need of a new type of security technology. However, with the speedy development of hardware, the existing security schemes can no longer meet the new requirements of anonymity and lightweight. New solutions that do not require complex calculations, such as certificateless cryptography, attract great attention from researchers. To resolve these difficulties, Wang et al. designed a new authentication architecture for the WBANs environment, which was claimed to be secure and efficient. However, in this paper, we will show that this scheme is prone to ephemeral key leakage attacks. Further, based on this authentication scheme, an anonymous certificateless scheme is proposed for lightweight devices. Meanwhile, user anonymity is fully protected. The proposed scheme is proved to be secure under a specific security model. In addition, we assess the security attributes our scheme meets through BAN logic and Scyther tool. The comparisons of time consumption and communication cost are given at the end of the paper, to demonstrate that our scheme performs prior to several previous schemes.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
Delegable zk-SNARKs with proxies
Jinrui SHA, Shengli LIU
Front. Comput. Sci.. 2024, 18 (5): 185812-.  
https://doi.org/10.1007/s11704-023-2782-9

Abstract   HTML   PDF (8703KB)

In this paper, we propose the concept of delegable zero knowledge succinct non-interactive arguments of knowledge (zk-SNARKs). The delegable zk-SNARK is parameterized by (μ,k,k,k). The delegable property of zk-SNARKs allows the prover to delegate its proving ability to μ proxies. Any k honest proxies are able to generate the correct proof for a statement, but the collusion of less than k proxies does not obtain information about the witness of the statement. We also define k-soundness and k-zero knowledge by taking into consider of multi-proxies.

We propose a construction of (μ,2t+1,t,t)- delegable zk-SNARK for the NPC language of arithmetic circuit satisfiability. Our delegable zk-SNARK stems from Groth’s zk-SNARK scheme (Groth16). We take advantage of the additive and multiplicative properties of polynomial-based secret sharing schemes to achieve delegation for zk-SNARK. Our secret sharing scheme works well with the pairing groups so that the nice succinct properties of Groth’s zk-SNARK scheme are preserved, while augmenting the delegable property and keeping soundness and zero-knowledge in the scenario of multi-proxies.

Figures and Tables | References | Supplementary Material | Related Articles | Metrics
16 articles