Author Search Result

[Author] Jun ZHAO(23hit)

1-20hit(23hit)

  • Complexity Metrics for Software Architectures

    Jianjun ZHAO  

     
    LETTER-Software Engineering

      Vol:
    E87-D No:8
      Page(s):
    2152-2156

    A large body of research in the measurement of software complexity at code level has been conducted, but little effort has been made to measure the architectural-level complexity of a software system. In this paper, we propose some architectural-level metrics which are appropriate for evaluating the architectural attributes of a software system. The main feature of our approach is to assess the architectural-level complexity of a software system by analyzing its formal architectural specification, and therefore the process of metric computation can be automated completely.

  • GRAPHULY: GRAPH U-Nets-Based Multi-Level Graph LaYout

    Kai YAN  Tiejun ZHAO  Muyun YANG  

     
    LETTER-Computer Graphics

      Pubricized:
    2022/09/16
      Vol:
    E105-D No:12
      Page(s):
    2135-2138

    Graph layout is a critical component in graph visualization. This paper proposes GRAPHULY, a graph u-nets-based neural network, for end-to-end graph layout generation. GRAPHULY learns the multi-level graph layout process and can generate graph layouts without iterative calculation. We also propose to use Laplacian positional encoding and a multi-level loss fusion strategy to improve the layout learning. We evaluate the model with a random dataset and a graph drawing dataset and showcase the effectiveness and efficiency of GRAPHULY in graph visualization.

  • CLCMiner: Detecting Cross-Language Clones without Intermediates

    Xiao CHENG  Zhiming PENG  Lingxiao JIANG  Hao ZHONG  Haibo YU  Jianjun ZHAO  

     
    PAPER-Software Engineering

      Pubricized:
    2016/11/21
      Vol:
    E100-D No:2
      Page(s):
    273-284

    The proliferation of diverse kinds of programming languages and platforms makes it a common need to have the same functionality implemented in different languages for different platforms, such as Java for Android applications and C# for Windows phone applications. Although versions of code written in different languages appear syntactically quite different from each other, they are intended to implement the same software and typically contain many code snippets that implement similar functionalities, which we call cross-language clones. When the version of code in one language evolves according to changing functionality requirements and/or bug fixes, its cross-language clones may also need be changed to maintain consistent implementations for the same functionality. Thus, it is needed to have automated ways to locate and track cross-language clones within the evolving software. In the literature, approaches for detecting cross-language clones are only for languages that share a common intermediate language (such as the .NET language family) because they are built on techniques for detecting single-language clones. To extend the capability of cross-language clone detection to more diverse kinds of languages, we propose a novel automated approach, CLCMiner, without the need of an intermediate language. It mines such clones from revision histories, based on our assumption that revisions to different versions of code implemented in different languages may naturally reflect how programmers change cross-language clones in practice, and that similarities among the revisions (referred to as clones in diffs or diff clones) may indicate actual similar code. We have implemented a prototype and applied it to ten open source projects implementations in both Java and C#. The reported clones that occur in revision histories are of high precisions (89% on average) and recalls (95% on average). Compared with token-based code clone detection tools that can treat code as plain texts, our tool can detect significantly more cross-language clones. All the evaluation results demonstrate the feasibility of revision-history based techniques for detecting cross-language clones without intermediates and point to promising future work.

  • Improving Feature-Rich Transition-Based Constituent Parsing Using Recurrent Neural Networks

    Chunpeng MA  Akihiro TAMURA  Lemao LIU  Tiejun ZHAO  Eiichiro SUMITA  

     
    PAPER-Natural Language Processing

      Pubricized:
    2017/06/05
      Vol:
    E100-D No:9
      Page(s):
    2205-2214

    Conventional feature-rich parsers based on manually tuned features have achieved state-of-the-art performance. However, these parsers are not good at handling long-term dependencies using only the clues captured by a prepared feature template. On the other hand, recurrent neural network (RNN)-based parsers can encode unbounded history information effectively, but they perform not well for small tree structures, especially when low-frequency words are involved, and they cannot use prior linguistic knowledge. In this paper, we propose a simple but effective framework to combine the merits of feature-rich transition-based parsers and RNNs. Specifically, the proposed framework incorporates RNN-based scores into the feature template used by a feature-rich parser. On English WSJ treebank and SPMRL 2014 German treebank, our framework achieves state-of-the-art performance (91.56 F-score for English and 83.06 F-score for German), without requiring any additional unlabeled data.

  • Wideband Adaptive Beamforming Algorithm for Conformal Arrays Based on Sparse Covariance Matrix Reconstruction

    Pei CHEN  Dexiu HU  Yongjun ZHAO  Chengcheng LIU  

     
    PAPER-Antennas and Propagation

      Pubricized:
    2017/08/22
      Vol:
    E101-B No:2
      Page(s):
    548-554

    Aiming at solving the performance degradation caused by the covariance matrix mismatch in wideband beamforming for conformal arrays, a novel adaptive beamforming algorithm is proposed in this paper. In this algorithm, the interference-plus-noise covariance matrix is firstly reconstructed to solve the desired signal contamination problem. Then, a sparse reconstruction method is utilized to reduce the high computational cost and the requirement of sampling data. A novel cost function is formulated by the focusing matrix and singular value decomposition. Finally, the optimization problem is efficiently solved in a second-order cone programming framework. Simulation results using a cylindrical array demonstrate the effectiveness and robustness of the proposed algorithm and prove that this algorithm can achieve superior performance over the existing wideband beamforming methods for conformal arrays.

  • Random Occurrence of Contact Welding in Electrical Endurance Tests

    Laijun ZHAO  Zhenbiao LI  Hansi ZHANG  Makoto HASEGAWA  

     
    PAPER

      Vol:
    E94-C No:9
      Page(s):
    1362-1368

    To clarify how the occurrence of contact welding is related to the series of arc duration characteristics in consecutive make and break operations, electrical endurance tests were conducted on commercially available automotive relays, and the voltage waveforms of make and break arcs between the electrodes were recorded with LabVIEW. Experimental results indicate that welding may occur suddenly or randomly with increasing number of operations. A single arc or a group of make or break arcs with a long arc duration does not necessarily result in contact welding, but a group of longer make or break arcs within a narrow range of operation numbers can cause imminent contact welding (such an effect can be called the “group of longer arcing duration effect”). It is confirmed that contact welding may occur in both make and break operations, but the welding probability during make operations is much higher than that during break operations.

  • Local-to-Global Structure-Aware Transformer for Question Answering over Structured Knowledge

    Yingyao WANG  Han WANG  Chaoqun DUAN  Tiejun ZHAO  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2023/06/27
      Vol:
    E106-D No:10
      Page(s):
    1705-1714

    Question-answering tasks over structured knowledge (i.e., tables and graphs) require the ability to encode structural information. Traditional pre-trained language models trained on linear-chain natural language cannot be directly applied to encode tables and graphs. The existing methods adopt the pre-trained models in such tasks by flattening structured knowledge into sequences. However, the serialization operation will lead to the loss of the structural information of knowledge. To better employ pre-trained transformers for structured knowledge representation, we propose a novel structure-aware transformer (SATrans) that injects the local-to-global structural information of the knowledge into the mask of the different self-attention layers. Specifically, in the lower self-attention layers, SATrans focus on the local structural information of each knowledge token to learn a more robust representation of it. In the upper self-attention layers, SATrans further injects the global information of the structured knowledge to integrate the information among knowledge tokens. In this way, the SATrans can effectively learn the semantic representation and structural information from the knowledge sequence and the attention mask, respectively. We evaluate SATrans on the table fact verification task and the knowledge base question-answering task. Furthermore, we explore two methods to combine symbolic and linguistic reasoning for these tasks to solve the problem that the pre-trained models lack symbolic reasoning ability. The experiment results reveal that the methods consistently outperform strong baselines on the two benchmarks.

  • rOOM: A Rust-Based Linux Out of Memory Kernel Component

    Linhan LI  Qianying ZHANG  Zekun XU  Shijun ZHAO  Zhiping SHI  Yong GUAN  

     
    PAPER

      Pubricized:
    2023/12/14
      Vol:
    E107-D No:3
      Page(s):
    245-256

    The Linux kernel has been applied in various security-sensitive fields, so ensuring its security is crucial. Vulnerabilities in the Linux kernel are usually caused by undefined behaviors of the C programming language, the most threatening of which are memory safety vulnerabilities. Both the software-based and hardware approaches to memory safety have disadvantages of poor performance, false positives, and poor compatibility. This paper explores the feasibility of using the safe programming language Rust to reconstruct a Linux kernel component and open-source the component's code. We leverage the Rust FFI mechanism to design a safe foreign interface layer to enable the reconstructed component to invoke other Linux functionalities, and then use Rust to reconstruct the component, during which we leverage Rust's type-safety and ownership mechanisms to improve its security, and finally export the C interface of the component to enable the invocation by the Linux kernel. The performance and memory overhead of the reconstructed component, referred to as “rOOM”, were evaluated, revealing a performance overhead of 8.9% in kernel mode, 5% in user mode, 3% in real time, and a memory overhead of 0.06%. These results suggest that it is possible to develop key components of the Linux kernel using Rust in terms of functionality, performance, and memory overhead.

  • Hierarchical Latent Alignment for Non-Autoregressive Generation under High Compression Ratio

    Wang XU  Yongliang MA  Kehai CHEN  Ming ZHOU  Muyun YANG  Tiejun ZHAO  

     
    PAPER-Natural Language Processing

      Pubricized:
    2023/12/01
      Vol:
    E107-D No:3
      Page(s):
    411-419

    Non-autoregressive generation has attracted more and more attention due to its fast decoding speed. Latent alignment objectives, such as CTC, are designed to capture the monotonic alignments between the predicted and output tokens, which have been used for machine translation and sentence summarization. However, our preliminary experiments revealed that CTC performs poorly on document abstractive summarization, where a high compression ratio between the input and output is involved. To address this issue, we conduct a theoretical analysis and propose Hierarchical Latent Alignment (HLA). The basic idea is a two-step alignment process: we first align the sentences in the input and output, and subsequently derive token-level alignment using CTC based on aligned sentences. We evaluate the effectiveness of our proposed approach on two widely used datasets XSUM and CNNDM. The results indicate that our proposed method exhibits remarkable scalability even when dealing with high compression ratios.

  • A Novel RNN-GBRBM Based Feature Decoder for Anomaly Detection Technology in Industrial Control Network

    Hua ZHANG  Shixiang ZHU  Xiao MA  Jun ZHAO  Zeng SHOU  

     
    PAPER-Industrial Control System Security

      Pubricized:
    2017/05/18
      Vol:
    E100-D No:8
      Page(s):
    1780-1789

    As advances in networking technology help to connect industrial control networks with the Internet, the threat from spammers, attackers and criminal enterprises has also grown accordingly. However, traditional Network Intrusion Detection System makes significant use of pattern matching to identify malicious behaviors and have bad performance on detecting zero-day exploits in which a new attack is employed. In this paper, a novel method of anomaly detection in industrial control network is proposed based on RNN-GBRBM feature decoder. The method employ network packets and extract high-quality features from raw features which is selected manually. A modified RNN-RBM is trained using the normal traffic in order to learn feature patterns of the normal network behaviors. Then the test traffic is analyzed against the learned normal feature pattern by using osPCA to measure the extent to which the test traffic resembles the learned feature pattern. Moreover, we design a semi-supervised incremental updating algorithm in order to improve the performance of the model continuously. Experiments show that our method is more efficient in anomaly detection than other traditional approaches for industrial control network.

  • On the Diagnosis of Two-Dimensional Grid of Processors

    Jun ZHAO  Fred J. MEYER  Nohpill PARK  Fabrizio LOMBARDI  

     
    PAPER

      Vol:
    E84-D No:11
      Page(s):
    1486-1499

    We examine diagnosis of processor array systems formed as two-dimensional grids, with boundaries, and either four or eight neighbors for each interior processor. We employ a parallel test schedule. Neighboring processors test each other and report the results. Our diagnostic objective is to find a fault-free processor or set of processors. The system may then be sequentially diagnosed by repairing those processors tested faulty according to the identified fault-free set. We establish an upper bound on the maximum number of faults that can be sustained without invalidating the test results under worst case conditions. We give test schedules and diagnostic algorithms that meet the upper bound as far as the highest order term. We compare these near optimal diagnostic algorithms to alternative algorithms--both new and already in the literature.

  • Hardware Architecture of the Fast Mode Decision Algorithm for H.265/HEVC

    Wenjun ZHAO  Takao ONOYE  Tian SONG  

     
    PAPER-VLSI Design Technology and CAD

      Vol:
    E98-A No:8
      Page(s):
    1787-1795

    In this paper, a specified hardware architecture of the Fast Mode Decision (FMD) algorithms presented by our previous work is proposed. This architecture is designed as an embedded mode dispatch module. On the basis of this module, some unnecessary modes can be skipped or the mode decision process can be terminated in advanced. In order to maintain a higher compatibility, the FMD algorithms are unitedly designed as an unique module that can be easily embedded into a common video codec for H.265/HEVC. The input and output interfaces between the proposed module and other parts of the codec are designed based on simple but effective protocol. Hardware synthesis results on FPGA demonstrate that the proposed architecture achieves a maximum frequency of about 193 MHz with less than 1% of the total resources consumed. Moreover, the proposed module can improve the overall throughput.

  • Generating Accurate Candidate Windows by Effective Receptive Field

    Baojun ZHAO  Boya ZHAO  Linbo TANG  Baoxian WANG  

     
    LETTER-Image

      Vol:
    E102-A No:12
      Page(s):
    1925-1927

    Towards involving the convolutional neural networks into the object detection field, many computer vision tasks have achieved favorable successes. In order to adapt targets with various scales, deep feature pyramid is widely used, since the traditional object detection methods detect different objects in Gaussian image pyramid. However, due to the mismatching between the anchors and the feature distributions of targets, the accurate detection for targets with various scales is still a challenge. Considering the differences between the theoretical receptive field and effective receptive field, we propose a novel anchor generation method, which takes the effective receptive field as the standard. The proposed method is evaluated on the PASCAL VOC dataset and shows the favorable results.

  • BCGL: Binary Classification-Based Graph Layout

    Kai YAN  Tiejun ZHAO  Muyun YANG  

     
    PAPER-Computer Graphics

      Pubricized:
    2022/05/30
      Vol:
    E105-D No:9
      Page(s):
    1610-1619

    Graph layouts reveal global or local structures of graph data. However, there are few studies on assisting readers in better reconstructing a graph from a layout. This paper attempts to generate a layout whose edges can be reestablished. We reformulate the graph layout problem as an edge classification problem. The inputs are the vertex pairs, and the outputs are the edge existences. The trainable parameters are the laid-out coordinates of the vertices. We propose a binary classification-based graph layout (BCGL) framework in this paper. This layout aims to preserve the local structure of the graph and does not require the total similarity relationships of the vertices. We implement two concrete algorithms under the BCGL framework, evaluate our approach on a wide variety of datasets, and draw comparisons with several other methods. The evaluations verify the ability of the BCGL in local neighborhood preservation and its visual quality with some classic metrics.

  • Syntax-Based Context Representation for Statistical Machine Translation

    Kehai CHEN  Tiejun ZHAO  Muyun YANG  

     
    PAPER-Natural Language Processing

      Pubricized:
    2018/08/24
      Vol:
    E101-D No:12
      Page(s):
    3226-3237

    Learning semantic representation for translation context is beneficial to statistical machine translation (SMT). Previous efforts have focused on implicitly encoding syntactic and semantic knowledge in translation context by neural networks, which are weak in capturing explicit structural syntax information. In this paper, we propose a new neural network with a tree-based convolutional architecture to explicitly learn structural syntax information in translation context, thus improving translation prediction. Specifically, we first convert parallel sentences with source parse trees into syntax-based linear sequences based on a minimum syntax subtree algorithm, and then define a tree-based convolutional network over the linear sequences to learn syntax-based context representation and translation prediction jointly. To verify the effectiveness, the proposed model is integrated into phrase-based SMT. Experiments on large-scale Chinese-to-English and German-to-English translation tasks show that the proposed approach can achieve a substantial and significant improvement over several baseline systems.

  • An Improved Closed-Form Method for Moving Source Localization Using TDOA, FDOA, Differential Doppler Rate Measurements

    Zhixin LIU  Dexiu HU  Yongsheng ZHAO  Yongjun ZHAO  

     
    PAPER-Sensing

      Pubricized:
    2018/12/03
      Vol:
    E102-B No:6
      Page(s):
    1219-1228

    This paper proposes an improved closed-form method for moving source localization using time difference of arrival (TDOA), frequency difference of arrival (FDOA) and differential Doppler rate measurements. After linearizing the measurement equations by introducing three additional parameters, a rough estimate is obtained by using the weighted least-square (WLS) estimator. To further refine the estimate, the relationship between additional parameters and source location is utilized. The proposed method gives a final closed-form solution without iteration or the extra mathematics operations used in existing methods by employing the basic idea of WLS processing. Numerical examples show that the proposed method exhibits better robustness and performance compared with several existing methods.

  • Biped: Bidirectional Prediction of Order Violations

    Xi CHANG  Zhuo ZHANG  Yan LEI  Jianjun ZHAO  

     
    PAPER-Software Engineering

      Pubricized:
    2014/10/29
      Vol:
    E98-D No:2
      Page(s):
    334-345

    Concurrency bugs do significantly affect system reliability. Although many efforts have been made to address this problem, there are still many bugs that cannot be detected because of the complexity of concurrent programs. Compared with atomicity violations, order violations are always neglected. Efficient and effective approaches to detecting order violations are therefore in urgent need. This paper presents a bidirectional predictive trace analysis approach, BIPED, which can detect order violations in parallel based on a recorded program execution. BIPED collects an expected-order execution trace into a layered bidirectional prediction model, which intensively represents two types of expected-order data flows in the bottom layer and combines the lock sets and the bidirectionally order constraints in the upper layer. BIPED then recognizes two types of candidate violation intervals driven by the bottom-layer model and then checks these recognized intervals bidirectionally based on the upper-layer constraint model. Consequently, concrete schedules can be generated to expose order violation bugs. Our experimental results show that BIPED can effectively detect real order violation bugs and the analysis speed is 2.3x-10.9x and 1.24x-1.8x relative to the state-of-the-art predictive dynamic analysis approaches and hybrid model based static prediction analysis approaches in terms of order violation bugs.

  • The Effect of Transverse Magnetic Field on Making & Breaking Arc Durations of Electrical Contact

    Yun LIU  Guangda XU  Laijun ZHAO  Zhenbiao LI  Makoto HASEGAWA  

     
    PAPER

      Vol:
    E95-C No:9
      Page(s):
    1481-1486

    Application of transverse magnetic field (TMF) is one of the most important ways to improve electric life and breaking capacity of DC relays. For better understanding of dependence of arc durations on transverse magnetic field, a series of experiments were conducted under an external transverse magnetic field with 12 pairs of AgSnO2 contacts in a DC 28 V 60 A/30 A/5 A circuit, respectively. By using permanent magnets, the transverse magnetic field was obtained and the magnetic flux density at the gap center was varied from 13 to 94 mT. The results show that breaking arc duration is decreased monotonically with increases in the magnetic flux density, but making arc duration isn't decreased monotonically with increases in the magnetic flux density. In addition, both the magnetic flux density and the breaking arc duration have threshold values Bl and Tbmin, respectively, which means the breaking arc duration is almost stable with the value Tbmin even if the magnetic flux density is higher than Bl.

  • Wideband Beamforming for Multipath Signals Based on Spatial Smoothing Method

    Chengcheng LIU  Dexiu HU  Yongjun ZHAO  

     
    BRIEF PAPER-Electromagnetic Theory

      Vol:
    E97-C No:11
      Page(s):
    1130-1133

    In this paper, the spatial smoothing (SS) method is extended to the wideband multipath case. By reordering the array input signal and the weight vector, the corresponding covariance matrix of each subarray can be constructed conveniently. Then, a novel wideband beamforming algorithm, based on the SS method (SS-WB), can be achieved by linearly constrained minimum variance (LCMV). Further improvement of the output signal-to-interference-plus-noise ratio (SINR) for SS-WB can be obtained by removing the desired signal in the observed array data with the reconstruction of covariance matrix, which is denoted as wideband beamformer based on modified SS method (MSS-WB). Both proposed algorithms can reduce the desired signal cancellation due to the super decorrelation ability of SS method and MSS-WB can lead to a significantly improved output SINR. The simulations verify their effectiveness in the multipath environment.

  • SPDebugger: A Fine-Grained Deterministic Debugger for Concurrency Code

    Ziyi LIN  Yilei ZHOU  Hao ZHONG  Yuting CHEN  Haibo YU  Jianjun ZHAO  

     
    PAPER-Software Engineering

      Pubricized:
    2016/12/20
      Vol:
    E100-D No:3
      Page(s):
    473-482

    When debugging bugs, programmers often prepare test cases to reproduce buggy behaviours. However, for concurrent programs, test cases alone are typically insufficient to reproduce buggy behaviours, due to the nondeterminism of multi-threaded executions. In literature, various approaches have been proposed to reproduce buggy behaviours for concurrency bugs deterministically, but to the best of our knowledge, they are still limited. In particular, we have recognized three debugging scenarios from programming practice, but existing approaches can handle only one of the scenarios. In this paper, we propose a novel approach, called SPDebugger, that provides finer-grained thread controlling over test cases, programs under test, and even third party library code, to reproduce the predesigned thread execution schedule. The evaluation shows that SPDebugger handles more debugging scenarios than the state-of-the-art tool, called IMUnit, with similar human effort.

1-20hit(23hit)

FlyerIEICE has prepared a flyer regarding multilingual services. Please use the one in your native language.