A domain decomposition method is widely utilized for analyzing large-scale electromagnetic problems. The method decomposes the target model into small independent subdomains. An electromagnetic analysis has inherently suffers from late convergence analyzed with iterative algorithms such as Krylov subspace algorithms. The DDM remedies this issue by decomposing the total system into subdomain problems and gathering the local results as an interface problem to adjust to achieve the total solution. In this paper we report the convergence properties of the domain decomposition method while modifying the size of local domain and the region shape on several mesh sizes. As experimental results show, the convergence speed depends on the number of interface problem variables and the selection of the local region shapes. In addition to that the convergence property differs according to the target frequencies. In general it is demonstrated that the convergence speed can be accelerated with large cubic subdomain shape. We propose the subdomain selection strategies based on the analysis of the condition numbers of the governing equation.
Hakan BERCAG Osman KUKRER Aykut HOCANIN
A new extended normalized least-mean-square (ENLMS) algorithm is proposed. A novel non-linear time-varying step-size (NLTVSS) formula is derived. The convergence rate of ENLMS increases due to NLTVSS as the number of data-reuse L is increased. ENLMS does not involve matrix inversion, and, thus, avoids numerical instability issues.
The steady-state and convergence performances are important indicators to evaluate adaptive algorithms. The step-size affects these two important indicators directly. Many relevant scholars have also proposed some variable step-size adaptive algorithms for improving performance. However, there are still some problems in these existing variable step-size adaptive algorithms, such as the insufficient theoretical analysis, the imbalanced performance and the unachievable parameter. These problems influence the actual performance of some algorithms greatly. Therefore, we intend to further explore an inherent relationship between the key performance and the step-size in this paper. The variation of mean square deviation (MSD) is adopted as the cost function. Based on some theoretical analyses and derivations, a novel variable step-size algorithm with a dynamic limited function (DLF) was proposed. At the same time, the sufficient theoretical analysis is conducted on the weight deviation and the convergence stability. The proposed algorithm is also tested with some typical algorithms in many different environments. Both the theoretical analysis and the experimental result all have verified that the proposed algorithm equips a superior performance.
Runde YU Zhuowen LI Zhe CHEN Gangyi DING
In order to solve the problems of copyrights infringement, high cost and complex process of rights protection in current media convergence center, a digital rights management system based on blockchain technology and IPFS (Inter Planetary File System) technology is proposed. Considering that large files such as video and audio cannot be stored on the blockchain directly, IPFS technology is adopted as the data expansion scheme for the data storage layer of the Ethereum platform, IPFS protocol is further used for distributed data storage and transmission of media content. In addition, smart contract is also used to uniquely identify digital rights through NFT (Non-fungible Tokens), which provides the characteristics of digital rights transferability and traceability, and realizes an open, transparent, tamper-proof and traceable digital rights management system for media convergence center. Several experimental results show that it has higher transaction success rate, lower storage consumption and transaction confirmation delay than existing scheme.
Wenkai LIU Cuizhu QIN Menglong WU Wenle BAI Hongxia DONG
Pose estimation is a research hot spot in computer vision tasks and the key to computer perception of human activities. The core concept of human pose estimation involves describing the motion of the human body through major joint points. Large receptive fields and rich spatial information facilitate the keypoint localization task, and how to capture features on a larger scale and reintegrate them into the feature space is a challenge for pose estimation. To address this problem, we propose a multi-scale convergence network (MSCNet) with a large receptive field and rich spatial information. The structure of the MSCNet is based on an hourglass network that captures information at different scales to present a consistent understanding of the whole body. The multi-scale receptive field (MSRF) units provide a large receptive field to obtain rich contextual information, which is then selectively enhanced or suppressed by the Squeeze-Excitation (SE) attention mechanism to flexibly perform the pose estimation task. Experimental results show that MSCNet scores 73.1% AP on the COCO dataset, an 8.8% improvement compared to the mainstream CMUPose method. Compared to the advanced CPN, the MSCNet has 68.2% of the computational complexity and only 55.4% of the number of parameters.
This paper proposes an efficient scheduling algorithm for the layered decoding of block low-density parity-check (LDPC) codes. To efficiently configure check node-based scheduling groups, the proposed algorithm utilizes the base matrix of the block LDPC code for a block-by-block scheduling group configuration; i.e., the proposed algorithm generates a scheduling group of check nodes, satisfying the weight condition of the layered decoding, which is performed in block units (including several check nodes). Therefore, unlike the conventional scheduling algorithms performed in node units, the proposed algorithm can efficiently generate scheduling groups for layered decoding at low computational complexity and memory requirements. In addition, to accelerate the decoding convergence speed, check nodes are allocated in each scheduling group such that messages from check nodes up to the current group are delivered as evenly as possible to bit nodes. Simulation results confirm that the proposed algorithm can accelerate decoding convergence compared to other block-based scheduling algorithms for layered decoding of block LDPC codes.
The hybrid implicit-explicit single-field finite-difference time-domain (HIE-SF-FDTD) method based on the wave equation of electric field is reformulated in a concise matrix-vector form. The global approximation error of the scheme is discussed theoretically. The second-order convergence of the HIE-SF-FDTD is numerically verified.
Numerous variable tap-length algorithms can be found in some literature and few strategies are derived from a basic theoretical formula. Thus, some algorithms lack of theoretical depth and their performance are unstable. In view of this point, the novel variable tap-length algorithm which is based on the mixed error cost function is presented in this letter. By analyzing the mixed expectation of the prior and the posterior error, the novel variable tap-length strategy is derived. The proposed algorithm has a more valid proximity to the optimal tap-length and a good convergence ability by the performance analysis. It can solve many deficiencies comprising large fluctuations of the tap-length, the high complexity and the weak steady-state ability. Simulation results demonstrate that the proposed algorithm equips good performance.
Ryota YOSHIMURA Ichiro MARUTA Kenji FUJIMOTO Ken SATO Yusuke KOBAYASHI
Particle filters have been widely used for state estimation problems in nonlinear and non-Gaussian systems. Their performance depends on the given system and measurement models, which need to be designed by the user for each target system. This paper proposes a novel method to design these models for a particle filter. This is a numerical optimization method, where the particle filter design process is interpreted into the framework of reinforcement learning by assigning the randomnesses included in both models of the particle filter to the policy of reinforcement learning. In this method, estimation by the particle filter is repeatedly performed and the parameters that determine both models are gradually updated according to the estimation results. The advantage is that it can optimize various objective functions, such as the estimation accuracy of the particle filter, the variance of the particles, the likelihood of the parameters, and the regularization term of the parameters. We derive the conditions to guarantee that the optimization calculation converges with probability 1. Furthermore, in order to show that the proposed method can be applied to practical-scale problems, we design the particle filter for mobile robot localization, which is an essential technology for autonomous navigation. By numerical simulations, it is demonstrated that the proposed method further improves the localization accuracy compared to the conventional method.
Ryota ISHIKAWA Masashi TAWADA Masao YANAGISAWA Nozomu TOGAWA
Recently, stochastic computing based on stochastic numbers attracts attention as an effective computation method, which realizes arithmetic operations by simple logic circuits with a tolerance of bit errors. When we input two or more identical values to a stochastic circuit, we require to duplicate a stochastic number. However, if bit streams of duplicated stochastic numbers are dependent on each other, their arithmetic operation results can be inaccurate. In this paper, we propose two stochastic number duplicators, called FSR and RRR. The stochastic numbers duplicated by the FSR and RRR duplicators have the equivalent values but have independent bit streams, effectively utilizing bit re-arrangement using randomized bit streams. Experimental evaluation results demonstrate that the RRR duplicator, in particular, obtains more accurate results even if a circuit has re-convergence paths, reducing the mean square errors by 20%-89% compared to a conventional stochastic number duplicator.
Nawfal AL-ZUBAIDI R-SMITH Lubomír BRANČÍK
Numerical inverse Laplace transform (NILT) methods are potential methods for time domain simulations, for instance the analysis of the transient phenomena in systems with lumped and/or distributed parameters. This paper proposes a numerical inverse Laplace transform method based originally on hyperbolic relations. The method is further enhanced by properly adapting several convergence acceleration techniques, namely, the epsilon algorithm of Wynn, the quotient-difference algorithm of Rutishauser and the Euler transform. The resulting accelerated models are compared as for their accuracy and computational efficiency. Moreover, an expansion to two dimensions is presented for the first time in the context of the accelerated hyperbolic NILT method, followed by the error analysis. The expansion is done by repeated application of one-dimensional partial numerical inverse Laplace transforms. A detailed static error analysis of the resulting 2D NILT is performed to prove the effectivness of the method. The work is followed by a practical application of the 2D NILT method to simulate voltage/current distributions along a transmission line. The method and application are programmed using the Matlab language.
This paper discusses the concept of PON standards convergence. The history of PON standardization is reviewed in brief as a way to explain how the industry arrived at its current divergent form. The reasons why convergence is favorable are enumerated, with a focus on what has changed since the last round of standardization. Finally, some paths forward are proposed.
Shinya MOCHIDUKI Yuki YOKOYAMA Keigo SUKEGAWA Hiroki SATO Miyuki SUGANUMA Mitsuho YAMADA
In this study, we first developed a simultaneous measurement system for accommodation and convergence eye movement and evaluated its precision. Then, using a stuffed animal as the target, whose depth should be relatively easy to perceive, we measured convergence eye movement and accommodation at the same time while a tablet displaying a 3D movie was moved in the depth direction. By adding the real 3D display depth movement to the movement of the 3D image, subjects showed convergence eye movement that corresponds appropriately to the dual change of parallax in the 3D movie and real display, even when a subject's convergence changed very little. Accommodation also changed appropriately according to the change in depth.
Takumi KIMURA Norikazu TAKAHASHI
Nonnegative Matrix Factorization (NMF) with sparseness and smoothness constraints has attracted increasing attention. When these properties are considered, NMF is usually formulated as an optimization problem in which a linear combination of an approximation error term and some regularization terms must be minimized under the constraint that the factor matrices are nonnegative. In this paper, we focus our attention on the error measure based on the Euclidean distance and propose a new iterative method for solving those optimization problems. The proposed method is based on the Hierarchical Alternating Least Squares (HALS) algorithm developed by Cichocki et al. We first present an example to show that the original HALS algorithm can increase the objective value. We then propose a new algorithm called the Gauss-Seidel HALS algorithm that decreases the objective value monotonically. We also prove that it has the global convergence property in the sense of Zangwill. We finally verify the effectiveness of the proposed algorithm through numerical experiments using synthetic and real data.
Tso-Cho CHEN Erl-Huei LU Chia-Jung LI Kuo-Tsang HUANG
In this paper, a weighted multiple bit flipping (WMBF) algorithman for decoding low-density parity-check (LDPC) codes is proposed first. Then the improved WMBF algorithm which we call the efficient weighted bit-flipping (EWBF) algorithm is developed. The EWBF algorithm can dynamically choose either multiple bit-flipping or single bit-flipping in each iteration according to the log-likelihood ratio of the error probability of the received bits. Thus, it can efficiently increase the convergence speed of decoding and prevent the decoding process from falling into loop traps. Compared with the parallel weighted bit-flipping (PWBF) algorithm, the EWBF algorithm can achieve significantly lower computational complexity without performance degradation when the Euclidean geometry (EG)-LDPC codes are decoded. Furthermore, the flipping criterion does not require any parameter adjustment.
Bin YANG Yuliang LU Kailong ZHU Guozheng YANG Jingwei LIU Haibo YIN
The rapid development of information techniques has lead to more and more high-dimensional datasets, making classification more difficult. However, not all of the features are useful for classification, and some of these features may even cause low classification accuracy. Feature selection is a useful technique, which aims to reduce the dimensionality of datasets, for solving classification problems. In this paper, we propose a modified bat algorithm (BA) for feature selection, called MBAFS, using a SVM. Some mechanisms are designed for avoiding the premature convergence. On the one hand, in order to maintain the diversity of bats, they are guided by the combination of a random bat and the global best bat. On the other hand, to enhance the ability of escaping from local optimization, MBAFS employs one mutation mechanism while the algorithm trapped into local optima. Furthermore, the performance of MBAFS was tested on twelve benchmark datasets, and was compared with other BA based algorithms and some well-known BPSO based algorithms. Experimental results indicated that the proposed algorithm outperforms than other methods. Also, the comparison details showed that MBAFS is competitive in terms of computational time.
Yufei HAN Mingjiang WANG Boya ZHAO
Improved fractional variable tap-length adaptive algorithm that contains Sigmoid limited fluctuation function and adaptive variable step-size of tap-length based on fragment-full error is presented. The proposed algorithm can solve many deficiencies in previous algorithm, comprising small convergence rate and weak anti-interference ability. The parameters are able to modify reasonably on the basis of different situations. The Sigmoid constrained function can decrease the fluctuant amplitude of the instantaneous errors effectively and improves the ability of anti-noise interference. Simulations demonstrate that the proposed algorithm equips better performance.
Shrinkage widely linear recursive least squares (SWL-RLS) and its improved version called structured shrinkage widely linear recursive least squares (SSWL-RLS) algorithms are proposed in this paper. By using the relationship between the noise-free a posterior and a priori error signals, the optimal forgetting factor can be obtained at each snapshot. In the implementation of algorithms, due to the a priori error signal known, we still need the information about the noise-free a priori error which can be estimated with a known formula. Simulation results illustrate that the proposed algorithms have faster convergence and better tracking capability than augmented RLS (A-RLS), augmented least mean square (A-LMS) and SWL-LMS algorithms.
Kenji FUJIKAWA Hiroaki HARAI Motoyuki OHMORI Masataka OHTA
We have developed an automatic network configuration technology for flexible and robust network construction. In this paper, we propose a two-or-more-level hierarchical link-state routing protocol in Hierarchical QoS Link Information Protocol (HQLIP). The hierarchical routing easily scales up the network by combining and stacking configured networks. HQLIP is designed not to recompute shortest-path trees from topology information in order to achieve a high-speed convergence of forwarding information base (FIB), especially when renumbering occurs in the network. In addition, we propose a fixed-midfix renumbering (FMR) method. FMR enables an even faster convergence when HQLIP is synchronized with Hierarchical/Automatic Number Allocation (HANA). Experiments demonstrate that HQLIP incorporating FMR achieves the convergence time within one second in the network where 22 switches and 800 server terminals are placed, and is superior to Open Shortest Path First (OSPF) in terms of a convergence time. This shows that a combination of HQLIP and HANA performs stable renumbering in link-state routing protocol networks.
Almost sure convergence coding theorems of one-shot and multi-shot Tunstall codes are proved for stationary memoryless sources. Coding theorem of one-shot Tunstall code is proved in the case that the leaf count of Tunstall tree increases. On the other hand, coding theorem is proved for multi-shot Tunstall code with increasing parsing count, under the assumption that the Tunstall tree grows as the parsing proceeds. In this result, it is clarified that the theorem for the one-shot Tunstall code is not a corollary of the theorem for the multi-shot Tunstall code. In the case of the multi-shot Tunstall code, it can be regarded that the coding theorem is proved for the sequential algorithm such that parsing and coding are processed repeatedly. Cartesian concatenation of trees and geometric mean of the leaf counts of trees are newly introduced, which play crucial roles in the analyses of multi-shot Tunstall code.