Takao WAHO Akihisa KOYAMA Hitoshi HAYASHI
Signal processing using delta-sigma modulated bit streams is reviewed, along with related topics in stochastic computing (SC). The basic signal processing circuits, adders and multipliers, are covered. In particular, the possibility of preserving the noise-shaping properties inherent in delta-sigma modulation during these operations is discussed. Finally, the root mean square error for addition and multiplication is evaluated, and the performance improvement of signal processing in the delta-sigma domain compared with SC is verified.
Yosuke IIJIMA Atsunori OKADA Yasushi YUMINAKA
In high-speed data communication systems, it is important to evaluate the quality of the transmitted signal at the receiver. At a high-speed data rate, the transmission line characteristics act as a high-frequency attenuator and contribute to the intersymbol interference (ISI) at the receiver. To evaluate ISI conditions, eye diagrams are widely used to analyze signal quality and visualize the ISI effect as an eye-opening rate. Various types of on-chip eye-opening monitors (EOM) have been proposed to adjust waveform-shaping circuits. However, the eye diagram evaluation of multi-valued signaling becomes more difficult than that of binary transmission because of the complicated signal transition patterns. Moreover, in severe ISI situations where the eye is completely closed, eye diagram evaluation does not work well. This paper presents a novel evaluation method using Two-dimensional(2D) symbol mapping and a linear mixture model (LMM) for multi-valued data transmission. In our proposed method, ISI evaluation can be realized by 2D symbol mapping, and an efficient quantitative analysis can be realized using the LMM. An experimental demonstration of four leveled pulse amplitude modulation(PAM-4) data transmission using a Cat5e cable 100 m is presented. The experimental results show that the proposed method can extract features of the ISI effect even though the eye is completely closed in the server condition.
Yasushi YUMINAKA Kazuharu NAKAJIMA Yosuke IIJIMA
This study investigates a two/three-dimensional (2D/3D) symbol-mapping technique that evaluates data transmission quality based on a four-level pulse-amplitude modulation (PAM-4) symbol transition. Multi-dimensional symbol transition mapping facilitates the visualization of the degree of interference (ISI). The simulation and experimental results demonstrated that the 2D symbol mapping can evaluate the PAM-4 data transmission quality degraded by ISI and visualize the equalization effect. Furthermore, potential applications of 2D mapping and its extension to 3D mapping were explored.
Hua HUANG Yiwen SHAN Chuan LI Zhi WANG
Image denoising is an indispensable process of manifold high level tasks in image processing and computer vision. However, the traditional low-rank minimization-based methods suffer from a biased problem since only the noisy observation is used to estimate the underlying clean matrix. To overcome this issue, a new low-rank minimization-based method, called nuclear norm minus Frobenius norm rank residual minimization (NFRRM), is proposed for image denoising. The propose method transforms the ill-posed image denoising problem to rank residual minimization problems through excavating the nonlocal self-similarity prior. The proposed NFRRM model can perform an accurate estimation to the underlying clean matrix through treating each rank residual component flexibly. More importantly, the global optimum of the proposed NFRRM model can be obtained in closed-form. Extensive experiments demonstrate that the proposed NFRRM method outperforms many state-of-the-art image denoising methods.
Shiyu YANG Tetsuya KANDA Daniel M. GERMAN Yoshiki HIGO
Stack Overflow, a leading Q&A platform for developers, is a substantial reservoir of Python code snippets. Nevertheless, the incompatibility issues between Python versions, particularly Python 2 and Python 3, introduce substantial challenges that can potentially jeopardize the utility of these code snippets. This empirical study dives deep into the challenges of Python version inconsistencies on the interpretation and application of Python code snippets on Stack Overflow. Our empirical study exposes the prevalence of Python version compatibility issues on Stack Overflow. It further emphasizes an apparent deficiency in version-specific identification, a critical element that facilitates the identification and utilization of Python code snippets. These challenges, primarily arising from the lack of backward compatibility between Python’s major versions, pose significant hurdles for developers relying on Stack Overflow for code references and learning. This study, therefore, signifies the importance of proactively addressing these compatibility issues in Python code snippets. It advocates for enhanced tools and strategies to assist developers in efficiently navigating through the Python version complexities on platforms like Stack Overflow. By highlighting these concerns and providing a potential remedy, we aim to contribute to a more efficient and effective programming experience on Stack Overflow and similar platforms.
At Crypto 2019, Gohr first adopted the neural distinguisher for differential cryptanalysis, and since then, this work received increasing attention. However, most of the existing work focuses on improving and applying the neural distinguisher, the studies delving into the intrinsic principles of neural distinguishers are finite. At Eurocrypt 2021, Benamira et al. conducted a study on Gohr’s neural distinguisher. But for the neural distinguishers proposed later, such as the r-round neural distinguishers trained with k ciphertext pairs or ciphertext differences, denoted as NDcpk_r (Gohr’s neural distinguisher is the special NDcpk_r with K = 1) and NDcdk_r , such research is lacking. In this work, we devote ourselves to study the intrinsic principles and relationship between NDcdk_r and NDcpk_r. Firstly, we explore the working principle of NDcd1_r through a series of experiments and find that it strongly relies on the probability distribution of ciphertext differences. Its operational mechanism bears a strong resemblance to that of NDcp1_r given by Benamira et al.. Therefore, we further compare them from the perspective of differential cryptanalysis and sample features, demonstrating the superior performance of NDcp1_r can be attributed to the relationships between certain ciphertext bits, especially the significant bits. We then extend our investigation to NDcpk_r, and show that its ability to recognize samples heavily relies on the average differential probability of k ciphertext pairs and some relationships in the ciphertext itself, but the reliance between k ciphertext pairs is very weak. Finally, in light of the findings of our research, we introduce a strategy to enhance the accuracy of the neural distinguisher by using a fixed difference to generate the negative samples instead of the random one. Through the implementation of this approach, we manage to improve the accuracy of the neural distinguishers by approximately 2% to 8% for 7-round Speck32/64 and 9-round Simon32/64.
Data sparsity has always been a problem in document classification, for which semi-supervised learning and few-shot learning are studied. An even more extreme scenario is to classify documents without any annotated data, but using only category names. In this paper, we introduce a nearest neighbor search-based method Con2Class to tackle this tough task. We intend to produce embeddings for predefined categories and predict category embeddings for all the unlabeled documents in a unified embedding space, such that categories can be easily assigned by searching the nearest predefined category in the embedding space. To achieve this, we propose confidence-driven contrastive learning, in which prompt-based templates are designed and MLM-maintained contrastive loss is newly proposed to finetune a pretrained language model for embedding production. To deal with the issue that no annotated data is available to validate the classification model, we introduce confidence factor to estimate the classification ability by evaluating the prediction confidence. The language model having the highest confidence factor is used to produce embeddings for similarity evaluation. Pseudo labels are then assigned by searching the semantically closest category name, which are further used to train a separate classifier following a progressive self-training strategy for final prediction. Our experiments on five representative datasets demonstrate the superiority of our proposed method over the existing approaches.
Hyebong CHOI Joel SHIN Jeongho KIM Samuel YOON Hyeonmin PARK Hyejin CHO Jiyoung JUNG
The design of automobile lamps requires accurate estimation of heat distribution to prevent overheating and deformation of the product. Traditional heat resistant analysis using Computational Fluid Dynamics (CFD) is time-consuming and requires expertise in thermofluid mechanics, making real-time temperature analysis less accessible to lamp designers. We propose a machine learning-based temperature prediction system for automobile lamp design. We trained our machine learning models using CFD results of various lamp designs, providing lamp designers real-time Heat-Resistant Analysis. Comprehensive tests on real lamp products demonstrate that our prediction model accurately estimates heat distribution comparable to CFD analysis within a minute. Our system visualizes the estimated heat distribution of car lamp design supporting quick decision-making by lamp designer. It is expected to shorten the product design process, improving the market competitiveness.
Haoran LUO Tengfei SHAO Shenglei LI Reiko HISHIYAMA
Makeup transfer is the process of applying the makeup style from one picture (reference) to another (source), allowing for the modification of characters’ makeup styles. To meet the diverse makeup needs of individuals or samples, the makeup transfer framework should accurately handle various makeup degrees, ranging from subtle to bold, and exhibit intelligence in adapting to the source makeup. This paper introduces a “3-level” adaptive makeup transfer framework, addressing facial makeup through two sub-tasks: 1. Makeup adaptation, utilizing feature descriptors and eyelid curve algorithms to classify 135 organ-level face shapes; 2. Makeup transfer, achieved by learning the reference picture from three branches (color, highlight, pattern) and applying it to the source picture. The proposed framework, termed “Face Shape Adaptive Makeup Transfer” (FSAMT), demonstrates superior results in makeup transfer output quality, as confirmed by experimental results.
Shuto HASEGAWA Koichiro ENOMOTO Taeko MIZUTANI Yuri OKANO Takenori TANAKA Osamu SAKAI
Melanin, which is responsible for the appearance of spots and freckles, is an important indicator in evaluating skin condition. To assess the efficacy of cosmetics, skin condition scoring is performed by analyzing the distribution and amount of melanin from microscopic images of the stratum corneum cells. However, the current practice of diagnosing skin condition using stratum corneum cells images relies heavily on visual evaluation by experts. The goal of this study is to develop a quantitative evaluation system for skin condition based on melanin within unstained stratum corneum cells images. The proposed system utilizes principal component regression to perform five-level scoring, which is then compared with visual evaluation scores to assess the system’s usefulness. Additionally, we evaluated the impact of indicators related to melanin obtained from images on the scores, and verified which indicators are effective for evaluation. In conclusion, we confirmed that scoring is possible with an accuracy of more than 60% on a combination of several indicators, which is comparable to the accuracy of visual assessment.
This paper presents MDX-Mixer, which improves music demixing (MDX) performance by leveraging source signals separated by multiple existing MDX models. Deep-learning-based MDX models have improved their separation performances year by year for four kinds of sound sources: “vocals,” “drums,” “bass,” and “other”. Our research question is whether mixing (i.e., weighted sum) the signals separated by state-of-the-art MDX models can obtain either the best of everything or higher separation performance. Previously, in singing voice separation and MDX, there have been studies in which separated signals of the same sound source are mixed with each other using time-invariant or time-varying positive mixing weights. In contrast to those, this study is novel in that it allows for negative weights as well and performs time-varying mixing using all of the separated source signals and the music acoustic signal before separation. The time-varying weights are estimated by modeling the music acoustic signals and their separated signals by dividing them into short segments. In this paper we propose two new systems: one that estimates time-invariant weights using 1×1 convolution, and one that estimates time-varying weights by applying the MLP-Mixer layer proposed in the computer vision field to each segment. The latter model is called MDX-Mixer. Their performances were evaluated based on the source-to-distortion ratio (SDR) using the well-known MUSDB18-HQ dataset. The results show that the MDX-Mixer achieved higher SDR than the separated signals given by three state-of-the-art MDX models.
With the rapid advancement of graphics processing units (GPUs), Virtual Reality (VR) experiences have significantly improved, enhancing immersion and realism. However, these advancements also raise security concerns in VR. In this paper, I introduce a new attack leveraging known WebVR vulnerabilities to track the activities of VR users. The proposed attack leverages the user’s hand motion information exposed to web attackers, demonstrating the capability to identify consumed content, such as 3D images and videos, and pilfer private drawings created in a 3D drawing app. To achieve this, I employed a machine learning approach to process controller sensor data and devised techniques to extract sensitive activities during the use of target apps. The experimental results demonstrate that the viewed content in the targeted content viewer can be identified with 90% accuracy. Furthermore, I successfully obtained drawing outlines that precisely match the user’s original drawings without performance degradation, validating the effectiveness of the attack.
Yuto ARIMURA Shigeru YAMASHITA
Stochastic Computing (SC) allows additions and multiplications to be realized with lower power than the conventional binary operations if we admit some errors. However, for many complex functions which cannot be realized by only additions and multiplications, we do not know a generic efficient method to calculate a function by using an SC circuit; it is necessary to realize an SC circuit by using a generic method such as polynomial approximation methods for such a function, which may lose the advantage of SC. Thus, there have been many researches to consider efficient SC realization for specific functions; an efficient SC square root circuit with a feedback circuit was proposed by D. Wu et al. recently. This paper generalizes the SC square root circuit with a feedback circuit; we identify a situation when we can implement a function efficiently by an SC circuit with a feedback circuit. As examples of our generalization, we propose SC circuits to calculate the n-th root calculation and division. We also show our analysis on the accuracy of our SC circuits and the hardware costs; our results show the effectiveness of our method compared to the conventional SC designs; our framework may be able to implement a SC circuit that is better than the existing methods in terms of the hardware cost or the calculation error.
Kaoru TAKEMURE Yusuke SAKAI Bagus SANTOSO Goichiro HANAOKA Kazuo OHTA
The existing discrete-logarithm-based two-round multi-signature schemes without using the idealized model, i.e., the Algebraic Group Model (AGM), have quite large reduction loss. This means that an implementation of these schemes requires an elliptic curve (EC) with a very large order for the standard 128-bit security when we consider concrete security. Indeed, the existing standardized ECs have orders too small to ensure 128-bit security of such schemes. Recently, Pan and Wagner proposed two two-round schemes based on the Decisional Diffie-Hellman (DDH) assumption (EUROCRYPT 2023). For 128-bit security in concrete security, the first scheme can use the NIST-standardized EC P-256 and the second can use P-384. However, with these parameter choices, they do not improve the signature size and the communication complexity over the existing non-tight schemes. Therefore, there is no two-round scheme that (i) can use a standardized EC for 128-bit security and (ii) has high efficiency. In this paper, we construct a two-round multi-signature scheme achieving both of them from the DDH assumption. We prove that an EC with at least a 321-bit order is sufficient for our scheme to ensure 128-bit security. Thus, we can use the NIST-standardized EC P-384 for 128-bit security. Moreover, the signature size and the communication complexity per one signer of our proposed scheme under P-384 are 1152 bits and 1535 bits, respectively. These are most efficient among the existing two-round schemes without using the AGM including Pan-Wagner’s schemes and non-tight schemes which do not use the AGM. Our experiment on an ordinary machine shows that for signing and verification, each can be completed in about 65 ms under 100 signers. This shows that our scheme has sufficiently reasonable running time in practice.
This paper proposes a scheme for reducing pilot interference in cell-free massive multiple-input multiple-output (MIMO) systems through scalable access point (AP) selection and efficient pilot allocation using the Grey Wolf Optimizer (GWO). Specifically, we introduce a bidirectional large-scale fading-based (B-LSFB) AP selection method that builds high-quality connections benefiting both APs and UEs. Then, we limit the number of UEs that each AP can serve and encourage competition among UEs to improve the scalability of this approach. Additionally, we propose a grey wolf optimization based pilot allocation (GWOPA) scheme to minimize pilot contamination. Specifically, we first define a fitness function to quantify the level of pilot interference between UEs, and then construct dynamic interference relationships between any UE and its serving AP sets using a weighted fitness function to minimize pilot interference. The simulation results shows that the B-LSFB strategy achieves scalability with performance similar to large-scale fading-based (LSFB) AP selection. Furthermore, the grey wolf optimization-based pilot allocation scheme significantly improves per-user net throughput with low complexity compared to four existing schemes.
Artificial intelligence and the introduction of Internet of Things technologies have benefited from technological advances and new automated computer system technologies. Eventually, it is now possible to integrate them into a single offline industrial system. This is accomplished through machine-to-machine communication, which eliminates the human factor. The purpose of this article is to examine security systems for machine-to-machine communication systems that rely on identification and authentication algorithms for real-time monitoring. The article investigates security methods for quickly resolving data processing issues by using the Security operations Center’s main machine to identify and authenticate devices from 19 different machines. The results indicate that when machines are running offline and performing various tasks, they can be exposed to data leaks and malware attacks by both the individual machine and the system as a whole. The study looks at the operation of 19 computers, 7 of which were subjected to data leakage and malware attacks. AnyLogic software is used to create visual representations of the results using wireless networks and algorithms based on previously processed methods. The W76S is used as a protective element within intelligent sensors due to its built-in memory protection. For 4 machines, the data leakage time with malware attacks was 70 s. For 10 machines, the duration was 150 s with 3 attacks. Machine 15 had the longest attack duration, lasting 190 s and involving 6 malware attacks, while machine 19 had the shortest attack duration, lasting 200 s and involving 7 malware attacks. The highest numbers indicated that attempting to hack a system increased the risk of damaging a device, potentially resulting in the entire system with connected devices failing. Thus, illegal attacks by attackers using malware may be identified over time, and data processing effects can be prevented by intelligent control. The results reveal that applying identification and authentication methods using a protocol increases cyber-physical system security while also allowing real-time monitoring of offline system security.
Xiangyu LI Ping RUAN Wei HAO Meilin XIE Tao LV
To achieve precise measurement without landing, the high-mobility vehicle-mounted theodolite needs to be leveled quickly with high precision and ensure sufficient support stability before work. After the measurement, it is also necessary to ensure that the high-mobility vehicle-mounted theodolite can be quickly withdrawn. Therefore, this paper proposes a hierarchical automatic leveling strategy and establishes a two-stage electromechanical automatic leveling mechanism model. Using coarse leveling of the first-stage automatic leveling mechanism and fine leveling of the second-stage automatic leveling mechanism, the model realizes high-precision and fast leveling of the vehicle-mounted theodolites. Then, the leveling control method based on repeated positioning is proposed for the first-stage automatic leveling mechanism. To realize the rapid withdrawal for high-mobility vehicle-mounted theodolites, the method ensures the coincidence of spatial movement paths when the structural parts are unfolded and withdrawn. Next, the leg static balance equation is constructed in the leveling state, and the support force detection method is discussed in realizing the stable support for vehicle-mounted theodolites. Furthermore, a mathematical model for “false leg” detection is established furtherly, and a “false leg” detection scheme based on the support force detection method is analyzed to significantly improve the support stability of vehicle-mounted theodolites. Finally, an experimental platform is constructed to perform the performance test for automatic leveling mechanisms. The experimental results show that the leveling accuracy of established two-stage electromechanical automatic leveling mechanism can reach 3.6″, and the leveling time is no more than 2 mins. The maximum support force error of the support force detection method is less than 15%, and the average support force error is less than 10%. In contrast, the maximum support force error of the drive motor torque detection method reaches 80.12%, and its leg support stability is much less than the support force detection method. The model and analysis method proposed in this paper can also be used for vehicle-mounted radar, vehicle-mounted laser measurement devices, vehicle-mounted artillery launchers and other types of vehicle-mounted equipment with high-precision and high-mobility working requirements.
Pengxu JIANG Yang YANG Yue XIE Cairong ZOU Qingyun WANG
Convolutional neural network (CNN) is widely used in acoustic scene classification (ASC) tasks. In most cases, local convolution is utilized to gather time-frequency information between spectrum nodes. It is challenging to adequately express the non-local link between frequency domains in a finite convolution region. In this paper, we propose a dual-path convolutional neural network based on band interaction block (DCNN-bi) for ASC, with mel-spectrogram as the model’s input. We build two parallel CNN paths to learn the high-frequency and low-frequency components of the input feature. Additionally, we have created three band interaction blocks (bi-blocks) to explore the pertinent nodes between various frequency bands, which are connected between two paths. Combining the time-frequency information from two paths, the bi-blocks with three distinct designs acquire non-local information and send it back to the respective paths. The experimental results indicate that the utilization of the bi-block has the potential to improve the initial performance of the CNN substantially. Specifically, when applied to the DCASE 2018 and DCASE 2020 datasets, the CNN exhibited performance improvements of 1.79% and 3.06%, respectively.
Ze Fu GAO Wen Ge YANG Yi Wen JIAO
Space is becoming increasingly congested and contested, which calls for effective means to conduct effective monitoring of high-value space assets, especially in Space Situational Awareness (SSA) missions, while there are imperfections in existing methods and corresponding algorithms. To overcome such a problem, this letter proposes an algorithm for accurate Connected Element Interferometry (CEI) in SSA based on more interpolation information and iterations. Simulation results show that: (i) after iterations, the estimated asymptotic variance of the proposed method can basically achieve uniform convergence, and the ratio of it to ACRB is 1.00235 in δ0 ∈ [-0.5, 0.5], which is closer to 1 than the current best AM algorithms; (ii) In the interval of SNR ∈ [-14dB, 0dB], the estimation error of the proposed algorithm decreases significantly, which is basically comparable to CRLB (maintains at 1.236 times). The research of this letter could play a significant role in effective monitoring and high-precision tracking and measurement with significant space targets during futuristic SSA missions.
Yi Wen JIAO Ze Fu GAO Wen Ge YANG
In future deep space communication missions, VLBI (Very Long Baseline Interferometry) based on antenna array technology remains a critical detection method, which urgently requires the improvement of synthesis performance for antenna array signals. Considering this, focusing on optimizing the traditional antenna grouping method applied in the phase estimation algorithm, this letter proposes a “L/2 to L/2” antenna grouping method based on the maximum correlation signal-to-noise ratio (SNR). Following this idea, a phase difference estimation algorithm named “Couple” is presented. Theoretical analysis and simulation verification illustrate that: when ρ < -10dB, the proposed “Couple” has the highest performance; increasing the number of antennas can significantly improve its synthetic loss performance and robustness. The research of this letter indicates a promising potential in supporting the rising deep space exploration and communication missions.