Wei LU Weidong WANG Ergude BAO Liqiang WANG Weiwei XING Yue CHEN
Web Service Composition (WSC) has been well recognized as a convenient and flexible way of service sharing and integration in service-oriented application fields. WSC aims at selecting and composing a set of initial services with respect to the Quality of Service (QoS) values of their attributes (e.g., price), in order to complete a complex task and meet user requirements. A major research challenge of the QoS-aware WSC problem is to select a proper set of services to maximize the QoS of the composite service meeting several QoS constraints upon various attributes, e.g. total price or runtime. In this article, a fast algorithm based on QoS-aware sampling (FAQS) is proposed, which can efficiently find the near-optimal composition result from sampled services. FAQS consists of five steps as follows. 1) QoS normalization is performed to unify different metrics for QoS attributes. 2) The normalized services are sampled and categorized by guaranteeing similar number of services in each class. 3) The frequencies of the sampled services are calculated to guarantee the composed services are the most frequent ones. This process ensures that the sampled services cover as many as possible initial services. 4) The sampled services are composed by solving a linear programming problem. 5) The initial composition results are further optimized by solving a modified multi-choice multi-dimensional knapsack problem (MMKP). Experimental results indicate that FAQS is much faster than existing algorithms and could obtain stable near-optimal result.
Leida LI Yu ZHOU Jinjian WU Jiansheng QIAN Beijing CHEN
Image retouching is fundamental in photography, which is widely used to improve the perceptual quality of a low-quality image. Traditional image quality metrics are designed for degraded images, so they are limited in evaluating the quality of retouched images. This letter presents a RETouched Image QUality Evaluation (RETIQUE) algorithm by measuring structure and color changes between the original and retouched images. Structure changes are measured by gradient similarity. Color colorfulness and saturation are utilized to measure color changes. The overall quality score of a retouched image is computed as the linear combination of gradient similarity and color similarity. The performance of RETIQUE is evaluated on a public Digitally Retouched Image Quality (DRIQ) database. Experimental results demonstrate that the proposed metric outperforms the state-of-the-arts.
Waqas ur RAHMAN Kwangsue CHUNG
In this paper we propose an adaptive bitrate (ABR) algorithm that selects the video rates by observing and controlling the playback buffer. In a Hypertext Transfer Protocol (HTTP) adaptive streaming session, buffer dynamics largely depend on the chunk sizes. First, we present the algorithm that selects the next video rates based on the current buffer level, while considering the upcoming chunk sizes. We aim to exploit the variation of chunk sizes of a variable bitrate (VBR) encoded video to optimize the tradeoff between the video rate and buffer underflow events while keeping a low frequency of video rate changes. To evaluate the performance of the proposed algorithm, we consider three scenarios: (i) the video flow does not compete with any cross traffic, (ii) the video flow shares the bottleneck link with competing TCP traffic, and (iii) two HTTP clients share the bottleneck. We show that the proposed algorithm selects a high playback video rate and avoids unnecessary rebuffering while keeping a low frequency of video rate changes. Furthermore, we show that the proposed algorithm changes the video quality gradually to guarantee the user's viewing experience.
Jinling ZHOU Xingchun DIAO Jianjun CAO Zhisong PAN
Compared to the traditional functional dependency (FD), the extended conditional functional dependency (CFD) has shown greater potential for detecting and repairing inconsistent data. CFDMiner is a widely used algorithm for mining constant-CFDs. But the search space of CFDMiner is too large, and there is still room for efficiency improvement. In this paper, an efficient pruning strategy is proposed to optimize the algorithm by reducing the search space. Both theoretical analysis and experiments have proved the optimized algorithm can produce the consistent results as the original CFDMiner.
Chao WANG Xuanqin MOU Lei ZHANG
In lossy image/video encoding, there is a compromise between the number of bits and the extent of distortion. Optimizing the allocation of bits to different sources, such as frames or blocks, can improve the encoding performance. In intra-frame encoding, due to the dependency among macro blocks (MBs) introduced by intra prediction, the optimization of bit allocation to the MBs usually has high complexity. So far, no practical optimal bit allocation methods for intra-frame encoding exist, and the commonly used method for intra-frame encoding is the fixed-QP method. We suggest that the QP selection inside an image/a frame can be optimized by aiming at the constant perceptual quality (CPQ). We proposed an iteration-based bit allocation scheme for H.264/AVC intra-frame encoding, in which all the local areas (which is defined by a group of MBs (GOMBs) in this paper) in the frame are encoded to have approximately the same perceptual quality. The SSIM index is used to measure the perceptual quality of the GOMBs. The experimental results show that the encoding performance on intra-frames can be improved greatly by the proposed method compared with the fixed-QP method. Furthermore, we show that the optimization on the intra-frame can bring benefits to the whole sequence encoding, since a better reference frame can improve the encoding of the subsequent frames. The proposed method has acceptable encoding complexity for offline applications.
Chen proposed an image quality assessment method to evaluate image quality at a ratio of noise in an image. However, Chen's method had some drawbacks that unnoticeable noise is reflected in the evaluation or noise position is not accurately detected. Therefore, in this paper, we propose a new image quality measurement scheme using the mean-centered WLNI (Weber's Law Noise Identifier) and the saliency map. The experimental results show that the proposed method outperforms Chen's and agrees more consistently with human visual judgment.
Yazhong ZHANG Jinjian WU Guangming SHI Xuemei XIE Yi NIU Chunxiao FAN
Reduced-reference (RR) image quality assessment (IQA) algorithm aims to automatically evaluate the distorted image quality with partial reference data. The goal of RR IQA metric is to achieve higher quality prediction accuracy using less reference information. In this paper, we introduce a new RR IQA metric by quantifying the difference of discrete cosine transform (DCT) entropy features between the reference and distorted images. Neurophysiological evidences indicate that the human visual system presents different sensitivities to different frequency bands. Moreover, distortions on different bands result in individual quality degradations. Therefore, we suggest to calculate the information degradation on each band separately for quality assessment. The information degradations are firstly measured by the entropy difference of reorganized DCT coefficients. Then, the entropy differences on all bands are pooled to obtain the quality score. Experimental results on LIVE, CSIQ, TID2008, Toyama and IVC databases show that the proposed method performs highly consistent with human perception with limited reference data (8 values).
Dae-Chul KIM Wang-Jun KYUNG Ho-Gun HA Yeong-Ho HA
The role of an optical low-pass filter (OLPF) in a digital still camera is to remove the high spatial frequencies that cause aliasing, thereby enhancing the image quality. However, this also causes some loss of detail. Yet, when an image is captured without the OLPF, moiré generally appears in the high spatial frequency region of the image. Accordingly, this paper presents a moiré reduction method that allows omission of the OLPF. Since most digital still cameras use a CCD or a CMOS with a Bayer pattern, moiré patterns and color artifacts are simultaneously induced by aliasing at high spatial frequencies. Therefore, in this study, moiré reduction is performed in both the luminance channel to remove the moiré patterns and the color channel to reduce color smearing. To detect the moiré patterns, the spatial frequency response (SFR) of the camera is first analyzed. The moiré regions are identified using patterns related to the SFR of the camera and then analyzed in the frequency domain. The moiré patterns are reduced by removing their frequency components, represented by the inflection point between the high-frequency and DC components in the moiré region. To reduce the color smearing, color changing regions are detected using the color variation ratios for the RGB channels and then corrected by multiplying with the average surrounding colors. Experiments confirm that the proposed method is able to reduce the moiré in both the luminance and color channels, while also preserving the detail.
The tree-based routing approach has been known as an efficient method for node mobility management and data packet transmission between two long-distance parties; however, its parameter adjustment must balance control overhead against the convergence speed of topology information according to node mobility. Meanwhile, location-based routing works more efficiently when the distance between the source and destination is relatively short. Therefore, this paper proposes a roadside unit (RSU) based hybrid routing protocol, called RSU-HRP that combines the strengths of both protocols while offsetting their weaknesses. In RSU-HRP, the tree construction is modified to take into account the link and route quality to construct a robust and reliable tree against high node mobility, and an optimized broadcast algorithm is developed to reduce control overhead induced by the advertisement message periodically sent from a roadside unit. In addition, the two routing methods are selectively used based on the computed distance in hops between a source and a destination. Simulation results show that RSU-HRP far outperforms TrafRoute in terms of packet delivery ratio, end-to-end delay, and control overhead in both Vehicle-to-Infrastructure and Vehicle-to-Vehicle communication models.
Daniel LAGO Edmundo MADEIRA Deep MEDHI
With the growth of cloud-based services, cloud data centers are experiencing large growth. A key component in a cloud data center is the network technology deployed. In particular, Ethernet technology, commonly deployed in cloud data centers, is already envisioned for 10 Tbps Ethernet. In this paper, we study and analyze the makespan, workload execution times, and virtual machine migrations as the network speed increases. In particular, we consider homogeneous and heterogeneous data centers, virtual machine scheduling algorithms, and workload scheduling algorithms. Results obtained from our study indicate that the increase in a network's speed reduces makespan and workloads execution times, while aiding in the increase of the number of virtual machine migrations. We further observed that the number of migrations' behaviors in relation to the speed of the networks also depends on the employed virtual machines scheduling algorithm.
Yusuke AMANO Gosuke OHASHI Shogo MORI Kazuya SAWADA Takeshi HOSHINO Yoshifumi SHIMODAIRA
The present study proposes a method for estimation of subjective image quality, for combinations of display physical factors, based on the Mahalanobis-Taguchi system in the field of quality engineering. The proposed method estimates subjective image quality by the estimated equation based on the Mahalanobis-Taguchi System and subjective evaluation experiments using the method of successive categories for images of which parameters are combinations of gamma, maximum luminance and minimum luminance. The estimated image quality is in good agreement with the experimental subjective image quality.
Perceptually optimized missing texture reconstruction via neighboring embedding (NE) is presented in this paper. The proposed method adopts the structural similarity (SSIM) index as a measure for representing texture reconstruction performance of missing areas. This provides a solution to the problem of previously reported methods not being able to perform perceptually optimized reconstruction. Furthermore, in the proposed method, a new scheme for selection of the known nearest neighbor patches for reconstruction of target patches including missing areas is introduced. Specifically, by monitoring the SSIM index observed by the proposed NE-based reconstruction algorithm, selection of known patches optimal for the reconstruction becomes feasible even if target patches include missing pixels. The above novel approaches enable successful reconstruction of missing areas. Experimental results show improvement of the proposed method over previously reported methods.
Selective visual attention is an integral mechanism of the human visual system that is often neglected when designing perceptually relevant image and video quality metrics. Disregarding attention mechanisms assumes that all distortions in the visual content impact equally on the overall quality perception, which is typically not the case. Over the past years we have performed several experiments to study the effect of visual attention on quality perception. In addition to gaining a deeper scientific understanding of this matter, we were also able to use this knowledge to further improve various quality prediction models. In this article, I review our work with the aim to increase awareness on the importance of visual attention mechanisms for the effective design of quality prediction models.
Shosuke SATO Masaharu NAKAGAWA Masahiro IWASAKI Fumihiko IMAMURA
In the case of a disaster such as an earthquake or a tsunami, the city, town, and village administration usually issues an evacuation advisory and other information through the Outdoor Public Address Speakers for the disaster reduction broadcasting system covering its area of jurisdiction. However, in areas those have previous experience of a disaster, people frequently voice the lack of audibility of the disaster reduction broadcast. In this research, we conducted a questionnaire survey on the residents in the central area of Ishinomaki City, Miyagi Prefecture, who are the victims of the Great East Japan Earthquake Disaster, on the audible quality of outdoor public address (PA) speakers of the disaster reduction broadcasting system so as to understand the current state of such broadcasts and to propose ideal methods of sending and receiving information at the time of a future disaster.
Ya-Shih HUANG Han-Yuan CHANG Juinn-Dar HUANG
The emerging three-dimensional (3D) technology is considered as a promising solution for achieving better performance and easier heterogeneous integration. However, the thermal issue becomes exacerbated primarily due to larger power density and longer heat dissipation paths. The thermal issue would also be critical once FPGAs step into the 3D arena. In this article, we first construct a fine-grained thermal resistive model for 3D FPGAs. We show that merely reducing the total power consumption and/or minimizing the power density in vertical direction is not enough for a thermal-aware 3D FPGA backend (placement and routing) flow. Then, we propose our thermal-aware backend flow named TherWare considering location-based heat balance. In the placement stage, TherWare not only considers power distribution of logic tiles in both lateral and vertical directions but also minimizes the interconnect power. In the routing stage, TherWare concentrates on overall power minimization and evenness of power distribution at the same time. Experimental results show that TherWare can significantly reduce the maximum temperature, the maximum temperature gradient, and the temperature deviation only at the cost of a minor increase in delay and runtime as compared with present arts.
Taek LEE Jung-Been LEE Hoh Peter IN
Adherence to coding conventions during the code production stage of software development is essential. Benefits include enabling programmers to quickly understand the context of shared code, communicate with one another in a consistent manner, and easily maintain the source code at low costs. In reality, however, programmers tend to doubt or ignore the degree to which the quality of their code is affected by adherence to these guidelines. This paper addresses research questions such as “Do violations of coding conventions affect the readability of the produced code?”, “What kinds of coding violations reduce code readability?”, and “How much do variable factors such as developer experience, project size, team size, and project maturity influence coding violations?” To respond to these research questions, we explored 210 open-source Java projects with 117 coding conventions from the Sun standard checklist. We believe our findings and the analysis approach used in the paper will encourage programmers and QA managers to develop their own customized and effective coding style guidelines.
Wyllian B. da SILVA Keiko V. O. FONSECA Alexandre de A. P. POHL
Digital video signals are subject to several distortions due to compression processes, transmission over noisy channels or video processing. Therefore, the video quality evaluation has become a necessity for broadcasters and content providers interested in offering a high video quality to the customers. Thus, an objective no-reference video quality assessment metric is proposed based on the sigmoid model using spatial-temporal features weighted by parameters obtained through the solution of a nonlinear least squares problem using the Levenberg-Marquardt algorithm. Experimental results show that when it is applied to MPEG-2 streams our method presents better linearity than full-reference metrics, and its performance is close to that achieved with full-reference metrics for H.264 streams.
Ryosuke KOBAYASHI Takumi KATO Kazuhiro AZUMA Yasushi YAMAO
Current mobile communication terminals are equipped with multiple RF circuits that cover all frequency bands assigned for the communication. In order to make efficient use of frequency spectrum and to reduce circuits in a terminal, a low-loss reconfigurable RF filter is necessary to flexibly change RF frequencies. In this paper, a new reconfigurable bandpass filter (BPF) having eight-frequency (three-bit) selection capability is proposed. It employs branch-line switched type variable resonators that provide low insertion loss. One of the design issues is how to control pass bandwidths among selectable frequencies. In order to analyze the bandwidth variation of the reconfigurable BPF, we calculate the changes of external Q and coupling coefficients. It is shown that the inductive coupling design can achieve less variation of bandwidth for the reconfigurable BPF, compared with commonly used capacitive coupling design. A prototype BPF on a printed circuit board with high dielectric constant substrate has been fabricated and evaluated in 2 GHz bands. It presents performance very close to the design results with respect to insertion loss, center frequency and passband bandwidth. Low insertion loss of less than 1 dB is achieved among the eight frequencies.
Hang LI Yafei ZHANG Jiabao WANG Yulong XU Yang LI Zhisong PAN
State-of-the-art background subtraction and foreground detection methods still face a variety of challenges, including illumination changes, camouflage, dynamic backgrounds, shadows, intermittent object motion. Detection of foreground elements via the robust principal component analysis (RPCA) method and its extensions based on low-rank and sparse structures have been conducted to achieve good performance in many scenes of the datasets, such as Changedetection.net (CDnet); however, the conventional RPCA method does not handle shadows well. To address this issue, we propose an approach that considers observed video data as the sum of three parts, namely a row-rank background, sparse moving objects and moving shadows. Next, we cast inequality constraints on the basic RPCA model and use an alternating direction method of multipliers framework combined with Rockafeller multipliers to derive a closed-form solution of the shadow matrix sub-problem. Our experiments have demonstrated that our method works effectively on challenging datasets that contain shadows.
Ryo NISHIOKA Takayuki NISHIO Masahiro MORIKURA Koji YAMAMOTO
The IEEE 802.11 wireless LAN (WLAN) is based on carrier sense multiple access with collision avoidance (CSMA/CA) protocol. CSMA/CA uses a backoff mechanism to avoid collisions among stations (STAs). One disadvantage of backoff mechanisms is that STAs must wait for some period of time before transmission, which degrades spectral efficiency. Moreover, a backoff algorithm cannot completely avoid collisions. We have proposed a novel medium access control (MAC) scheme called the visual recognition-based medium access control (VRMAC) scheme, which uses an LED-camera communication technique. STAs send media-access request messages by blinking their LEDs in VRMAC scheme. An access point (AP) receives the messages via its camera, and then allocates transmission opportunities to the STAs by transmitting control frames. Since the transmission rate of the LED-camera communication is lower than WLAN transmission, the delay of access requesting causes and it could decrease the system throughput of the VRMAC system based WLAN. We reveal the effect of the delay for TCP flows and propose enhanced access procedures to eliminate the effect of the delay. Our simulation results demonstrate that VRMAC scheme increases the system throughput in UDP and TCP traffic. Moreover, the scenario-based evaluations reveal that VRMAC scheme also decreases the session delay which is a metric of quality of experience (QoE) for TCP applications.