1-3hit |
Template tracking has been extensively studied in Computer Vision with a wide range of applications. A general framework is to construct a parametric model to predict movement and to track the target. The difference in intensity between the pixels belonging to the current region and the pixels of the selected target allows a straightforward prediction of the region position in the current image. Traditional methods track the object based on the assumption that the relationship between the intensity difference and the region position is linear or non-linear. They will result in bad tracking performance when just one model is adopted. This paper proposes a method, called as Mixture Hyperplanes Approximation, which is based on finite mixture of generalized linear regression models to perform robust tracking. Moreover, a fast learning strategy is discussed, which improves the robustness against noise. Experiments demonstrate the performance and stability of Mixture Hyperplanes Approximation.
Hakaru TAMUKOH Keiichi HORIO Takeshi YAMAKAWA
This paper describes a new fast learning algorithm for Self-Organizing Map employing a "rough comparison winner-take-all" and its digital hardware architecture. In rough comparison winner-take-all algorithm, the winner unit is roughly and strictly assigned in early and later learning stage, respectively. It realizes both of high accuracy and fast learning. The digital hardware of the self-organizing map with proposed WTA algorithm is implemented using FPGA. Experimental results show that the designed hardware is superior to other hardware with respect to calculation speed.
Hiroki YOSHIMURA Tadaaki SHIMIZU Naoki ISU Kazuhiro SUGATA
A noise reduction filter composed of a sandglass-type neural network (Sandglass-type Neural network Noise Reduction Filter: SNNRF) was proposed in the present paper. Sandglass-type neural network (SNN) has symmetrical layer construction, and consists of the same number of units in input and output layers and less number of units in a hidden layer. It is known that SNN has the property of processing signals which is equivalent to KL expansion after learning. We applied the recursive least square (RLS) method to learning of SNNRF, so that the SNNRF became able to process on-line noise reduction. This paper showed theoretically that SNNRF behaves most optimally when the number of units in the hidden layer is equal to the rank of covariance matrix of signal component included in input signal. Computer experiments confirmed that SNNRF acquired appropriate characteristics for noise reduction from input signals, and remarkably improved the SN ratio of the signals.