|
1.IntroductionScale-adaptive tracking using only color information is a difficult and important problem. One of the most successful approaches using only color information is the kernel-based approach, which has become popular due to its tracking speed.1, 2 However, due to the instability of color-based tracking, research to stabilize the tracking has been done by introducing the scale space,3 changing the kernel function,4, 5 using multiple kernels or patches,6, 7, 8, 9 incorporating temporal filtering,10 or discriminating local reliable regions.11 Most color-based scale-adaptive tracking methods have utilized the histogram to determine the target size.1, 2, 3, 4, 12 However, in certain cases, the histogram fails to provide a good estimate of the target size (e.g., in cases of occlusion or the appearance of colors in the background similar to the target colors). This is due to the fact that the histogram is related to the numbers of pixels corresponding to the target colors. Therefore, sometimes colors in the background similar to the target colors cause the target region to spread out to the background region. In other cases, the target window sometimes shrinks too much, because small partial regions of the target show a similar distribution of colors as the entire target region. In this letter, we propose a stable scale-adaptive tracking algorithm that combines the color centroids–based tracking algorithm in Ref. 13 with a new scale-adaptation algorithm also based on the use of centroids. The advantages of using centroids in the computation of the scale are as follows: first, centroids have a direct relationship with the scale of the target, which makes the estimation of the scale simple and fast. Second, centroids can be filtered by filtering algorithms, leaving only the reliable centroids to be used in the rescaling. The proposed tracking algorithm produces stable scale-adaptive tracking results even in difficult cases when the background is cluttered or contains target colors, or when occlusion occurs. 2.Scale-Adaptive Centroid-Based Tracking AlgorithmIn this section, we first introduce the color centroids–based target localization algorithm. Then, we propose the scale-adaptation algorithm, which is also based on the use of the color centroids. The combination of the two algorithms will be explained at the end of this section. 2.1.Target Window Shifting Using Centroids of the Target ColorsIn Ref. 13, we proposed a stable target-shifting algorithm that shifts the current target location to the next target location based on the use of centroids of the target colors. The target position in the current frame ( [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _0$\end{document} ) is obtained by the area-weighted mean of the color centroids in the current frame, i.e., Eq. 1[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{equation} \mathop {{\bf {\hat{y}}}}\nolimits _0 = {\frac{\sum \nolimits _{u = 1}^m \,\mathop {\hat{q}}\nolimits _u {\bf {C}}_u^n}{\sum \nolimits _{u = 1}^m \,\mathop {\hat{q}}\nolimits _u }}, \end{equation} \end{document}The location of the target in the next frame ( [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _1$\end{document} ) can be calculated in the same way as [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _0$\end{document} using [TeX:] \documentclass[12pt]{minimal}\begin{document}${\bf {C}}_u^{n+1}$\end{document} , the color centroids in the next frame, instead of [TeX:] \documentclass[12pt]{minimal}\begin{document}${\bf {C}}_u^n$\end{document} in Eq. 1. Then, the shifting vector [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _{\rm shift}$\end{document} which shifts the location of the target in the current frame to the location in the next frame, is computed as [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _{\rm shift} = \mathop {{\bf {\hat{y}}}}\nolimits _1 - \mathop {{\bf {\hat{y}}}}\nolimits _0$\end{document} . 2.2.Scale Adaptation Based on Centroid Difference VectorsAlthough most scale-adaptation algorithms utilize histograms of the target colors, we utilize the centroids of the colors for the rescaling of the new target region. The reason for this is based on the following facts:
For scale adaptation, we first divide the target region (which is centered at [TeX:] \documentclass[12pt]{minimal}\begin{document}$\mathop {{\bf {\hat{y}}}}\nolimits _1$\end{document} ) into four equal subregions, as shown in Fig. 1. Figures 1, 1 show the target in the current frame and Figs. 1, 1 show the target in the next frame. The scale of the target changes in the next frame. After dividing the target region into subregions, we calculate the centroids in each subregion, independently. Figures 1, 1 show the calculated centroids in the current frame, and Figs. 1, 1 show those in the next frame. As shown in Figs. 1 and 1 or in Figs. 1, 1, the distances between the centroids in different subregions corresponding to the same colors have a direct relationship to the size of the target region. For example, in Figs. 1, 1, the ratio between [TeX:] \documentclass[12pt]{minimal}\begin{document}$d_{AD,u,x}^n$\end{document} and [TeX:] \documentclass[12pt]{minimal}\begin{document}$d_{AD,u,x}^{n + 1}$\end{document} can be used as an estimate of the ratio of the target's width in the next frame to that in the current frame. We first make some notations. We denote by [TeX:] \documentclass[12pt]{minimal}\begin{document}$d_{A,B,u,x}^{n + 1}$\end{document} the absolute difference in the x-coordinates between the centroids in subregions A and B corresponding to the color u: Eq. 2[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{equation} d_{{\rm A,B},u,x}^{n + 1} = \big|C_{{\rm A},u,x}^{n + 1} - C_{{\rm B},u,x}^{n + 1}\big|, \end{equation} \end{document}Eq. 3[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{eqnarray} d_{u,x}^{n + 1} &=& {W_{{\rm A},{\rm B},u}}d_{{\rm A},{\rm B},u,x}^{n + 1} + {W_{{\rm A},{\rm D},u}}d_{{\rm A},{\rm D},u,x}^{n + 1} + {W_{{\rm C},{\rm B},u}}d_{{\rm C},{\rm B},u,x}^{n + 1}\nonumber\\ && +\, {W_{{\rm C},{\rm D},u}}d_{{\rm C},{\rm D},u,x}^{n + 1} \end{eqnarray} \end{document}Eq. 4[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{eqnarray} {W_{{\rm A},{\rm B},u}} = {{{N_{{\rm A},u}}{N_{{\rm B},u}}} \over {{N_{{\rm A},u}}{N_{{\rm B},u}} + {N_{{\rm A},u}}{N_{{\rm D},u}} + {N_{{\rm C},u}}{N_{{\rm B},u}} + {N_{{\rm C},u}}{N_{{\rm D},u}}}},\nonumber\!\!\!\!\! \\ \end{eqnarray} \end{document}The zooming factor corresponding to the color u is then defined as Eq. 5[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{equation} z_{u,x}^{n + 1} = {{d_{u,x}^{n + 1}} \over {d_{u,x}^n}}. \end{equation} \end{document}However, it is not guaranteed that the centroid that corresponds to a certain color u is reliable. Therefore, we perform filtering on the estimated zooming factors to obtain a pair of reliable zooming factors. The filtering first trims off all the extreme values because the size of the target cannot change dramatically. After the extreme zooming factor values have been trimmed off, the remaining zooming factor values are averaged. The filtering suitable for this kind of problem is the α-trimmed mean filtering,14 which is performed by first sorting the remaining zooming factor values to get Eq. 6[TeX:] \documentclass[12pt]{minimal}\begin{document} \begin{equation} \big\lbrace z_x^n(1),\ldots ,z_x^n(k)\big\rbrace , \,\, \big\lbrace z_y^n(1),\ldots ,z_y^n(k)\big\rbrace , \end{equation} \end{document}2.3.Scale-Adaptive Tracking AlgorithmThe principle steps of the tracking algorithm that combines the centroid-based target shifting with the centroid-based scale adaptation can be described as follows:
3.Experimental ResultsWe performed several experiments and compared the results to the Camshift,1 the meanshift blob,3 the mean-shift particle filtering,10 and the Kwon–Lee algorithm,9 which we chose as a representative of patch-based tracking. For all the algorithms, the background colors were not removed when obtaining the initial target colors. Figure 2 shows the experimental results when the background is cluttered. This is a very difficult case for color-based tracking, because colors in the background similar to the target's colors are included in the target window. In this case, all the algorithms except the proposed one fail in the tracking. Figure 3 shows the case where the object experiences fast enlarging with a complex background. The mean-shift blob and mean-shift particle-filtering tracking fail in this case and are not shown here. The Kwon–Lee algorithm can track the object with a small initial target window, including not too much of the background colors. In comparison, the proposed scheme succeeds in the tracking even with a rather large target window. Table 1 shows the quantitative error results for the rain sequence. Table 1Summarization of the localization and scaling errors.
AcknowledgmentsThis work was supported by the National Research Foundation of Korea grant funded by the Korea government (MEST) (Grant No. 2011-0000096) and The “Dongseo Frontier Project” Research Fund of 2009 by Dongseo University. ReferencesG. R. Bradski,
“Computer vision face tracking for use in a perceptual user interface,”
Intel Technol. J., 2 1
–15
(1998). Google Scholar
D. Comaniciu, V. Ramesh and P. Meer,
“Kernel-based object tracking,”
IEEE Trans. Pattern Anal. Mach. Intell., 25 564
–575
(2003). https://doi.org/10.1109/TPAMI.2003.1195991 Google Scholar
R. Collins,
“Mean-shift blob tracking through scale space,”
(2003). Google Scholar
A. Yilmaz,
“Object tracking by asymmetric kernel mean shift with automatic scale and orientation selection,”
(2007). Google Scholar
C. Yang, R. Duraiswami, and L. Davis,
“Efficient mean-shift tracking via a new similarity measure,”
(2005). Google Scholar
G. Hager, M. Dewan, and C. Stewart,
“Multiple kernel tracking with SSD,”
(2004). Google Scholar
J. Jeyakar, R. Venkatesh Babu, K. R. Ramakrishnan,
“Robust object tracking using local kernels and background information,”
(2007). Google Scholar
F. Porikli and O. Tuzel,
“Multi-kernel object tracking,”
1234
–1237
(2005). Google Scholar
J. Kwon and K. Lee,
“Tracking of a non-rigid object via patch-based dynamic appearance modeling and adaptive basin hopping monte carlo sampling,”
(2009). Google Scholar
A. Naeem, S. Mills, and T. Pridmore,
“Structured combination of particle filter and kernel mean-shift tracking,”
(2006). Google Scholar
J. Fan, Y. Wu, and S. Dai,
“Discriminative spatial attention for robust tracking,”
(2010). Google Scholar
H. Wang, D. Suter, K. Schindler, and C. Shen,
“Adaptive object tracking based on an effective appearance filter,”
IEEE Trans. Pattern Anal. Mach. Intell., 29
(9), 1661
–1667
(2007). https://doi.org/10.1109/TPAMI.2007.1112 Google Scholar
S. H. Lee and M. G. Kang,
“Motion tracking based on area and level set weighted centroid shifting,”
IET Comput. Vis., 4
(2), 73
–84
(June 2010). https://doi.org/10.1049/iet-cvi.2008.0017 Google Scholar
R. Oten and R. Figueiredo,
“Adaptive alpha-trimmed mean filters under deviations from assumed noise model,”
IEEE Trans. Image Process., 13 627
–639
(2004). https://doi.org/10.1109/TIP.2003.821115 Google Scholar
|