Article Search
닫기

Original Article

Split Viewer

International Journal of Fuzzy Logic and Intelligent Systems 2018; 18(3): 182-189

Published online September 25, 2018

https://doi.org/10.5391/IJFIS.2018.18.3.182

© The Korean Institute of Intelligent Systems

Detection and Tracking of Multiple Moving Vehicles with a UAV

Min-Hyuck Lee, and Seokwon Yeom

School of Computer and Communication Engineering, Daegu University, Gyeongsan, Korea

Correspondence to :
Seokwon Yeom (yeom@daegu.ac.kr)

Received: June 1, 2018; Revised: September 4, 2018; Accepted: September 4, 2018

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Multiple object detection and tracking are essential for video surveillance. A drone or unmanned aerial vehicle (UAV) is very useful for aerial surveillance due to the efficiency of capturing remote scenes. This paper addresses the detection and tracking of moving vehicles with a UAV. A UAV can capture video sequences above the road at a distance. The detection step consists of frame differencing followed by thresholding, morphological filtering, and removing false alarms considering the true size of vehicles. The centroids of the detected areas are considered as measurements for tracking. Tracking is performed with Kalman filtering to estimate the state of the target based on the dynamic state model and measured positions. In the experiment, three moving cars are captured at a long distance by a drone. Experimental results show that the proposed method well detects moving cars and achieves good accuracy in tracking their dynamic state.

Keywords: UAV/drone imaging, Object detection, Multiple target tracking, Kalman filtering

Multiple object detection and tracking are important for security and surveillance [1]. Recently, there is increased usage of unmanned aerial vehicles (UAVs) or drones for aerial surveillance due to the efficiency of capturing remote scenes [2]. The image resolution for objects located far from the camera is low. Further, blurring and noise can degrade image quality. Various studies have been conducted on visual detection and tracking of moving objects. Methods based on background subtraction and frame difference were studied in [38]. Gaussian mixture modeling (GMM) was used to analyze background and target regions in [35]. Scale invariant feature transform (SIFT) was adopted to extract foreground objects from the background scene in [6, 7]. In [8], background was subtracted under Gaussian mixture assumption followed by morphological filters. In [9], hypothesis fitting was adopted for vehicle detection. Long-range moving objects were detected by background subtraction in [10].

In addition to the above-mentioned research, many studies have been conducted on multiple-target tracking [1114]. High maneuvering, closely located multiple targets, heavy clutter (false alarm), and low detection probability are often considered difficulties to overcome [11]. Kalman filtering provides a solution for estimating the state of the target in real time. It is known to be optimal under independent Gaussian noise assumption [12, 13]. When multiple measurements are detected in a frame, data association is required to assign the measurements to the established tracks [14].

Vison-based target tracking with UAVs has been researched in [1522]. Correlation filtering-based tracking was studied in [15]. Closely located-object tracking was performed with feature matching and multi-inertial sensing data in [16]. Pedestrians were tracked by template-matching in [17, 18]. Small animals were tracked with a freely moving camera in [19]. A moving ground target was tracked in dense obstacles areas with UAVs in [20]. Multiple moving targets are tracked by Kalman filtering and IMM filtering in [21] and [22], respectively. In [23], position and velocity were estimated with multiple sensor fusion.

In this paper, we address the detection and tracking of multiple moving vehicles obtained by a UAV, which captures videos above the road from a long distance. To achieve this goal, we propose to integrate two systems: one is detection based on image processing, which manipulates pixels in the stationary scene, and the other is tracking based on Kalman filtering, which estimates dynamic states (position and velocity) of targets in the time domain. Object detection is performed through frame differencing and thresholding, morphological filtering, and removing false alarms based on the minimum size of vehicles. Two frames separated by a constant interval are subtracted and turned into a binary image by thresholding, which can represent the difference between the current and the past frames. Two morphological operations, erosion and dilation, are applied to the binary image to extract regions of interest (ROIs), which are the disjoint alternative areas. The ROIs are the candidate areas of moving vehicles. False ROIs are removed by the minimum ROI size, which is based on the true size of moving vehicles. The centroids of ROIs are the results of detection. The centroids represent the measured positions of targets, which will be processed in the next tracking stage. Tracking is performed by Kalman filtering to estimate the state of the target. A nearly constant velocity (NCV) model is assumed for the dynamic state of the target [11]. A position gating process excludes the measurement outside the validation region. The nearest measurement-to-track association scheme assigns one measurement to one track [21]. The tracks are initialized by a two-point differencing initialization method with a maximum speed gating [11]. In the experiments, three moving cars are captured at a long distance by a UAV. The camera built in a UAV points directly downward to the road. Root-mean-squared errors (RMSEs) are obtained for position and velocity in x and y directions. Experimental results show that the location and the speed are well detected and tracked with good accuracy by the proposed method.

The remainder of the paper is organized as follows. Object detection is discussed in Section 2. Multiple target tracking is presented in Section 3. Section 4 demonstrates experimental results. The conclusion follows in Section 5.

A block diagram of object detection is described in Figure 1. First, frame difference is obtained between the current and the past frames. A thresholding step follows to generate a binary image as:

IT(m,n;k)={1,ifI(m,n;k)-I(m,n;k-kd)>θT,0,otherwise,}m=1,,M,   n=1,,N,

where I(m, n; k) and I(m, n; kkd) are the kth frame and the (kkd)th frame, respectively, kd is a constant interval for frame differencing, θT is a thresholding value, andM and N are pixel sizes in x and y directions, respectively. Two morphological operations, erosion and dilation, are applied sequentially to the binary image to segment moving vehicles. Erosion and dilation filtering, respectively, are defined as [24]:

IE={lElITc=ϕ},ID={lDlIEϕ},

where El = {e + l|eE} and Dl = {d + l|dD}; E and D are the structuring elements for erosion and dilation, respectively. It is noted that l is an integer value less than the image size. Morphological filters extract ROIs, which are the disjoint alternative areas. At last, false ROIs are removed by the minimum ROI size as:

Oi(m,n)={1,Size{Oi}θS,0,otherwise},

where Oi is the ith ROI and θs is the minimum size of ROI. The center of each ROI is considered as a measurement for target tracking in the next section.

3.1 System Modeling

The dynamic state of the target is modeled as a NCV model; the target’s maneuvering is modeled by the uncertainty of the process noise, which is assumed to follow the Gaussian distribution. The following is the discrete state equation of a target:

xt(k+1)=F(Δ)xt(k)+q(Δ)v(k),   t=1,,NT,F(Δ)=[1Δ000100001Δ0001],   q(Δ)=[Δ2/20Δ00Δ2/20Δ],

where xt(k) = [xt(k) vtx(k) yt(k) vty(k)]T is the state vector at frame k, xt(k) and yt(k) are the positions in x and y directions, respectively, vtx(k) and vty(k) are the velocities in x and y directions, respectively, T denotes matrix transposition, Δ is the sampling time, and v(k) is a process noise vector, which is Gaussian white noise with the covariance matrix Q=diag([σx2σy2]), and NT is the number of targets. A measurement vector is composed of two components in x and y directions. The following is the measurement equation:

zt(k)=[ztx(k)zty(k)]=Hxt(k)+w(k),H=[10000010],

where w(k) is a measurement noise vector, which is Gaussian white noise with the covariance matrix R=diag([rx2ry2]).

3.2 Two-Point Differencing Initialization with Velocity Gating

The initial state of each target is calculated by the two-point differencing method with a maximum speed gating [14]. The initial state and covariance matrix of target t become:

x^t(kk)=[x^t(kk)v^tx(kk)y^t(kk)v^ty(kk)]=[ztx(k)ztx(k)-ztx(k-1)Δzty(k)zty(k)-zty(k-1)Δ],Pt(kk)=[rx2rx2Δ00rx2Δ2rx2Δ20000ry2ry2Δ00ry2Δ2ry2Δ2].

The state is confirmed as an initial state of a track if the following speed gating is satisfied:

[v^tx(kk)]2+[v^ty(kk)]2Vmax.

3.3 Prediction and Filter Gain

The state and covariance predictions are iteratively computed with the initial values as:

x^t(kk-1)=Fx^t(k-1k-1),Pt(kk-1)=FPt(k-1k-1)FT+Q,

where t(k|k−1) and Pt(k|k−1), respectively, are the state and the covariance prediction for target t and frame k. The residual covariance St(k) and the filter gain Wt(k), respectively, are obtained as:

St(k)=HPt(kk-1)HT+R,Wt(k)=Pt(kk-1)HTSt(k)-1.

3.4 Data Association

Data association is the process of assigning multiple measurements to established tracks. Measurement gating is a preprocess that reduces the number of candidate measurements for data association. The measurement gating is performed by chi-square hypothesis testing with the assumption of the Gaussian measurement residuals [11]. Measurements become valid for target t at frame k if they exist in the validation region as

Zt(k)={zm(k)νtm(k)T[St(k)]-1νtm(k)γ,m=1,,mt(k)},νtm(k)=zm(k)-Hxt(kk-1),

where zm(k) is the mth measurement vector at frame k, γ is the gating size, and mt(k) is the number of candidate measurements for target t at frame k. The nearest neighbor rule associates a measurement with track t as:

m^t=argminm=1,,mt(k)νtm(k)T[St(k)]-1νtm(k).

3.5 State Estimate and Covariance Update

The state estimate and the covariance matrix of targets are updated as follows:

x^t(kk)=x^t(kk-1)+Wt(k)νtm^t(k),Pt(kk)=Pt(kk-1)-Wt(k)St(k)Wt(k)T.

If mt(k) is equal to zero, i.e., no measurement is associated with target t at frame k, they merely become the predictions of the state and the covariance as:

x^t(kk)=x^t(kk-1),Pt(kk)=Pt(kk-1).

3.6 Performance Evaluation

Several metrics are used for performance evaluation, position error, velocity error, and RMSE of position and velocity. The position error of target t at frame k is obtained as:

epxt(k)=x^t(kk)-xttrue(k),epyt(k)=y^t(kk)-yttrue(k),

where xttrue(k) and yttrue(k) are the ground truth of position of target t at frame k, in x and y directions, respectively. The ground truth of positions of targets is obtained manually at each scene. The RMSE of position is obtained as:

RMSEp(t)=1Kt(f)-Kt(s)+1k=Kt(s)Kt(f){[epxt(k)]2+[epyt(k)]2},

where Kt(s) and Kt(f) are the first and last frame where target t is estimated. The velocity error of target t at frame k is obtained in both x and y directions as:

evxt(k)=v^tx(kk)-vtxtrue(k),evyt(k)=v^ty(kk)-ytytrue(k),

where the ground truth of the velocity is approximated as:

vtxtrue(k)=12δΔ[xttrue(k+δ)-xttrue(k-δ)],vtytrue(k)=12δΔ[yttrue(k+δ)-yttrue(k-δ)].

The RMSE of velocity is obtained as:

RMSEv(t)=1Kt(f)-Kt(s)+1k=Kt(s)Kt(f){[evxt(k)]2+[evyt(k)]2}.

4.1 Scenario Description

A drone (Phantom 4 Advanced) captures three moving vehicles at 30 frames per second. A total of 372 frames are captured for more than 12 seconds. The size of one frame is 4096×2160 pixels; the frame is reduced to 20% size, 820×432 pixels, for efficient image processing. The drone hovers (stays still in the air) while capturing video sequences at a height of 92 m above the road. The camera points directly downwards while capturing video sequences. One pixel corresponds to 0.1344 m. There are three cars moving, with different initialization positions and maneuverings. The first car (Car 1) is located at the center of the scene from the first frame and moves to the right until it disappear after the 155th frame. The second car (Car 2) moves to the left from the first frame to the 289th frame. The third car (Car 3) moves in the same direction as the second car from the 74th to the 372nd frame. Figure 2(a) shows Cars 1 and 2 at the first frame, Figure 2(b) shows Cars 1–3 at the 150th frame, and Figure 2(c) shows Car 3 at the 300th frame. In the figures, the red circles indicate the moving vehicles.

4.2 Detection of Moving Objects

A total of 368 frames from the 5th frame to the 372nd frame is processed to detect moving objects, since kd in Eq. (1) is set at 4, θT in Eq. (1) is set at 30, and E and D in Eqs. (2) and (3) are set at [1]9×9 and [1]15×15, respectively. The parameters for thresholding and morphological operations are determined heuristically when better results are produced; θs in Eq. (4) is set at 90 considering the true size of moving vehicles. The first row in Figure 3 shows the detection process of Figure 2(a). The second and third rows are of Figure 2(b) and (c), respectively. Figure 3(a) and (b), respectively, are the frame differencing and thresholding to generate a binary image. Figure 3(c) shows the morphological filtering (erosion and dilation) result; the centroids of the segmented areas are marked with an asterisk (*). The ROI window is a rectangle that includes disjoint detected areas. Figure 3(d) shows the segmented areas by the ROI windows after removing false ROIs according to Eq. (4).

Figure 4 shows all centroids for 368 frames. Table 1 shows the detection rate of three cars. Average detection rate is around 96%. The low detection rate of Car 3 is mainly due to the tree occlusion and the same background color (white sidewalk) as the target. A total of 10 false alarms are detected, and two of them are caused by the split measurements from one car.

4.3 Multiple Target Tracking

The centroids in Figure 4 become input measurements for tracking. The sampling time Δ in Eq. (6) is 0.033 s since the frame rate is 30 fps. The standard deviation of process and the measurement noise are set at σx = σy = 30 m/s2 and rx = ry = 1.5 m, respectively; γ for the gating test is set at 4; and σx and σy are conventionally set between 0.5amax and amax where amax is the maximum acceleration of the target [14]; and rx and ry are set at half the length of the vehicle. The validation region in Eq. (14) is chi-square distributed with the degree of freedom is equal to the dimension of measurement vectors, thus the probability mass corresponds to 86.5% when γ is set at 4. The maximum speed for speed gating of two-point differencing initialization is set at 30 m/s. The initial-state vectors of Cars 1 and 2 are initialized by the measured position in the 5th and 6th frames; the initial state of Car 3 is initialized by the 74th and 75th frames. Cars 1, 2, and 3 are estimated until 155th, 289th, and 372nd frame for 150, 284, and 298 frames, respectively, including initialization. Figure 5(a)–(c) are the results of tracking Cars 1–3, respectively. No false tracks are established, and no track termination scheme is applied.

Figure 6(a) shows the ground truth of the position of Car 1 from the first to the 155th frame. Figures 6(b) and c are the position vectors of Cars 2 and 3, respectively, from the first frame to the 289th frame, and from the 74th frame to the 372nd frame. Figure 7(a)–(c) shows the approximated ground truth of the velocity in x and y directions of Car 1, 2, and 3, obtained by Eq. (24); δis set at 26 when the least average velocity RMSE is produced. Figures 8 and 9 show the position in Eq. (21) and velocity errors in Eq. (23), respectively. Large position errors in x direction are mainly caused by measurement errors. High speed generates greater measurement errors. Therefore, measurement errors during detection are bigger in the direction of the car’s movement. Velocity errors are generated during tracking. It is noted that the velocity error differences in x and y directions are small compared with those of position errors.

The sampling time is set differently in the following experiments. Figure 10 shows the tracking result with Δ = 0.1 s, that is, measurements are acquired every third frame from the original data. Cars 1 and 2 are initialized by the 5th and 8th frames and Car 3 is initialized by the 74th and 77th frames. Cars 1, 2, and 3 are estimated until 155th, 287th, and 371th frame, respectively, thus Cars 1, 2, and 3 are estimated for 50, 94, and 99 frames, respectively, including initialization. Figure 11 is the tracking results with every sixth frame (Δ = 0.2 s) updated. Cars 1 and 2 are initialized by the 5th and 11th frames and Car 3 is initialized by the 77th and 83rd frames. Cars 1, 2, and 3 are estimated for 25, 47, and 49 frames, respectively, until 155th, 298th, and 371st frame. The relative computational times are 1, 0.51, and 0.34 for Δ = 0.033, Δ = 0.1, and Δ = 0.2 s, respectively.

Table 2 shows the RMSE of position in Eq. (22). The large position RMSEs of Cars 2 and 3 are due to the high maneuvering of the circular motion. The average position RMSE is bigger over many sampling times, as shown in Table 2. Finer sampling steps can reduce position errors in Car 1 and 2, costing more computational resources. Table 3 shows the RMSE of velocity in Eq. (25). The velocity RMSE is also bigger for Cars 2 and 3 than Car 1. It is shown that the average velocity RMSE decreases as the sampling time increases.

In this paper, multiple moving objects are captured over long distances by a UAV. The locations of objects are detected using a background scene. Three targets are tracked with Kalman filtering. Experimental results show that the proposed algorithm detects and tracks moving objects with good accuracy. Position and velocity errors are obtained for different maneuvering targets and sampling times. It is shown that the proposed system can adjust the sampling time to build a more efficient system in computational resources. This property is favorable for drone applications because the computational resources in a drone are very limited. This study also proposes new applications for drones, such as traffic control and smart car as well as security and defense. Tracking a large number of targets with high maneuvering remains for future study.

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (Grant Number: 2017R1D1 A3B03031668).

Fig. 1.

Block diagram of moving object detection.


Fig. 2.

(a) Cars 1 and 2 at the first frame; (b) Cars 13 at the 150th frame; (c) Car 3 at the 300th frame. Moving cars are indicated by the red circles.


Fig. 3.

(a) Frame differencing; (b) thresholding; (c) morphological filtering and centroids; (d) segmented region by ROI windows after false alarm removal.


Fig. 4.

Measurements including false alarms for 368 frames.


Fig. 5.

Tracking results whenΔ = 0.033 s: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 6.

Ground truth of position: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 7.

Approximated ground truth of velocity: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 8.

Position error: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 9.

Velocity error: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 10.

Tracking results whenΔ = 0.1 s: (a) Car 1, (b) Car 2, (c) Car 3.


Fig. 11.

Tracking results whenΔ = 0.2 s: (a) Car 1, (b) Car 2, (c) Car 3.


Table. 1.

Table 1. Detection rates.

Number of framesNumber of detectionsDetection rate (%)
Car 1151151100
Car 228528298.95
Car 329927391.3
Total/average73570696.05

Table. 2.

Table 2. Position RMSE.

Sampling time (s)RMSE (m)Average
Car 1Car 2Car 3
0.0330.8851.6191.5671.345
0.10.891.6841.5611.378
0.20.9251.6861.6361.416

Table. 3.

Table 3. Velocity RMSE.

Sampling time (s)RMSE (m/s)Average
Car 1Car 2Car 3
0.0331.4721.712.1141.765
0.11.3771.9831.8441.734
0.21.2872.0152.3661.89

  1. Hampapur, A, Brown, L, Connell, J, Pankanti, S, Senior, AW, and Tian, YL . Smart surveillance: applications, technologies and implications., Proceedings of the 2003 Joint Conference of the 4th International Conference on Information, Communications and Signal Processing and 4th Pacific Rim Conference on Multimedia, 2003, Singapore, Array, pp.1133-1138. https://doi.org/10.1109/ICICS.2003.1292637
    CrossRef
  2. Kumar, R, Sawhney, H, Samarasekera, S, Hsu, S, Tao, H, and Guo, Y (2001). Aerial video surveillance and exploitation. Proceedings of the IEEE. 89, 1518-1539. https://doi.org/10.1109/5.959344
    CrossRef
  3. Chung, W, Kim, Y, Kim, YJ, and Kim, D . A two-stage foreground propagation for moving object detection in a non-stationary., Proceedings of 2016 13th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), 2016, Colorado Springs, CO, Array, pp.187-193. https://doi.org/10.1109/AVSS.2016.7738024
    CrossRef
  4. Yeh, CH, Lin, CY, Lai, HE, Muchtar, K, and Sun, MT (2017). Three-pronged compensation and hysteresis thresholding for moving object detection in real-time video surveillance. IEEE Transactions on Industrial Electronics. 64, 4945-4955. https://doi.org/10.1109/tie.2017.2669881
    CrossRef
  5. Wu, Y, He, X, and Nguyen, TQ (2017). Moving object detection with a freely moving camera via background motion subtraction. IEEE Transactions on Circuits and Systems for Video Technology. 27, 236-248. https://doi.org/10.1109/tcsvt.2015.2493499
    CrossRef
  6. Mousse, M, Motamed, C, and Ezin, E . An adaptive algorithm for fast moving object detection from dynamic background based on codebook., Proceedings of 2016 4th International Conference on Robotics and Mechatronics (ICROM), 2016, Tehran, Iran, Array, pp.584-588. https://doi.org/10.1109/ICRoM.2016.7886808
    CrossRef
  7. Zhang, Y, Huang, X, Li, J, Liu, X, Zhang, H, and Xing, X . Research of moving object detection algorithm in transmission lines under complex background., Proceedings of 2016 International Conference on Condition Monitoring and Diagnosis (CMD), 2016, Array, pp.176-179. https://doi.org/10.1109/cmd.2016.7757799
    CrossRef
  8. Olugboja, A, and Wang, Z . Detection of moving objects using foreground detector and improved morphological filter., Proceedings of the 3rd International Conference on Information Science and Control Engineering, 2016, Beijing, China, Array, pp.329-333. https://doi.org/10.1109/icisce.2016.80
    CrossRef
  9. Cheon, M, and Lee, H (2017). Vision-based vehicle detection system applying hypothesis fitting. International Journal of Fuzzy Logic and Intelligent Systems. 17, 58-67. http://dx.doi.org/10.5391/IJFIS.2017.17.2.58
    CrossRef
  10. Yeom, S, Lee, MH, and Cho, IJ . Long-range moving object detection based on background subtraction., Proceedings of the 18th International Symposium on Advanced Intelligent Systems (ISIS), 2017, Daegu, Korea, pp.1082-1085.
  11. Bar-Shalom, Y, and Li, XR (1995). Multitarget-Multisensor Tracking: Principles and Techniques. Storrs CT: YBS Publishing
  12. Stone, LD, Streit, RL, Corwin, TL, and Bell, KL (2014). Bayesian Multiple Target Tracking. Boston, MA: Artech House
  13. Hamid, KR, Talukder, A, and Ehtesanul Islam, AKM (2018). Implementation of fuzzy aided Kalman filter for tracking a moving object in two-dimensional space. International Journal of Fuzzy Logic and Intelligent Systems. 18, 85-96. https://doi.org/10.5391/IJFIS.2018.18.2.85
    CrossRef
  14. Yeom, S, Kirubarajan, T, and Bar-Shalom, Y (2004). Track segment association, fine-step IMM, and initialization with Doppler for improved track performance. IEEE Transactions on Aerospace and Electronic Systems. 40, 293-309. https://doi.org/10.1109/TAES.2004.1292161
    CrossRef
  15. Chaudhary, K, Zhao, M, Shi, F, Chen, X, Okada, K, and Inaba, M . Robust real-time visual tracking using dual-frame deep comparison network integrated with correlation filters., Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2017, Array, pp.6837-6842. https://doi.org/10.1109/IROS.2017.8206604
    CrossRef
  16. Chen, P, Dang, Y, Liang, R, Zhu, W, and He, X (2018). Realtime object tracking on a drone with multi-inertial sensing Data. IEEE Transactions on Intelligent Transportation Systems. 19, 131-139. https://doi.org/10.1109/TITS.2017.2750091
    CrossRef
  17. Bian, C, Yang, Z, Zhang, T, and Xiong, H . Pedestrian tracking from an unmanned aerial vehicle., Proceedings of 2016 IEEE 13th International Conference on Signal Processing (ICSP), 2016, Array, pp.1067-1071. https://doi.org/10.1109/ICSP.2016.7877993
    CrossRef
  18. Haag, K, Dotenco, S, and Gallwitz, F . Correlation filter based visual trackers for person pursuit using a low-cost quadrotor., Proceedings of the 15th International Conference on Innovations for Community Services (I4CS), 2015, Array, pp.1-8. https://doi.org/10.1109/I4CS.2015.7294481
    CrossRef
  19. Risse, B, Mangan, M, Del Pero, L, and Webb, B . Visual tracking of small animals in cluttered natural environments using a freely moving camera., Proceedings of IEEE International Conference on Computer Vision Workshops, 2017, Venice, Italy, Array, pp.2840-2849. https://doi.org/10.1109/ICCVW.2017.335
    CrossRef
  20. Kim, J, and Kim, Y . Moving ground target tracking in dense obstacle areas using UAVs., Proceedings of the 17th IFAC World Congress, 2008, Toulouse, France, pp.8552-8557.
  21. Lee, MH, and Yeom, S . Vehicle tracking based on Kalman filtering with drone imaging., Presented in the 6th International Conference on Green and Human Information Technology (ICGHIT), 2018, Chiang Mai, Thailand.
  22. Lee, MH, and Yeom, S (2018). Multiple target detection and tracking on urban roads with a drone. Journal of Intelligent & Fuzzy Systems. https://doi.org/10.3233/JIFS-169847
    CrossRef
  23. Cho, Y, and Lee, H (2017). A position and velocity estimation using multifarious and multiple sensor fusion. International Journal of Fuzzy Logic and Intelligent Systems. 17, 121-128. https://doi.org/10.5391/IJFIS.2017.17.2.121
    CrossRef
  24. Gonzalez, RC, and Woods, RE (2008). Digital Image Processing. Upper Saddle River, NJ: Pearson Prentice- Hall

Minhyuck Lee is currently pursuing a B.S. degree in School of Computer and Communication Engineering at Daegu University in South Korea. He has been working on researches related to image processing using UAV. His research interests include intelligent image processing and target tracking.

E-mail: dool0331@daegu.ac.kr

Seokwon Yeom received the M.S. and B.S. degrees in Electronics Engineering from Korea University and the Ph.D. degree in Electrical and Computer Engineering from the University of Connecticut. He is currently a professor of School of Computer and Communication Engineering at Daegu University in Korea. He is now performing several research projects related to image and signal processing. His research interests include image processing, target tracking, machine learning, and optical information processing.

E-mail: yeom@daegu.ac.kr

Article

Original Article

International Journal of Fuzzy Logic and Intelligent Systems 2018; 18(3): 182-189

Published online September 25, 2018 https://doi.org/10.5391/IJFIS.2018.18.3.182

Copyright © The Korean Institute of Intelligent Systems.

Detection and Tracking of Multiple Moving Vehicles with a UAV

Min-Hyuck Lee, and Seokwon Yeom

School of Computer and Communication Engineering, Daegu University, Gyeongsan, Korea

Correspondence to:Seokwon Yeom (yeom@daegu.ac.kr)

Received: June 1, 2018; Revised: September 4, 2018; Accepted: September 4, 2018

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0/) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Multiple object detection and tracking are essential for video surveillance. A drone or unmanned aerial vehicle (UAV) is very useful for aerial surveillance due to the efficiency of capturing remote scenes. This paper addresses the detection and tracking of moving vehicles with a UAV. A UAV can capture video sequences above the road at a distance. The detection step consists of frame differencing followed by thresholding, morphological filtering, and removing false alarms considering the true size of vehicles. The centroids of the detected areas are considered as measurements for tracking. Tracking is performed with Kalman filtering to estimate the state of the target based on the dynamic state model and measured positions. In the experiment, three moving cars are captured at a long distance by a drone. Experimental results show that the proposed method well detects moving cars and achieves good accuracy in tracking their dynamic state.

Keywords: UAV/drone imaging, Object detection, Multiple target tracking, Kalman filtering

1. Introduction

Multiple object detection and tracking are important for security and surveillance [1]. Recently, there is increased usage of unmanned aerial vehicles (UAVs) or drones for aerial surveillance due to the efficiency of capturing remote scenes [2]. The image resolution for objects located far from the camera is low. Further, blurring and noise can degrade image quality. Various studies have been conducted on visual detection and tracking of moving objects. Methods based on background subtraction and frame difference were studied in [38]. Gaussian mixture modeling (GMM) was used to analyze background and target regions in [35]. Scale invariant feature transform (SIFT) was adopted to extract foreground objects from the background scene in [6, 7]. In [8], background was subtracted under Gaussian mixture assumption followed by morphological filters. In [9], hypothesis fitting was adopted for vehicle detection. Long-range moving objects were detected by background subtraction in [10].

In addition to the above-mentioned research, many studies have been conducted on multiple-target tracking [1114]. High maneuvering, closely located multiple targets, heavy clutter (false alarm), and low detection probability are often considered difficulties to overcome [11]. Kalman filtering provides a solution for estimating the state of the target in real time. It is known to be optimal under independent Gaussian noise assumption [12, 13]. When multiple measurements are detected in a frame, data association is required to assign the measurements to the established tracks [14].

Vison-based target tracking with UAVs has been researched in [1522]. Correlation filtering-based tracking was studied in [15]. Closely located-object tracking was performed with feature matching and multi-inertial sensing data in [16]. Pedestrians were tracked by template-matching in [17, 18]. Small animals were tracked with a freely moving camera in [19]. A moving ground target was tracked in dense obstacles areas with UAVs in [20]. Multiple moving targets are tracked by Kalman filtering and IMM filtering in [21] and [22], respectively. In [23], position and velocity were estimated with multiple sensor fusion.

In this paper, we address the detection and tracking of multiple moving vehicles obtained by a UAV, which captures videos above the road from a long distance. To achieve this goal, we propose to integrate two systems: one is detection based on image processing, which manipulates pixels in the stationary scene, and the other is tracking based on Kalman filtering, which estimates dynamic states (position and velocity) of targets in the time domain. Object detection is performed through frame differencing and thresholding, morphological filtering, and removing false alarms based on the minimum size of vehicles. Two frames separated by a constant interval are subtracted and turned into a binary image by thresholding, which can represent the difference between the current and the past frames. Two morphological operations, erosion and dilation, are applied to the binary image to extract regions of interest (ROIs), which are the disjoint alternative areas. The ROIs are the candidate areas of moving vehicles. False ROIs are removed by the minimum ROI size, which is based on the true size of moving vehicles. The centroids of ROIs are the results of detection. The centroids represent the measured positions of targets, which will be processed in the next tracking stage. Tracking is performed by Kalman filtering to estimate the state of the target. A nearly constant velocity (NCV) model is assumed for the dynamic state of the target [11]. A position gating process excludes the measurement outside the validation region. The nearest measurement-to-track association scheme assigns one measurement to one track [21]. The tracks are initialized by a two-point differencing initialization method with a maximum speed gating [11]. In the experiments, three moving cars are captured at a long distance by a UAV. The camera built in a UAV points directly downward to the road. Root-mean-squared errors (RMSEs) are obtained for position and velocity in x and y directions. Experimental results show that the location and the speed are well detected and tracked with good accuracy by the proposed method.

The remainder of the paper is organized as follows. Object detection is discussed in Section 2. Multiple target tracking is presented in Section 3. Section 4 demonstrates experimental results. The conclusion follows in Section 5.

2. Object Detection with Frame Differencing

A block diagram of object detection is described in Figure 1. First, frame difference is obtained between the current and the past frames. A thresholding step follows to generate a binary image as:

IT(m,n;k)={1,ifI(m,n;k)-I(m,n;k-kd)>θT,0,otherwise,}m=1,,M,   n=1,,N,

where I(m, n; k) and I(m, n; kkd) are the kth frame and the (kkd)th frame, respectively, kd is a constant interval for frame differencing, θT is a thresholding value, andM and N are pixel sizes in x and y directions, respectively. Two morphological operations, erosion and dilation, are applied sequentially to the binary image to segment moving vehicles. Erosion and dilation filtering, respectively, are defined as [24]:

IE={lElITc=ϕ},ID={lDlIEϕ},

where El = {e + l|eE} and Dl = {d + l|dD}; E and D are the structuring elements for erosion and dilation, respectively. It is noted that l is an integer value less than the image size. Morphological filters extract ROIs, which are the disjoint alternative areas. At last, false ROIs are removed by the minimum ROI size as:

Oi(m,n)={1,Size{Oi}θS,0,otherwise},

where Oi is the ith ROI and θs is the minimum size of ROI. The center of each ROI is considered as a measurement for target tracking in the next section.

3. Target Tracking with Kalman Filtering

3.1 System Modeling

The dynamic state of the target is modeled as a NCV model; the target’s maneuvering is modeled by the uncertainty of the process noise, which is assumed to follow the Gaussian distribution. The following is the discrete state equation of a target:

xt(k+1)=F(Δ)xt(k)+q(Δ)v(k),   t=1,,NT,F(Δ)=[1Δ000100001Δ0001],   q(Δ)=[Δ2/20Δ00Δ2/20Δ],

where xt(k) = [xt(k) vtx(k) yt(k) vty(k)]T is the state vector at frame k, xt(k) and yt(k) are the positions in x and y directions, respectively, vtx(k) and vty(k) are the velocities in x and y directions, respectively, T denotes matrix transposition, Δ is the sampling time, and v(k) is a process noise vector, which is Gaussian white noise with the covariance matrix Q=diag([σx2σy2]), and NT is the number of targets. A measurement vector is composed of two components in x and y directions. The following is the measurement equation:

zt(k)=[ztx(k)zty(k)]=Hxt(k)+w(k),H=[10000010],

where w(k) is a measurement noise vector, which is Gaussian white noise with the covariance matrix R=diag([rx2ry2]).

3.2 Two-Point Differencing Initialization with Velocity Gating

The initial state of each target is calculated by the two-point differencing method with a maximum speed gating [14]. The initial state and covariance matrix of target t become:

x^t(kk)=[x^t(kk)v^tx(kk)y^t(kk)v^ty(kk)]=[ztx(k)ztx(k)-ztx(k-1)Δzty(k)zty(k)-zty(k-1)Δ],Pt(kk)=[rx2rx2Δ00rx2Δ2rx2Δ20000ry2ry2Δ00ry2Δ2ry2Δ2].

The state is confirmed as an initial state of a track if the following speed gating is satisfied:

[v^tx(kk)]2+[v^ty(kk)]2Vmax.

3.3 Prediction and Filter Gain

The state and covariance predictions are iteratively computed with the initial values as:

x^t(kk-1)=Fx^t(k-1k-1),Pt(kk-1)=FPt(k-1k-1)FT+Q,

where t(k|k−1) and Pt(k|k−1), respectively, are the state and the covariance prediction for target t and frame k. The residual covariance St(k) and the filter gain Wt(k), respectively, are obtained as:

St(k)=HPt(kk-1)HT+R,Wt(k)=Pt(kk-1)HTSt(k)-1.

3.4 Data Association

Data association is the process of assigning multiple measurements to established tracks. Measurement gating is a preprocess that reduces the number of candidate measurements for data association. The measurement gating is performed by chi-square hypothesis testing with the assumption of the Gaussian measurement residuals [11]. Measurements become valid for target t at frame k if they exist in the validation region as

Zt(k)={zm(k)νtm(k)T[St(k)]-1νtm(k)γ,m=1,,mt(k)},νtm(k)=zm(k)-Hxt(kk-1),

where zm(k) is the mth measurement vector at frame k, γ is the gating size, and mt(k) is the number of candidate measurements for target t at frame k. The nearest neighbor rule associates a measurement with track t as:

m^t=argminm=1,,mt(k)νtm(k)T[St(k)]-1νtm(k).

3.5 State Estimate and Covariance Update

The state estimate and the covariance matrix of targets are updated as follows:

x^t(kk)=x^t(kk-1)+Wt(k)νtm^t(k),Pt(kk)=Pt(kk-1)-Wt(k)St(k)Wt(k)T.

If mt(k) is equal to zero, i.e., no measurement is associated with target t at frame k, they merely become the predictions of the state and the covariance as:

x^t(kk)=x^t(kk-1),Pt(kk)=Pt(kk-1).

3.6 Performance Evaluation

Several metrics are used for performance evaluation, position error, velocity error, and RMSE of position and velocity. The position error of target t at frame k is obtained as:

epxt(k)=x^t(kk)-xttrue(k),epyt(k)=y^t(kk)-yttrue(k),

where xttrue(k) and yttrue(k) are the ground truth of position of target t at frame k, in x and y directions, respectively. The ground truth of positions of targets is obtained manually at each scene. The RMSE of position is obtained as:

RMSEp(t)=1Kt(f)-Kt(s)+1k=Kt(s)Kt(f){[epxt(k)]2+[epyt(k)]2},

where Kt(s) and Kt(f) are the first and last frame where target t is estimated. The velocity error of target t at frame k is obtained in both x and y directions as:

evxt(k)=v^tx(kk)-vtxtrue(k),evyt(k)=v^ty(kk)-ytytrue(k),

where the ground truth of the velocity is approximated as:

vtxtrue(k)=12δΔ[xttrue(k+δ)-xttrue(k-δ)],vtytrue(k)=12δΔ[yttrue(k+δ)-yttrue(k-δ)].

The RMSE of velocity is obtained as:

RMSEv(t)=1Kt(f)-Kt(s)+1k=Kt(s)Kt(f){[evxt(k)]2+[evyt(k)]2}.

4. Results

4.1 Scenario Description

A drone (Phantom 4 Advanced) captures three moving vehicles at 30 frames per second. A total of 372 frames are captured for more than 12 seconds. The size of one frame is 4096×2160 pixels; the frame is reduced to 20% size, 820×432 pixels, for efficient image processing. The drone hovers (stays still in the air) while capturing video sequences at a height of 92 m above the road. The camera points directly downwards while capturing video sequences. One pixel corresponds to 0.1344 m. There are three cars moving, with different initialization positions and maneuverings. The first car (Car 1) is located at the center of the scene from the first frame and moves to the right until it disappear after the 155th frame. The second car (Car 2) moves to the left from the first frame to the 289th frame. The third car (Car 3) moves in the same direction as the second car from the 74th to the 372nd frame. Figure 2(a) shows Cars 1 and 2 at the first frame, Figure 2(b) shows Cars 1–3 at the 150th frame, and Figure 2(c) shows Car 3 at the 300th frame. In the figures, the red circles indicate the moving vehicles.

4.2 Detection of Moving Objects

A total of 368 frames from the 5th frame to the 372nd frame is processed to detect moving objects, since kd in Eq. (1) is set at 4, θT in Eq. (1) is set at 30, and E and D in Eqs. (2) and (3) are set at [1]9×9 and [1]15×15, respectively. The parameters for thresholding and morphological operations are determined heuristically when better results are produced; θs in Eq. (4) is set at 90 considering the true size of moving vehicles. The first row in Figure 3 shows the detection process of Figure 2(a). The second and third rows are of Figure 2(b) and (c), respectively. Figure 3(a) and (b), respectively, are the frame differencing and thresholding to generate a binary image. Figure 3(c) shows the morphological filtering (erosion and dilation) result; the centroids of the segmented areas are marked with an asterisk (*). The ROI window is a rectangle that includes disjoint detected areas. Figure 3(d) shows the segmented areas by the ROI windows after removing false ROIs according to Eq. (4).

Figure 4 shows all centroids for 368 frames. Table 1 shows the detection rate of three cars. Average detection rate is around 96%. The low detection rate of Car 3 is mainly due to the tree occlusion and the same background color (white sidewalk) as the target. A total of 10 false alarms are detected, and two of them are caused by the split measurements from one car.

4.3 Multiple Target Tracking

The centroids in Figure 4 become input measurements for tracking. The sampling time Δ in Eq. (6) is 0.033 s since the frame rate is 30 fps. The standard deviation of process and the measurement noise are set at σx = σy = 30 m/s2 and rx = ry = 1.5 m, respectively; γ for the gating test is set at 4; and σx and σy are conventionally set between 0.5amax and amax where amax is the maximum acceleration of the target [14]; and rx and ry are set at half the length of the vehicle. The validation region in Eq. (14) is chi-square distributed with the degree of freedom is equal to the dimension of measurement vectors, thus the probability mass corresponds to 86.5% when γ is set at 4. The maximum speed for speed gating of two-point differencing initialization is set at 30 m/s. The initial-state vectors of Cars 1 and 2 are initialized by the measured position in the 5th and 6th frames; the initial state of Car 3 is initialized by the 74th and 75th frames. Cars 1, 2, and 3 are estimated until 155th, 289th, and 372nd frame for 150, 284, and 298 frames, respectively, including initialization. Figure 5(a)–(c) are the results of tracking Cars 1–3, respectively. No false tracks are established, and no track termination scheme is applied.

Figure 6(a) shows the ground truth of the position of Car 1 from the first to the 155th frame. Figures 6(b) and c are the position vectors of Cars 2 and 3, respectively, from the first frame to the 289th frame, and from the 74th frame to the 372nd frame. Figure 7(a)–(c) shows the approximated ground truth of the velocity in x and y directions of Car 1, 2, and 3, obtained by Eq. (24); δis set at 26 when the least average velocity RMSE is produced. Figures 8 and 9 show the position in Eq. (21) and velocity errors in Eq. (23), respectively. Large position errors in x direction are mainly caused by measurement errors. High speed generates greater measurement errors. Therefore, measurement errors during detection are bigger in the direction of the car’s movement. Velocity errors are generated during tracking. It is noted that the velocity error differences in x and y directions are small compared with those of position errors.

The sampling time is set differently in the following experiments. Figure 10 shows the tracking result with Δ = 0.1 s, that is, measurements are acquired every third frame from the original data. Cars 1 and 2 are initialized by the 5th and 8th frames and Car 3 is initialized by the 74th and 77th frames. Cars 1, 2, and 3 are estimated until 155th, 287th, and 371th frame, respectively, thus Cars 1, 2, and 3 are estimated for 50, 94, and 99 frames, respectively, including initialization. Figure 11 is the tracking results with every sixth frame (Δ = 0.2 s) updated. Cars 1 and 2 are initialized by the 5th and 11th frames and Car 3 is initialized by the 77th and 83rd frames. Cars 1, 2, and 3 are estimated for 25, 47, and 49 frames, respectively, until 155th, 298th, and 371st frame. The relative computational times are 1, 0.51, and 0.34 for Δ = 0.033, Δ = 0.1, and Δ = 0.2 s, respectively.

Table 2 shows the RMSE of position in Eq. (22). The large position RMSEs of Cars 2 and 3 are due to the high maneuvering of the circular motion. The average position RMSE is bigger over many sampling times, as shown in Table 2. Finer sampling steps can reduce position errors in Car 1 and 2, costing more computational resources. Table 3 shows the RMSE of velocity in Eq. (25). The velocity RMSE is also bigger for Cars 2 and 3 than Car 1. It is shown that the average velocity RMSE decreases as the sampling time increases.

5. Conclusion

In this paper, multiple moving objects are captured over long distances by a UAV. The locations of objects are detected using a background scene. Three targets are tracked with Kalman filtering. Experimental results show that the proposed algorithm detects and tracks moving objects with good accuracy. Position and velocity errors are obtained for different maneuvering targets and sampling times. It is shown that the proposed system can adjust the sampling time to build a more efficient system in computational resources. This property is favorable for drone applications because the computational resources in a drone are very limited. This study also proposes new applications for drones, such as traffic control and smart car as well as security and defense. Tracking a large number of targets with high maneuvering remains for future study.

Acknowledgments

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (Grant Number: 2017R1D1 A3B03031668).

Fig 1.

Figure 1.

Block diagram of moving object detection.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 2.

Figure 2.

(a) Cars 1 and 2 at the first frame; (b) Cars 13 at the 150th frame; (c) Car 3 at the 300th frame. Moving cars are indicated by the red circles.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 3.

Figure 3.

(a) Frame differencing; (b) thresholding; (c) morphological filtering and centroids; (d) segmented region by ROI windows after false alarm removal.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 4.

Figure 4.

Measurements including false alarms for 368 frames.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 5.

Figure 5.

Tracking results whenΔ = 0.033 s: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 6.

Figure 6.

Ground truth of position: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 7.

Figure 7.

Approximated ground truth of velocity: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 8.

Figure 8.

Position error: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 9.

Figure 9.

Velocity error: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 10.

Figure 10.

Tracking results whenΔ = 0.1 s: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Fig 11.

Figure 11.

Tracking results whenΔ = 0.2 s: (a) Car 1, (b) Car 2, (c) Car 3.

The International Journal of Fuzzy Logic and Intelligent Systems 2018; 18: 182-189https://doi.org/10.5391/IJFIS.2018.18.3.182

Table 1 . Detection rates.

Number of framesNumber of detectionsDetection rate (%)
Car 1151151100
Car 228528298.95
Car 329927391.3
Total/average73570696.05

Table 2 . Position RMSE.

Sampling time (s)RMSE (m)Average
Car 1Car 2Car 3
0.0330.8851.6191.5671.345
0.10.891.6841.5611.378
0.20.9251.6861.6361.416

Table 3 . Velocity RMSE.

Sampling time (s)RMSE (m/s)Average
Car 1Car 2Car 3
0.0331.4721.712.1141.765
0.11.3771.9831.8441.734
0.21.2872.0152.3661.89

References

  1. Hampapur, A, Brown, L, Connell, J, Pankanti, S, Senior, AW, and Tian, YL . Smart surveillance: applications, technologies and implications., Proceedings of the 2003 Joint Conference of the 4th International Conference on Information, Communications and Signal Processing and 4th Pacific Rim Conference on Multimedia, 2003, Singapore, Array, pp.1133-1138. https://doi.org/10.1109/ICICS.2003.1292637
    CrossRef
  2. Kumar, R, Sawhney, H, Samarasekera, S, Hsu, S, Tao, H, and Guo, Y (2001). Aerial video surveillance and exploitation. Proceedings of the IEEE. 89, 1518-1539. https://doi.org/10.1109/5.959344
    CrossRef
  3. Chung, W, Kim, Y, Kim, YJ, and Kim, D . A two-stage foreground propagation for moving object detection in a non-stationary., Proceedings of 2016 13th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), 2016, Colorado Springs, CO, Array, pp.187-193. https://doi.org/10.1109/AVSS.2016.7738024
    CrossRef
  4. Yeh, CH, Lin, CY, Lai, HE, Muchtar, K, and Sun, MT (2017). Three-pronged compensation and hysteresis thresholding for moving object detection in real-time video surveillance. IEEE Transactions on Industrial Electronics. 64, 4945-4955. https://doi.org/10.1109/tie.2017.2669881
    CrossRef
  5. Wu, Y, He, X, and Nguyen, TQ (2017). Moving object detection with a freely moving camera via background motion subtraction. IEEE Transactions on Circuits and Systems for Video Technology. 27, 236-248. https://doi.org/10.1109/tcsvt.2015.2493499
    CrossRef
  6. Mousse, M, Motamed, C, and Ezin, E . An adaptive algorithm for fast moving object detection from dynamic background based on codebook., Proceedings of 2016 4th International Conference on Robotics and Mechatronics (ICROM), 2016, Tehran, Iran, Array, pp.584-588. https://doi.org/10.1109/ICRoM.2016.7886808
    CrossRef
  7. Zhang, Y, Huang, X, Li, J, Liu, X, Zhang, H, and Xing, X . Research of moving object detection algorithm in transmission lines under complex background., Proceedings of 2016 International Conference on Condition Monitoring and Diagnosis (CMD), 2016, Array, pp.176-179. https://doi.org/10.1109/cmd.2016.7757799
    CrossRef
  8. Olugboja, A, and Wang, Z . Detection of moving objects using foreground detector and improved morphological filter., Proceedings of the 3rd International Conference on Information Science and Control Engineering, 2016, Beijing, China, Array, pp.329-333. https://doi.org/10.1109/icisce.2016.80
    CrossRef
  9. Cheon, M, and Lee, H (2017). Vision-based vehicle detection system applying hypothesis fitting. International Journal of Fuzzy Logic and Intelligent Systems. 17, 58-67. http://dx.doi.org/10.5391/IJFIS.2017.17.2.58
    CrossRef
  10. Yeom, S, Lee, MH, and Cho, IJ . Long-range moving object detection based on background subtraction., Proceedings of the 18th International Symposium on Advanced Intelligent Systems (ISIS), 2017, Daegu, Korea, pp.1082-1085.
  11. Bar-Shalom, Y, and Li, XR (1995). Multitarget-Multisensor Tracking: Principles and Techniques. Storrs CT: YBS Publishing
  12. Stone, LD, Streit, RL, Corwin, TL, and Bell, KL (2014). Bayesian Multiple Target Tracking. Boston, MA: Artech House
  13. Hamid, KR, Talukder, A, and Ehtesanul Islam, AKM (2018). Implementation of fuzzy aided Kalman filter for tracking a moving object in two-dimensional space. International Journal of Fuzzy Logic and Intelligent Systems. 18, 85-96. https://doi.org/10.5391/IJFIS.2018.18.2.85
    CrossRef
  14. Yeom, S, Kirubarajan, T, and Bar-Shalom, Y (2004). Track segment association, fine-step IMM, and initialization with Doppler for improved track performance. IEEE Transactions on Aerospace and Electronic Systems. 40, 293-309. https://doi.org/10.1109/TAES.2004.1292161
    CrossRef
  15. Chaudhary, K, Zhao, M, Shi, F, Chen, X, Okada, K, and Inaba, M . Robust real-time visual tracking using dual-frame deep comparison network integrated with correlation filters., Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2017, Array, pp.6837-6842. https://doi.org/10.1109/IROS.2017.8206604
    CrossRef
  16. Chen, P, Dang, Y, Liang, R, Zhu, W, and He, X (2018). Realtime object tracking on a drone with multi-inertial sensing Data. IEEE Transactions on Intelligent Transportation Systems. 19, 131-139. https://doi.org/10.1109/TITS.2017.2750091
    CrossRef
  17. Bian, C, Yang, Z, Zhang, T, and Xiong, H . Pedestrian tracking from an unmanned aerial vehicle., Proceedings of 2016 IEEE 13th International Conference on Signal Processing (ICSP), 2016, Array, pp.1067-1071. https://doi.org/10.1109/ICSP.2016.7877993
    CrossRef
  18. Haag, K, Dotenco, S, and Gallwitz, F . Correlation filter based visual trackers for person pursuit using a low-cost quadrotor., Proceedings of the 15th International Conference on Innovations for Community Services (I4CS), 2015, Array, pp.1-8. https://doi.org/10.1109/I4CS.2015.7294481
    CrossRef
  19. Risse, B, Mangan, M, Del Pero, L, and Webb, B . Visual tracking of small animals in cluttered natural environments using a freely moving camera., Proceedings of IEEE International Conference on Computer Vision Workshops, 2017, Venice, Italy, Array, pp.2840-2849. https://doi.org/10.1109/ICCVW.2017.335
    CrossRef
  20. Kim, J, and Kim, Y . Moving ground target tracking in dense obstacle areas using UAVs., Proceedings of the 17th IFAC World Congress, 2008, Toulouse, France, pp.8552-8557.
  21. Lee, MH, and Yeom, S . Vehicle tracking based on Kalman filtering with drone imaging., Presented in the 6th International Conference on Green and Human Information Technology (ICGHIT), 2018, Chiang Mai, Thailand.
  22. Lee, MH, and Yeom, S (2018). Multiple target detection and tracking on urban roads with a drone. Journal of Intelligent & Fuzzy Systems. https://doi.org/10.3233/JIFS-169847
    CrossRef
  23. Cho, Y, and Lee, H (2017). A position and velocity estimation using multifarious and multiple sensor fusion. International Journal of Fuzzy Logic and Intelligent Systems. 17, 121-128. https://doi.org/10.5391/IJFIS.2017.17.2.121
    CrossRef
  24. Gonzalez, RC, and Woods, RE (2008). Digital Image Processing. Upper Saddle River, NJ: Pearson Prentice- Hall

Share this article on :

Related articles in IJFIS