Introduction
Multiple object tracking has become an essential module in human–computer interaction [1], automated surveillance [2], and vehicle navigation systems [3, 4]. With the rapid development of sensor technology, event cameras (or dynamic vision sensors, DVSs) have come into the market thanks to their superior performance in fast-moving scenarios and high dynamic scenes. Although many research efforts have been made, the algorithms designed for event cameras are still far from mature. In particular, event cameras generate significantly different data compared to conventional image sensors (see Section 3.1 for details). There exist two technical routes for event data processing and interpretation. The first tries to establish an event-based processing framework, while the second makes use of the off-the-shelf image processing approaches to accelerate the development. The first route may have a better chance to unleash the full potential of the sensor, but most existing methods belong to the second. Although data from event cameras can be approximately represented in frames, it is still not clear that if these data can be efficiently and effectively utilised by traditional and modern machine vision algorithms, such as feature extraction and convolutional neural networks (CNNs).
This work studies the visual tracking problem based on event cameras, Fig. 1 demonstrates the diagram of the proposed framework, which contains two stages. At stage one, offline training has been done based on YOLO detector [5] before tracking, such that object detection results with objects’ types could be obtained. At stage two, online training and tracking are in a simultaneous process. The real-time generated event count images are fed into the offline-trained detector and the correlation filter (CF). After data association, those assignments are sent to a Kalman ensemble as measurements. Meanwhile, the CF also outputs its tracked results to the Kalman ensemble as supplementary measurements. Reinitialisation is triggered once inconsistencies are found between outputs from the ensemble and the CF, such that the ensemble is adaptive to various conditions with changes of appearance and scale. The main contributions are listed as follows:
-
An object tracking framework has been presented for event cameras, aiming at real-time applications for ground vehicles. The tracking framework contains a YOLO-based [5] offline-trained deep detector and a CF-based [6] online-trained tracker, which complements each other.
-
A fusion scheme, which incorporates the results from the detector and the tracker, has been proposed based on the Kalman filter. With automated initialisation and track maintenance strategies in the fusion scheme, real-time intervene-free operations could be attained.
-
Features in event count images have been demonstrated well-learned by deep networks and feature extractors. Robustness and accuracy of the tracker have been presented using data in actual driving scenarios.
[IMAGE OMITTED. SEE PDF]
The paper is organised as follows: Section 2 presents related work in event-based object tracking. Section 3 introduces the event camera and its unique data format, and the essential modules in an online tracker. Then, the proposed tracking framework is elaborated in Section 4, followed by evaluation results in Section 5. Lastly, Section 6 concludes the paper.
Related work
Object tracking has been extensively studied in the past decades. In [7], different implementations for the five building blocks in a tracking framework, namely the motion model, the feature extractor, the observation model, the model updater, and the ensemble post-processor have been evaluated to help design a tracker. Authors in [8] review multiple object tracking approaches thoroughly. A multiple object tracker (MOT) is not a simple combination of single object trackers. In practice, manual initialisation is often infeasible; thus an initialisation scheme needs to be proposed for MOTs. Moreover, a maintenance mechanism is essential in MOTs to create new tracks, delete obsolete tracks and assign consistent labels, since the number of objects may vary at any time. Besides, data association is necessary for assigning detections to trackers such that the ensemble post-processor can update objects’ positions according to latest observations.
Recently, the concept of ‘tracking by detection’ has become increasingly popular owing to its discriminative nature compared to conventional generative model-based trackers [9–11]. Many event-based detectors have been proposed: As an early attempt, a hierarchical spiking model for object recognition has been introduced in [12], where asynchronous data are fed into the spiking neural networks that take advantage of accurate timing from event cameras. An end-to-end object detector has been proposed for DVS in [13], where the feature extraction network is designed by considering unique data structure from DVS; moreover, the adaptive temporal pooling is applied to balance triggered events between rapid and slow motions. A comparison of object detection performance between traditional image frames and event frames has been presented in [14], which shows the advantages of event-based approaches in fast and low-light conditions. Authors in [15] propose two neural network architectures for object detection based on the YOLO [5] detector.
As for event-based object tracking, an event camera has been used to assist MOT in [16], demonstrating superior performance for fast-moving objects. In [17], an expectation maximisation-based probabilistic multi-hypothesis tracker is presented for MOT with false alarm observations. In [18], a moving object detector has been proposed for event cameras. After motion correction, those areas with different event timestamps are deemed as moving objects. Then the detected objects are tracked using a Kalman filter. The similar tracking scheme is also implemented in [19]. An event-driven stereo vision system has been developed in [20], where cluster tracking and three-dimensional (3D) reconstruction are complementarily integrated such that ambiguity could be solved when occlusion occurred. By simple morphological operations, the moving objects are extracted and tracked using a Kalman filter with the constant acceleration model. Authors in [21] propose an online tracker-detector integration framework, where the local tracker is based on a support vector machine classifier, and the user-initialised global detector is responsible for failure recovery.
We have listed some state-of-the-art methods in Table 1 for comparison. Compared with [21], this work considers multiple objects in the tracking framework. Compared with [18, 19], all three papers use the Kalman filter as tracking framework, but this work focuses on improving label continuity and tracking stability. The offline-trained detector not only brings a priori information to enable self-initialisation but also improves final performance by providing drift-free, independent results for each frame. The online-trained tracker provides persistent results once initialised, and is free from drifting issues with the help of self-reinitialisation.
Table 1 Object tracking approaches using event cameras
| No. of objects | Configuration | Tracking scheme | |
| [18] | multiple | monocular | probabilistic |
| [20] | multiple | stereo | event matching-based |
| [21] | single | monocular | optimisation-based |
| [22] | single | monocular | probabilistic |
| [19] | multiple | monocular | probabilistic |
| this work | multiple | monocular | probabilistic |
Preliminaries
In this preliminary section, we first introduce event cameras and their data formats for those readers who are not familiar with event cameras. Then we present how to represent the appearance of a region of interest to generate ‘features’, and how to utilise the features in single object tracking, and finally, how to generalise to MOT problems from a single object tracker.
Event cameras and event frames
Event cameras detect intensity changes at each pixel independently. When the change is larger than a user-defined threshold, an ‘event’ would be triggered, where denote pixel coordinates, I is the intensity, and t represents the timestamp [The explicit data format varies based on the sensor model. Events as where p is the polarity are common in the literature [23].]. From these events, diverse processing techniques can be used to generate event-based data for easier implementation. By aggregating events in an interval as a tuple set , different types of event frames can be obtained. We write the fixed time interval for event frames as ; denotes the timestamp of the first event in each interval; represents the intensity at pixel for the k th frame. An event binary image is
Essentials of the online tracker
As this paper is not a comprehensive review, only those building blocks that have been applied to our framework are explained here for readers’ convenience.
Appearance representation
A natural idea to represent the region of interest on an image is to simply use the intensity values inside the region and write them as the vector form. To include more information, intensities in colour channels instead of grayscale have been used. However, those raw intensities are sometimes too specific to describe the substantial features of the interested region. The histogram of oriented gradients (HOGs) [24, 25] has been demonstrated as a useful feature in object detection. In addition to hand-crafted features, it is noted that learned features [26–28] using data-driven techniques sometimes perform better. YOLO is one of the examples that shows excellent object detection results using deep convolutional features. In the proposed framework, deep features have been used in YOLO, while either intensity features or HOG could be applied to the CF.
Correlation filter
The filter-based object trackers model the problem as regression. Given a set of features with target points as training samples and their corresponding user-defined responses , we aim to train a filter which gives the maximum correlation output on these points. Suppose the regression model is linear , the problem becomes to find the model parameter that leads to minimum squared regression error:
Authors in [6] have noted that the regression problem could be further accelerated when the training data is composed of cyclic shifts. Moreover, Kernelised CFs (KCFs) have been derived to tackle kernel ridge regression problems. The proposed approach is generalised to multi-channel images, such that more features including HOG could be applied in appearance representation. Considering both speed and accuracy, the KCF is used for online learning in this work.
Data association
One additional concern for MOT compared to single object tracking is the data association, which assigns detected objects to existing tracks. The problem is usually formulated as the minimisation of total costs, which, could be solved by the Hungarian algorithm [30, 31]. Particularly, suppose we have n detections pending assignment to m tracks, and a cost matrix representing the cost of assigning the jth detection to the ith track. The algorithm outputs pairs of tracks and corresponding detections as the first and second columns in matrix respectively, where l denotes the number of successfully paired assignments.
Offline–online multiple object tracking
Offline learning
The offline learning follows the YOLOv3 framework [32] where the ground truth is manually labelled with bounding boxes on the objects in event count images. Although YOLO can detect multiple classes, only cars are considered in off-line training in this work due to a lack of manually-labelled data. The features are then extracted by undergoing the CNN of Darknet-53. Finally, the model is learned or fine-tuned by minimising the loss between the ground truth and the regression results.
Online learning and tracking
The online learning and tracking procedure, as the core contribution of this work, is summarised in Algorithm 1. The proposed tracking framework maintains a list of tracks that keeps necessary properties (such as Kalman filter estimates, type, CF status and age variables) of all track candidates. A priori estimate and error covariance are computed from the function predictKalman, and assignHungarian tries to find the best match between tracks and detections. Then, estimated bounding boxes on the current image are used as positive samples for online learning in correlationFilter. Correction is done in correctKalman to incorporate online CF and offline detector together. The function maintainTracks removes obsolete tracks, creates new tracks and updates their properties according to assignment results. The conditions of track removal is detailed in the Section 4.2.1.
System modelling in Kalman ensemble
As the Kalman filter has been widely used in object tracking [33, 34], the recursive equations are omitted for simplicity. In this work, notation denotes the estimated value; at discrete time k, subscript and represent the predicted (a priori) estimate and the updated (a posteriori) estimate, respectively.
We consider the following linear process model between k and :
The measurement model at time k is written as
It is noted that is not always available as (i) the detector may fail to recognise the object in each frame; or (ii) the data association may not assign the detected bounding boxes to their corresponding tracks correctly. In this case, the ensemble still works in a short time by correcting objects’ position using , which is persistent once initialised. If neither nor is available, the ensemble predicts objects’ position for a certain iterations , then removes those tracks if measurements cannot be recovered.
Cost function in data association
We use the assignment algorithm described in Section 3.2.3. The values in play a central role in algorithm performance. In this work, we modified the cost function [35] at the time k such that both position and type from the detector are considered in assignment:
[IMAGE OMITTED. SEE PDF]
Initialisation and reinitialisation of CF
Since the proposed tracking framework aims to work continuously without users’ intervention, it is necessary to design an initialisation scheme for the CF to deal with new objects. As the CF does not consider scale variation, reinitialisation is required to update objects’ bounding boxes so that the regions out of the objects could be excluded in the CF. Moreover, due to the change of object motion, appearance and illumination conditions, results from the CF may drift unboundedly. The situation also needs to be saved by a reinitialisation scheme. In this work, a Boolean flag initialisedCF with the default status false would be assigned for each track. At time k, the initialisation would be done to set initialisedCF = true if all following conditions are satisfied:
-
there is a reliable track from the Kalman ensemble; and
-
the track is with the status initialisedCF = false; and
-
the measurement is available.
-
the difference of central coordinates between the CF and Kalman ensemble exceeds the user-defined threshold; or
-
the difference of bounding boxes’ size between the CF and Kalman ensemble exceeds the user-defined threshold; or
-
the bounding box from the CF is too close to the borders of image.
Evaluation in driving scenarios
We have tested the approach on self-collected event data, a long sequence of 1663 event count images in urban areas, from CelePixel CeleX5 Sensor. The data show common behaviors of vehicles such as lane change and overtaking. Occlusion and scale variation are frequent due to the heavy traffic. Some typical scenarios in the dataset are shown in Fig. 3. The performance metrics used in this work include the multiple object tracking accuracy (MOTA) [36, 37] and the maximum label number (MLN). MOTA is defined as
[IMAGE OMITTED. SEE PDF]
Although quantitative metrics related to bounding boxes tracking accuracy (e.g. multiple object tracking precision) are not available as the proposed approach focuses on providing smooth and continuous tracking results, the improved performance can still be demonstrated by qualitative results. It is noted that the position error of the proposed tracker is related to the performance of the YOLO detector and the CF since both are providing position measurements, while the errors on bounding box width and height are influenced by YOLO detector solely. Moreover, it is essential to tune Kalman parameters in practice, as the output confidence level in the deep learning-based model is not in the sense of probability but from the proximity to cost function. The full evaluation video has been made available online at .
General results
We firstly run the YOLO detector alone as a baseline, which shows without considering as there is no numbered label for YOLO detection results. With the proposed approach, it is observed that MOTA has been raised to 0.69 when using the linear kernel, HOG feature and track removal threshold . The metrics used for MOTA calculation could be found in Fig. 4. In particular, by implementing the proposed framework, the numbers of FN and TP become much smoother, as shown in Fig. 4 especially from to 750. Moreover, there is no FP for the proposed approach after removing unreliable tracks, at the cost of a slower response to new objects. In the testing sequence, no sudden change of identification indices for the same object has been found; thus, the fragmentation .
[IMAGE OMITTED. SEE PDF]
Fig. 5 shows the relation between MLN and the user-defined parameter . It is obvious that MLN becomes smaller with a larger , because a larger retains the track longer when the detection result is not available. However, a larger would also lead to higher drifting errors in case of no correction from the detector. In this work, by incorporating CF results, the tracking performance could be improved by setting a larger .
[IMAGE OMITTED. SEE PDF]
Without the online-trained tracker, the whole tracking framework still works by considering the output as the sole measurement vector. We have compared the performance between two configurations: (i) the partial configuration without CF; and (ii) the full configuration with offline-online learning. It is noted that the partial configuration only contains single information source on object position. If an object is not detected, its position will be predicted by constant velocity motion model. In the circumstance of false negative detections, the full configuration is more accurate and robust, as demonstrated in Fig. 6. The real-time performance between two configurations are shown in Table 2.
Table 2 Real-time performance of the proposed tracking framework
| Partial configuration | Full configuration | ||
| Feature | Intensity | HOG | |
| 459 FPS | Linear kernal | 55.7 FPS | 50.4 FPS |
| Gaussian kernal | 33.8 FPS | 34.6 FPS |
[IMAGE OMITTED. SEE PDF]
The time consumption covers core operations in the framework such as CF, Kalman filtering, and data association, and does not include YOLO detection as it is executed outside the loop. The FPS is computed on a mobile workstation running MATLAB R2018b, with an Intel Core i5-8250U CPU and 8 GB memory. HOG orientation number and cell size are set to 6 and 4, respectively. Although KCF could be used to improve the performance by bringing features into high-dimensional spaces [6], it lowers computational efficiency. It is noted that the performance gap between different kernels is small when a powerful feature is selected. These results are similar to [7].
Influence of Kalman parameters
It is well known that parameters in Kalman ensemble, namely noise covariance matrices, affect final results. In all experiments, we set
Conclusion
This paper presents a real-time tracking framework based on event count images from event cameras. Compared to conventional tracking-by-detection, an offline-online training scheme has been presented to improve the tracking robustness and accuracy. Deep features, HOG or raw intensity features have been demonstrated effective in the detector and the tracker. A Kalman ensemble, together with the proposed self-initialisation scheme, has been designed to incorporate measurements from the detector and the tracker such that the tracking consistency is enhanced with bounded position drifts.
Although some achievements have been made, the limitations of this work include: (i) offline training is required, which limits objects to pre-defined categories. (ii) Experiments are still at the preliminary stage, where only cars in urban areas have been considered in the detector. In the future, the authors believe that the detector could be enhanced by introducing more data in a variety of environments. As for feature selection in CF, the possibilities of applying other features (such as LBP [38], Haar [39] and other deep features) that describe event count images could be explored. The data association could be improved by adding appearance discrepancies between detected and predicted bounding boxes into the assignment cost function. It is also promising to reduce the time consumption by exploiting data compression mechanisms.
Sridhar, S., Mueller, F., Zollhöfer, M. et al.: ‘Real‐time joint tracking of a hand manipulating an object from RGB‐D input’. European Conf. on Computer Vision, Springer, Amsterdam, The Netherlands, 2016, pp. 294–310
Javed, O., Shah, M.: ‘Tracking and object classification for automated surveillance’. European Conf. on Computer Vision, Springer, Copenhagen, Denmark, 2002, pp. 343–357
Mendes, A., Bento, L.C., Nunes, U.: ‘Multi‐target detection and tracking with a laser scanner’. IEEE Intelligent Vehicles Symp., 2004, Parma, Italy, 2004, pp. 796–801
Zhao, D., Fu, H., Xiao, L. et al.: ‘Multi‐object tracking with correlation filter for autonomous vehicle’, Sensors, 2018, 18, (7), p. 2004 (doi:
Redmon, J., Divvala, S., Girshick, R. et al.: ‘You only look once: unified, real‐time object detection’. Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 2016, pp. 779–788
Henriques, J.F., Caseiro, R., Martins, P. et al.: ‘High‐speed tracking with kernelized correlation filters’, IEEE Trans. Pattern Anal. Mach. Intell., 2014, 37, (3), pp. 583–596 (doi:
Wang, N., Shi, J., Yeung, D.‐Y. et al.: ‘Understanding and diagnosing visual tracking systems’. Proc. of the IEEE Int. Conf. on Computer Vision, Las Condes, Chile, 2015, pp. 3101–3109
Luo, W., Xing, J., Milan, A. et al.: ‘Multiple object tracking: a literature review’. arXiv preprint arXiv:1409.7618, 2014
Okuma, K., Taleghani, A., De Freitas, N. et al.: ‘A boosted particle filter: multitarget detection and tracking’. European Conf. on Computer Vision, Springer, Prague, Czech Republic, 2004, pp. 28–39
Avidan, S.: ‘Ensemble tracking’, IEEE Trans. Pattern Anal. Mach. Intell., 2007, 29, (2), pp. 261–271 (doi:
Breitenstein, M.D., Reichlin, F., Leibe, B. et al.: ‘Robust tracking‐by‐detection using a detector confidence particle filter’. 2009 IEEE 12th Int. Conf. on Computer Vision, Kyoto, Japan, 2009, pp. 1515–1522
Orchard, G., Meyer, C., Etienne‐Cummings, R. et al.: ‘HFirst: a temporal approach to object recognition’, IEEE Trans. Pattern Anal. Mach. Intell., 2015, 37, (10), pp. 2028–2040 (doi:
Li, J., Shi, F., Liu, W. et al.: ‘Adaptive temporal pooling for object detection using dynamic vision sensor’. British Machine Vision Conf. (BMVC), London, UK, 2017
Iacono, M., Weber, S., Glover, A. et al.: ‘Towards event‐driven object detection with off‐the‐shelf deep learning’. 2018 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), Madrid, Spain, 2018, pp. 1–9
Cannici, M., Ciccone, M., Romanoni, A. et al.: ‘Event‐based convolutional networks for object detection in neuromorphic cameras’. arXiv preprint arXiv:1805.07931, 2018
Saner, D., Wang, O., Heinzle, S. et al.: ‘High‐speed object tracking using an asynchronous temporal contrast sensor’. VMV, Darmstadt, Germany, 2014, pp. 87–94
Cheung, B., Rutten, M., Davey, S. et al.: ‘Probabilistic multi hypothesis tracker for an event based sensor’. 2018 21st Int. Conf. on Information Fusion (FUSION), Cambridge UK, 2018, pp. 1–8
Mitrokhin, A., Fermüller, C., Parameshwara, C. et al.: ‘Event‐based moving object detection and tracking’. 2018 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS) IEEE, Madrid, Spain, 2018, pp. 1–9
Barranco, F., Fermuller, C., Ros, E.: ‘Real‐time clustering and multi‐target tracking using event‐based sensors’. 2018 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems (IROS), Madrid, Spain, 2018, pp. 5764–5769
Camuñas‐Mesa, L.A., Serrano‐Gotarredona, T., Ieng, S.‐H. et al.: ‘Event‐driven stereo visual tracking algorithm to solve object occlusion’, IEEE Trans. Neural Netw. Learn. Syst., 2018, 29, (9), pp. 4223–4237 (doi:
Ramesh, B., Zhang, S., Lee, Z.W. et al.: ‘Long‐term object tracking with a moving event camera’. British Machine Vision Conf., Newcastle, UK, 2018, vol. 2
Liu, H., Moeys, D.P., Das, G. et al.: ‘Combined frame‐and event‐based detection and tracking’. 2016 IEEE Int. Symp. on Circuits and Systems (ISCAS), Montreal, Canada, 2016, pp. 2511–2514
Gallego, G., Delbrück, T., Orchard, G. et al.: ‘Event‐based vision: a survey’. CoRR, vol. abs/1904.08405, 2019. Available at http://arxiv.org/abs/1904.08405=0pt
Dalal, N., Triggs, B.: ‘Histograms of oriented gradients for human detection’. Int. Conf. on Computer Vision & Pattern Recognition (CVPR'05), IEEE Computer Society, San Diego, CA, USA, 2005, vol. 1, pp. 886–893
Vondrick, C., Khosla, A., Malisiewicz, T. et al.: ‘Hoggles: visualizing object detection features’. Proc. of the IEEE Int. Conf. on Computer Vision, Darling Harbour, Australia, 2013, pp. 1–8
Wang, N., Yeung, D.‐Y.: ‘Learning a deep compact image representation for visual tracking’, Adv. Neural. Inf. Process. Syst., 2013, 26, pp. 809–817
Danelljan, M., Hager, G., Shahbaz Khan, F. et al.: ‘Convolutional features for correlation filter based visual tracking’. Proc. of the IEEE Int. Conf. on Computer Vision Workshops, Las Condes, Chile, 2015, pp. 58–66
Yi, K.M., Trulls, E., Lepetit, V. et al.: ‘Lift: learned invariant feature transform’. European Conf. on Computer Vision, Springer, Amsterdam, The Netherlands, 2016, pp. 467–483
Rifkin, R., Yeo, G., Poggio, T. et al.: ‘Regularized least‐squares classification’, Nato Sci. Series Sub Series III Comput. Syst. Sci., 2003, 190, pp. 131–154
Munkres, J.: ‘Algorithms for the assignment and transportation problems’, J. Soc. Ind. Appl. Math., 1957, 5, (1), pp. 32–38 (doi:
Miller, M.L., Stone, H.S., Cox, I.J.: ‘Optimizing murty's ranked assignment method’, IEEE Trans. Aerosp. Electron. Syst., 1997, 33, (3), pp. 851–862 (doi:
Redmon, J., Farhadi, A.: ‘YOLOv3: an incremental improvement’. arXiv preprint arXiv:1804.02767, 2018
Kim, D.Y., Jeon, M.: ‘Data fusion of radar and image measurements for multi‐object tracking via kalman filtering’, Inf. Sci., 2014, 278, pp. 641–652 (doi:
Kulikov, G.Y., Kulikova, M.V.: ‘The accurate continuous‐discrete extended Kalman filter for radar tracking’, IEEE Trans. Signal Process., 2015, 64, (4), pp. 948–958 (doi:
Altendorfer, R., Wirkert, S.: ‘Why the association log‐likelihood distance should be used for measurement‐to‐track association’. 2016 IEEE Intelligent Vehicles Symp. (IV), Gotenburg, Sweden, 2016, pp. 258–265
Ristani, E., Solera, F., Zou, R. et al.: ‘Performance measures and a data set for multi‐target, multi‐camera tracking’. European Conf. on Computer Vision, Springer, 2016, Amsterdam, The Netherlands, pp. 17–35
Bernardin, K., Stiefelhagen, R.: ‘Evaluating multiple object tracking performance: the clear mot metrics’, J. Image Video Process., 2008, 2008, p. 1 (doi:
Ojala, T., Pietikäinen, M., Mäenpää, T.: ‘Multiresolution gray‐scale and rotation invariant texture classification with local binary patterns’, IEEE Trans. Pattern Anal. Mach. Intell., 2002, 24, (7), pp. 971–987 (doi:
Viola, P., Jones, M.: ‘Rapid object detection using a boosted cascade of simple features’. IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society Conference on Computer Vision and Pattern Recognition CVPR (1), Kauai, HI, USA, 2001, vol. 1, pp. 511–518
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2020. This work is published under http://creativecommons.org/licenses/by-nc/3.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Compared with conventional image sensors, event cameras have been attracting attention thanks to their potential in environments under fast motion and high dynamic range (HDR). To tackle the lost‐track issue due to fast illumination changes under HDR scene such as tunnels, an object tracking framework has been presented based on event count images from an event camera. The framework contains an offline‐trained detector and an online‐trained tracker which complement each other: The detector benefits from pre‐labelled data during training, but may have false or missing detections; the tracker provides persistent results for each initialised object but may suffer from drifting issues or even failures. Besides, process and measurement equations have been modelled, and a Kalman fusion scheme has been proposed to incorporate measurements from the detector and the tracker. Self‐initialisation and track maintenance in the fusion scheme ensure autonomous real‐time tracking without user intervene. With self‐collected event data in urban driving scenarios, experiments have been conducted to show the performance of the proposed framework and the fusion scheme.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer





