Back to Current Issues

Utility Person Detection and Multi-View Video Tracking Annotation Model

M.Senbagapriya, Dr. P.Sumitra,

PG and Research Department of Computer Science Vivekananda College of Arts and Sciences for Women (Autonomous) Elayamapalayam.
:10.22362/ijcert/2017/v4/i8/xxxx [UNDER PROCESS]

In this thesis a generic methodology for the semi-automatic generation of reliable position annotations for evaluating multi-camera people-trackers on large video data sets. Most of the annotation data are automatically computed, by estimating a consensus tracking result from multiple existing trackers and people detectors and classifying it as either reliable or not. A small subset of the data, composed of tracks with insufficient reliability, is verified by a human using a simple binary decision task, a process faster than marking the correct person position. The proposed framework is generic and can handle additional trackers. In this thesis studied the most commonly used face edge detection techniques of Enhanced Sobel Edge Annotation Algorithm (ESEAA). Higher-level edge detection techniques and appropriate programming tools only facilitate the process but do not make it a simple task.

M.Senbagapriya, & Dr. P.Sumitra. (2017). Utility Person Detection and Multi-View Video Tracking Annotation Model. International Journal of Computer Engineering In Research Trends, 4(8), 346-355. Retrieved from

Keywords : Image processing, Digital Image Processing, Analog Image Processing Two dimensional signals

[1]	M. Liem and D. Gavrila, “A comparative study on multi-person tracking using overlapping cameras,” in-proc. 9th Int. Comput. Vis. Syst., 2013, pp. 203–212.
[2]	C. Vondrick, D. Patterson, and D. Ramanan, “Efficiently scaling up crowdsourced video annotation,” Int. J. Comput. Vis., vol. 101, no. 1, pp. 184–204, Jan. 2013, doi: 10.1007/s11263-012-0564-1.
[3]	A. Utasi and C. Benedek, “A multi-view annotation tool for people detection evaluation,” in-proc. VIGTA, 2012, pp. 1–6.
[4]	L.ˇ Cehovin, M. Kristan, and A. Leonardis, “Is my new tracker better than yours?” in proc. WACV, Mar. 2014, pp. 540–547.
[5]	L. Marcenaro, P. Morerio, and C. S. Regazzoni, “Performance evaluation of multi-camera visual tracking,” in Proc. AVSS, Sep. 2012, pp. 464–469.
[6]	A. Milan, K. Schindler, and S. Roth, “Challenges of ground truth evaluation of multi-target tracking,” in Proc. CVPRW, Jun. 2013, pp. 735–742.
[7]	M. Kristan et al., “The visual object tracking VOT2013 challenge results,” in Proc. ICCVW, Dec. 2013, pp. 98–111.
[8]	S. Vijayanarasimhan and K. Grauman, “Active frame selection for label propagation in videos,” in-proc. ECCV, 2012, pp. 496–509.
[9]	I. Kavasidis, S. Palazzo, R. Di Salvo, D. Giordano, and C. Spampinato, “A semi-automatic tool for detection and tracking ground truth generation in videos,” in Proc. VIGTA, 2012, pp. 1–5.
[10]	I. Kavasidis, S. Palazzo, R. D. Salvo, D. Giordano, and C. Spampinato, “An innovative Web-based collaborative platform for video annotation,” Multimedia Tools Appl., vol. 70, no. 1, pp. 413–432, May 2013.
[11]	 Ajin P Thomas, Sruthi P.S, Jerry Rachel Jacob, Vandana V Nair, Reeba R,” Survey on Different Applications of Image Processing.” International Journal of Computer Engineering In Research Trends.,vol.4,no.2,pp. 13-19,2017. 
[12]	Trisha Chakraborty, Nikita Nalawade, Abhishri Manjre, Akanksha Sarawgi, Pranali P Chaudhari,” Review of Various Image Processing Techniques for Currency Note Authentication.” International Journal of Computer Engineering In Research Trends.,vol.3,no.3,pp. 119-122,2016. 
[13]	Gunjan, Er. Madan Lal,” Investigation of Various Image Steganography Techniques in Spatial Domain.” International Journal of Computer Engineering In Research Trends., vol.3,no.6,pp. 347-351,2016. 
[14]	 G.Prasanthi, A.Somasekhar,” Anti-Theft Tracking and Controlling Of Vehicle According Us.” International Journal of Computer Engineering In Research Trends., vol.2, no.12, pp. 898-903, 2015. 

DOI Link : Not yet assigned

Download :

Refbacks : Currently There are no refbacks

Quick Links


Science Central

Score: 13.30

Submit your paper to [email protected]