<?xml version="1.0" encoding="ISO-8859-1"?><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id>1665-6423</journal-id>
<journal-title><![CDATA[Journal of applied research and technology]]></journal-title>
<abbrev-journal-title><![CDATA[J. appl. res. technol]]></abbrev-journal-title>
<issn>1665-6423</issn>
<publisher>
<publisher-name><![CDATA[Universidad Nacional Autónoma de México, Instituto de Ciencias Aplicadas y Tecnología]]></publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id>S1665-64232012000600015</article-id>
<title-group>
<article-title xml:lang="en"><![CDATA[Object Detection with Vocabularies of Space-time Descriptors]]></article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Hernandez-Heredia]]></surname>
<given-names><![CDATA[Y.]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[González-Linares]]></surname>
<given-names><![CDATA[J.M.ª]]></given-names>
</name>
<xref ref-type="aff" rid="A03"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Guil]]></surname>
<given-names><![CDATA[N.]]></given-names>
</name>
<xref ref-type="aff" rid="A03"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Ortiz]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<xref ref-type="aff" rid="A02"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Hernandez]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Cózar]]></surname>
<given-names><![CDATA[J.R.]]></given-names>
</name>
<xref ref-type="aff" rid="A03"/>
</contrib>
</contrib-group>
<aff id="A01">
<institution><![CDATA[,Universidad de las Ciencias Informáticas Centro de Geoinformática y Señales Digitales ]]></institution>
<addr-line><![CDATA[Habana ]]></addr-line>
<country>Cuba</country>
</aff>
<aff id="A02">
<institution><![CDATA[,Universidad de las Ciencias Informáticas Vicerrectoria de Tecnología ]]></institution>
<addr-line><![CDATA[Habana ]]></addr-line>
<country>Cuba</country>
</aff>
<aff id="A03">
<institution><![CDATA[,Universidad de Málaga Escuela Técnica Superior de Ingeniería Informática Departamento de Arquitectura de Computadores]]></institution>
<addr-line><![CDATA[Málaga ]]></addr-line>
<country>España</country>
</aff>
<pub-date pub-type="pub">
<day>00</day>
<month>12</month>
<year>2012</year>
</pub-date>
<pub-date pub-type="epub">
<day>00</day>
<month>12</month>
<year>2012</year>
</pub-date>
<volume>10</volume>
<numero>6</numero>
<fpage>950</fpage>
<lpage>956</lpage>
<copyright-statement/>
<copyright-year/>
<self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_arttext&amp;pid=S1665-64232012000600015&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_abstract&amp;pid=S1665-64232012000600015&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_pdf&amp;pid=S1665-64232012000600015&amp;lng=en&amp;nrm=iso"></self-uri><abstract abstract-type="short" xml:lang="en"><p><![CDATA[This paper presents a novel framework for objects detection in security and broadcast videos. Our method assumes that object classes are unknown in advance and exploit the temporal-space properties of the videos for the creation of a vocabulary that describes these classes. Local space-time features have recently became a popular video representation for action recognition and object detection. Several methods for feature localization and description have been proposed in the literature and promising recognition results were demonstrated for a number of action classes. In this work we propose the use of different kinds of descriptors for the creation of vocabularies for different detection object task. For a better description of the videos we carry out a background model, tryring to clean up and follow the areas where there are objects. The points of interest in the videos to characterize the objects are calculated with a temporary variant of the famous Harris corner detector. With the descriptors obtained from the points of interest, a vocabulary is realized usingthe kinds of videos we want to train. Then we obtained the frequency histograms between the videos for training and the vocabulary so, with a binary classifier obtain the trained classes and following the same procedure without the vocabulary realized the detection and monitoring of the objects. The new method presented is also compared with a state of the art method, obtaining better results in both accuracy and false object rejection.]]></p></abstract>
<abstract abstract-type="short" xml:lang="es"><p><![CDATA[Este artículo presenta un método novedoso para la detección de objetos en videos de seguridad y de transmisión de televisión. Nuestro método supone que las clases de objetos son desconocidas por adelantado y explota las propiedades temporales y espaciales de los videos para la creación de un vocabulario que describe estas clases. Las características locales del espacio y el tiempo se han convertido recientemente en una representación popular de los vídeos para el reconocimiento de acciones y la detección objetos. En estudios recientes se han propuesto varios métodos para la localización y descripción de características de videos y han demostrado resultados prometedores de reconocimiento para clases de acción de personas y objetos. En este trabajo proponemos el uso de diferentes tipos de descriptores para la creación de vocabularios para tareas de detección de objetos diferentes. Para una mejor descripción de los videos generamos el modelo del fondo para tratar de limpiar y seguir las zonas donde están los objetos. Los puntos de interés de los videos para caracterizar a los objetos se calculan con una variante temporal del famoso detector de esquinas Harris. Con los descriptores obtenidos de los puntos de interés se realiza un vocabulario con las clases de videos que se quieran entrenar. Luego se obtienen los histogramas de frecuencia entre los videos de entrenamiento y el vocabulario para con un clasificador binario obtener las clases entrenadas y siguiendo el mismo procedimiento sin el vocabulario realizar la detección y seguimiento de los objetos. El nuevo método presentado también se compara con propuestas actuales para situaciones similares, obteniendo mejores resultados en la precisión y el rechazo de objetos falsos.]]></p></abstract>
<kwd-group>
<kwd lng="en"><![CDATA[object detection]]></kwd>
<kwd lng="en"><![CDATA[video segmentation]]></kwd>
<kwd lng="en"><![CDATA[vocabulary]]></kwd>
<kwd lng="en"><![CDATA[binary classifier]]></kwd>
</kwd-group>
</article-meta>
</front><body><![CDATA[  	    <p align="center"><font face="verdana" size="4"><b>Object Detection with Vocabularies of Space&#45;time Descriptors</b></font></p>  	    <p align="center"><font face="verdana" size="2">&nbsp;</font></p>  	    <p align="center"><font face="verdana" size="2"><b>Y. Hernandez&#45;Heredia*<sup>1</sup>, J.M.<sup>a</sup> Gonz&aacute;lez&#45;Linares<sup>3</sup>, N. Guil<sup>3</sup>, J. Ortiz<sup>2</sup>, R. Hernandez<sup>1</sup>, J.R. C&oacute;zar<sup>3</sup></b></font></p>  	    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>  	    <p align="justify"><font face="verdana" size="2"><sup>1</sup> <i>Centro de Geoinform&aacute;tica y Se&ntilde;ales Digitales Universidad de las Ciencias Inform&aacute;ticas Cuba, Habana.</i> *<a href="mailto:yhernandezh@uci.cu">yhernandezh@uci.cu</a>.</font></p>  	    <p align="justify"><font face="verdana" size="2"><sup>2</sup> <i>Vicerrectoria de Tecnolog&iacute;a Universidad de las Ciencias Inform&aacute;ticas Cuba, Habana</i>.</font></p>  	    <p align="justify"><font face="verdana" size="2"><sup>3</sup> <i>Departamento de Arquitectura de Computadores Universidad de M&aacute;laga, E.T.S.I. Inform&aacute;tica Espa&ntilde;a, M&aacute;laga.</i></font></p>  	    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>  	    <p align="justify"><font face="verdana" size="2"><b>ABSTRACT</b></font></p>  	    ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2">This paper presents a novel framework for objects detection in security and broadcast videos. Our method assumes that object classes are unknown in advance and exploit the temporal&#45;space properties of the videos for the creation of a vocabulary that describes these classes. Local space&#45;time features have recently became a popular video representation for action recognition and object detection. Several methods for feature localization and description have been proposed in the literature and promising recognition results were demonstrated for a number of action classes.</font></p>  	    <p align="justify"><font face="verdana" size="2">In this work we propose the use of different kinds of descriptors for the creation of vocabularies for different detection object task. For a better description of the videos we carry out a background model, tryring to clean up and follow the areas where there are objects. The points of interest in the videos to characterize the objects are calculated with a temporary variant of the famous Harris corner detector. With the descriptors obtained from the points of interest, a vocabulary is realized usingthe kinds of videos we want to train. Then we obtained the frequency histograms between the videos for training and the vocabulary so, with a binary classifier obtain the trained classes and following the same procedure without the vocabulary realized the detection and monitoring of the objects.</font></p>  	    <p align="justify"><font face="verdana" size="2">The new method presented is also compared with a state of the art method, obtaining better results in both accuracy and false object rejection.</font></p>  	    <p align="justify"><font face="verdana" size="2"><b>Keywords:</b> object detection, video segmentation, vocabulary, binary classifier.</font></p>  	    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>  	    <p align="justify"><font face="verdana" size="2"><b>RESUMEN</b></font></p>  	    <p align="justify"><font face="verdana" size="2">Este art&iacute;culo presenta un m&eacute;todo novedoso para la detecci&oacute;n de objetos en videos de seguridad y de transmisi&oacute;n de televisi&oacute;n. Nuestro m&eacute;todo supone que las clases de objetos son desconocidas por adelantado y explota las propiedades temporales y espaciales de los videos para la creaci&oacute;n de un vocabulario que describe estas clases. Las caracter&iacute;sticas locales del espacio y el tiempo se han convertido recientemente en una representaci&oacute;n popular de los v&iacute;deos para el reconocimiento de acciones y la detecci&oacute;n objetos. En estudios recientes se han propuesto varios m&eacute;todos para la localizaci&oacute;n y descripci&oacute;n de caracter&iacute;sticas de videos y han demostrado resultados prometedores de reconocimiento para clases de acci&oacute;n de personas y objetos.</font></p>  	    <p align="justify"><font face="verdana" size="2">En este trabajo proponemos el uso de diferentes tipos de descriptores para la creaci&oacute;n de vocabularios para tareas de detecci&oacute;n de objetos diferentes. Para una mejor descripci&oacute;n de los videos generamos el modelo del fondo para tratar de limpiar y seguir las zonas donde est&aacute;n los objetos. Los puntos de inter&eacute;s de los videos para caracterizar a los objetos se calculan con una variante temporal del famoso detector de esquinas Harris. Con los descriptores obtenidos de los puntos de inter&eacute;s se realiza un vocabulario con las clases de videos que se quieran entrenar. Luego se obtienen los histogramas de frecuencia entre los videos de entrenamiento y el vocabulario para con un clasificador binario obtener las clases entrenadas y siguiendo el mismo procedimiento sin el vocabulario realizar la detecci&oacute;n y seguimiento de los objetos.</font></p>  	    <p align="justify"><font face="verdana" size="2">El nuevo m&eacute;todo presentado tambi&eacute;n se compara con propuestas actuales para situaciones similares, obteniendo mejores resultados en la precisi&oacute;n y el rechazo de objetos falsos.</font></p>  	    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>  	    ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2"><a href="/pdf/jart/v10n6/v10n6a15.pdf" target="_blank">DESCARGAR ART&Iacute;CULO EN FORMATO PDF</a></font></p>  	    <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>  	    <p align="justify"><font face="verdana" size="2"><b><i>References</i></b></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;1&#93; Yingzi, D. Unsupervised approach to color video thresholding. s.l. : Optical Engineering, 2004.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834001&pid=S1665-6423201200060001500001&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;2&#93; Alfredo, M. Vision AIBO, ITAM. 2008.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834003&pid=S1665-6423201200060001500002&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;3&#93; Cipolla, R., et al. Semantic texton forests forimage categorization and segmentation. s.l. : Computer Vision and Pattern Recognition, CVPR08. IEEE Conference, 2008.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834005&pid=S1665-6423201200060001500003&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;4&#93; Laptev, I. histograms, Improving object detection with boosted. INRIA Rennes, France : Image and Vision Computing, 2009.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834007&pid=S1665-6423201200060001500004&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;5&#93; Fergus, R., et al. Object class recognition by unsupervised scale&#45;invariant learning. Oxford, UK : Computer Vision and Pattern Recognition, CVPR03 IEEE Computer Society Conference, 2003.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834009&pid=S1665-6423201200060001500005&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;6&#93; Laptev, I., et al. Learning realistic human actions from movies. Anchorage, Alaska, USA : Computer Vision and Pattern Recognition, CVPR08., 2008.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834011&pid=S1665-6423201200060001500006&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;7&#93; Burl, M.C., et al. probabilistic approach to object recognition using local photometry and global geometry. s.l. : Computer Vision, ECCV98, 1998.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834013&pid=S1665-6423201200060001500007&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;8&#93; Weber, M., et al. Unsupervised Learning of Models for Recognition. Dublin, Ireland : Computer Vision &#45; ECCV 2000, 2000. 978&#45;3&#45;540&#45;67685&#45;0.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834015&pid=S1665-6423201200060001500008&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;9&#93; Laptev, I. and Lindeberg, T. On Space&#45;time interest points. Springer : International Journal of Computer Vision, Kluwer Academic Publishers, 2005. 0920&#45;5691.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834017&pid=S1665-6423201200060001500009&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;10&#93; Willems, G. and Tuytelaars, T. An efficient dense and scale&#45;invariant spatio&#45;temporal interest point detector. s.l. : Lecture Notes in Computer Science, 2008. 978&#45;3&#45;540&#45;88685&#45;3.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834019&pid=S1665-6423201200060001500010&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;11&#93; Dollar, P., et al. Behavior recognition via sparse spatio&#45;temporal features. USA : Visual Surveillance and Performance Evaluation of Tracking and Surveillance. 2nd Joint IEEE International Workshop, 2005.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834021&pid=S1665-6423201200060001500011&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;12&#93; Triggs, B. and Dalal, N. Histograms of oriented gradients for human detection. San Diego, CA, USA : Computer Vision and Pattern Recognition, CVPR05, 2005.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834023&pid=S1665-6423201200060001500012&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;13&#93; Lazebnik, S., et al. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. New York, NY, USA : Computer Vision and Pattern Recognition, CVPR06, 2006.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834025&pid=S1665-6423201200060001500013&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;14&#93; Gool, L.V. Bag of visual words model: recognizing object categories. England : Oxford University, 2007.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834027&pid=S1665-6423201200060001500014&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;15&#93; Wallach, H.M. Topic Modeling: Beyond Bag&#45;of&#45;Words. New York USA : ICML '06 Proceedings of the 23rd international conference on Machine learning, 2006. 1&#45;59593&#45;383&#45;2.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834029&pid=S1665-6423201200060001500015&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;16&#93; Lempitsky, V. and Gall, J. Class&#45;Specific Hough Forests for Object Detection. Miami, FL, USA : Computer Vision and Pattern Recognition, CVPR09, 2009.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834031&pid=S1665-6423201200060001500016&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;17&#93; Herbrich, R. Bayes Point Machines. Department of Engineering Mathematics, Bristol University, United Kingdom: The Journal of Machine Learning Research, 2001.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834033&pid=S1665-6423201200060001500017&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;18&#93; R&uuml;ping, S. mySVM &#45; a support vector machine. 2010.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834035&pid=S1665-6423201200060001500018&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;19&#93; Dragonfly Interactive. Nec Laboratories, INC America. 2008.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834037&pid=S1665-6423201200060001500019&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;20&#93; Thorsten, J. Support Vector Machine for Complex Outputs. 2008.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834039&pid=S1665-6423201200060001500020&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>  	    <!-- ref --><p align="justify"><font face="verdana" size="2">&#91;21&#93; Chang, C. and Lin, C. LIBSVM &#45; A Library for Support Vector Machines. s.l. : ACM Transactions on Intelligent Systems and Technology (TIST), 2011.    &nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=4834041&pid=S1665-6423201200060001500021&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --></font></p>      ]]></body><back>
<ref-list>
<ref id="B1">
<label>1</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Yingzi]]></surname>
<given-names><![CDATA[D.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Unsupervised approach to color video thresholding. s.l.]]></article-title>
<source><![CDATA[Optical Engineering]]></source>
<year>2004</year>
</nlm-citation>
</ref>
<ref id="B2">
<label>2</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Alfredo]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
</person-group>
<source><![CDATA[]]></source>
<year>2008</year>
<publisher-name><![CDATA[ITAM]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B3">
<label>3</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Cipolla]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Semantic texton forests forimage categorization and segmentation. s.l.]]></article-title>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR08. IEEE Conference]]></source>
<year>2008</year>
</nlm-citation>
</ref>
<ref id="B4">
<label>4</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Laptev]]></surname>
<given-names><![CDATA[I.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[histograms, Improving object detection with boosted. INRIA Rennes, France]]></article-title>
<source><![CDATA[Image and Vision Computing]]></source>
<year>2009</year>
</nlm-citation>
</ref>
<ref id="B5">
<label>5</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Fergus]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Object class recognition by unsupervised scale-invariant learning]]></article-title>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR03 IEEE Computer Society Conference]]></source>
<year>2003</year>
<publisher-loc><![CDATA[Oxford ]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B6">
<label>6</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Laptev]]></surname>
<given-names><![CDATA[I.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Learning realistic human actions from movies]]></article-title>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR08]]></source>
<year>2008</year>
<publisher-loc><![CDATA[Anchorage^eAlaska Alaska]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B7">
<label>7</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Burl]]></surname>
<given-names><![CDATA[M.C.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[probabilistic approach to object recognition using local photometry and global geometry. s.l.]]></article-title>
<source><![CDATA[Computer Vision, ECCV98]]></source>
<year>1998</year>
</nlm-citation>
</ref>
<ref id="B8">
<label>8</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Weber]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Unsupervised Learning of Models for Recognition]]></article-title>
<source><![CDATA[Computer Vision - ECCV 2000]]></source>
<year>2000</year>
<publisher-loc><![CDATA[Dublin ]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B9">
<label>9</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Laptev]]></surname>
<given-names><![CDATA[I.]]></given-names>
</name>
<name>
<surname><![CDATA[Lindeberg]]></surname>
<given-names><![CDATA[T.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[On Space-time interest points]]></article-title>
<source><![CDATA[Springer : International Journal of Computer Vision]]></source>
<year>2005</year>
<publisher-name><![CDATA[Kluwer Academic Publishers]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B10">
<label>10</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Willems]]></surname>
<given-names><![CDATA[G.]]></given-names>
</name>
<name>
<surname><![CDATA[Tuytelaars]]></surname>
<given-names><![CDATA[T.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[An efficient dense and scale-invariant spatio-temporal interest point detector]]></article-title>
<source><![CDATA[Lecture Notes in Computer Science]]></source>
<year>2008</year>
</nlm-citation>
</ref>
<ref id="B11">
<label>11</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Dollar]]></surname>
<given-names><![CDATA[P.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Behavior recognition via sparse spatio-temporal features. USA : Visual Surveillance and Performance Evaluation of Tracking and Surveillance]]></article-title>
<source><![CDATA[2nd Joint IEEE International Workshop]]></source>
<year>2005</year>
</nlm-citation>
</ref>
<ref id="B12">
<label>12</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Triggs]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
<name>
<surname><![CDATA[Dalal]]></surname>
<given-names><![CDATA[N.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Histograms of oriented gradients for human detection]]></article-title>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR05]]></source>
<year>2005</year>
<publisher-loc><![CDATA[San Diego^eCA CA]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B13">
<label>13</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lazebnik]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
</person-group>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR06]]></source>
<year>2006</year>
<publisher-loc><![CDATA[New York^eNY NY]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B14">
<label>14</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Gool]]></surname>
<given-names><![CDATA[L.V.]]></given-names>
</name>
</person-group>
<source><![CDATA[Bag of visual words model: recognizing object categories]]></source>
<year>2007</year>
<publisher-name><![CDATA[Oxford University]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B15">
<label>15</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Wallach]]></surname>
<given-names><![CDATA[H.M.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Topic Modeling: Beyond Bag-of-Words]]></article-title>
<source><![CDATA[ICML '06 Proceedings of the 23rd international conference on Machine learning]]></source>
<year>2006</year>
<publisher-loc><![CDATA[New York ]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B16">
<label>16</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Lempitsky]]></surname>
<given-names><![CDATA[V.]]></given-names>
</name>
<name>
<surname><![CDATA[Gall]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Class-Specific Hough Forests for Object Detection]]></article-title>
<source><![CDATA[Computer Vision and Pattern Recognition, CVPR09]]></source>
<year>2009</year>
<publisher-loc><![CDATA[Miami^eFL FL]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B17">
<label>17</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Herbrich]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<source><![CDATA[The Journal of Machine Learning Research]]></source>
<year>2001</year>
</nlm-citation>
</ref>
<ref id="B18">
<label>18</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Rüping]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
</person-group>
<source><![CDATA[mySVM - a support vector machine]]></source>
<year>2010</year>
</nlm-citation>
</ref>
<ref id="B19">
<label>19</label><nlm-citation citation-type="book">
<source><![CDATA[Dragonfly Interactive]]></source>
<year>2008</year>
<publisher-name><![CDATA[Nec Laboratories, INC America]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B20">
<label>20</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Thorsten]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
</person-group>
<source><![CDATA[Support Vector Machine for Complex Outputs]]></source>
<year>2008</year>
</nlm-citation>
</ref>
<ref id="B21">
<label>21</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Chang]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
<name>
<surname><![CDATA[Lin]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[LIBSVM - A Library for Support Vector Machines]]></article-title>
<source><![CDATA[ACM Transactions on Intelligent Systems and Technology (TIST)]]></source>
<year>2011</year>
</nlm-citation>
</ref>
</ref-list>
</back>
</article>
