<?xml version="1.0" encoding="ISO-8859-1"?><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance">
<front>
<journal-meta>
<journal-id>1405-5546</journal-id>
<journal-title><![CDATA[Computación y Sistemas]]></journal-title>
<abbrev-journal-title><![CDATA[Comp. y Sist.]]></abbrev-journal-title>
<issn>1405-5546</issn>
<publisher>
<publisher-name><![CDATA[Instituto Politécnico Nacional, Centro de Investigación en Computación]]></publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id>S1405-55462009000300002</article-id>
<title-group>
<article-title xml:lang="en"><![CDATA[AsistO: A Qualitative MDP-based Recommender System for Power Plant Operation]]></article-title>
<article-title xml:lang="es"><![CDATA[AsistO: Un Sistema de Recomendaciones basado en MDPs Cualitativos para la Operación de Plantas Generadoras]]></article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Reyes]]></surname>
<given-names><![CDATA[Alberto]]></given-names>
</name>
<xref ref-type="aff" rid="A01"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Sucar]]></surname>
<given-names><![CDATA[L. Enrique]]></given-names>
</name>
<xref ref-type="aff" rid="A02"/>
</contrib>
<contrib contrib-type="author">
<name>
<surname><![CDATA[Morales]]></surname>
<given-names><![CDATA[Eduardo F.]]></given-names>
</name>
<xref ref-type="aff" rid="A02"/>
</contrib>
</contrib-group>
<aff id="A01">
<institution><![CDATA[,Instituto de Investigaciones Eléctricas  ]]></institution>
<addr-line><![CDATA[Cuernavaca Morelos]]></addr-line>
<country>México</country>
</aff>
<aff id="A02">
<institution><![CDATA[,Instituto Nacional de Astrofísica, Óptica y Electrónica  ]]></institution>
<addr-line><![CDATA[Tonantzintla Puebla]]></addr-line>
<country>México</country>
</aff>
<pub-date pub-type="pub">
<day>00</day>
<month>09</month>
<year>2009</year>
</pub-date>
<pub-date pub-type="epub">
<day>00</day>
<month>09</month>
<year>2009</year>
</pub-date>
<volume>13</volume>
<numero>1</numero>
<fpage>5</fpage>
<lpage>20</lpage>
<copyright-statement/>
<copyright-year/>
<self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_arttext&amp;pid=S1405-55462009000300002&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_abstract&amp;pid=S1405-55462009000300002&amp;lng=en&amp;nrm=iso"></self-uri><self-uri xlink:href="http://www.scielo.org.mx/scielo.php?script=sci_pdf&amp;pid=S1405-55462009000300002&amp;lng=en&amp;nrm=iso"></self-uri><abstract abstract-type="short" xml:lang="en"><p><![CDATA[This paper proposes a novel and practical model-based learning approach with iterative refinement for solving continuous (and hybrid) Markov decision processes. Initially, an approximate model is learned using conventional sampling methods and solved to obtain a policy. Iteratively, the approximate model is refined using variance in the utility values as partition criterion. In the learning phase, initial reward and transition functions are obtained by sampling the state-action space. The samples are used to induce a decision tree predicting reward values from which an initial partition of the state space is built. The samples are also used to induce a factored MDP. The state abstraction is then refined by splitting states only where the split is locally important. The main contributions of this paper are the use of sampling to construct an abstraction, and a local refinement process of the state abstraction based on utility variance. The proposed technique was tested in AsistO, an intelligent recommender system for power plant operation, where we solved two versions of a complex hybrid continuous-discrete problem. We show how our technique approximates a solution even in cases where standard methods explode computationally.]]></p></abstract>
<abstract abstract-type="short" xml:lang="es"><p><![CDATA[Este artículo propone una técnica novedosa y práctica de aprendizaje basada en modelos con refinamiento iterativo para resolver procesos de decisión de Markov (MDPs) continuos. Inicialmente, se aprende un modelo aproximado usando métodos de muestreo convencionales, el cual se resuelve para obtener una política. Iterativamente, el modelo aproximado se refina con base en la varianza de los valores de la utilidad esperada. En la fase de aprendizaje, se obtienen las funciones de recompensa inmediata y de transición mediante muestras del tipo estado-acción. Éstas primero se usan para inducir un árbol de decisión que predice los valores de recompensa y a partir del cual se construye una partición inicial del espacio de estados. Posteriormente, las muestras también se usan para inducir un MDP factorizado. Finalmente, la abstracción de espacio de estados resultante se refina dividiendo aquellos estados donde pueda haber cambios en la política. Las contribuciones principales de este trabajo son el uso de datos para construir una abstracción inicial, y el proceso de refinamiento local basado en la varianza de la utilidad. La técnica propuesta fue probada en AsistO, un sistema inteligente de recomendaciones para la operación de plantas generadoras de electricidad, donde resolvimos dos versiones de un problema complejo con variables híbridas continuas y discretas. Aquí mostramos como nuestra técnica aproxima una solución aun en casos donde los métodos estándar explotan computacionalmente.]]></p></abstract>
<kwd-group>
<kwd lng="en"><![CDATA[Recommender systems]]></kwd>
<kwd lng="en"><![CDATA[power plants]]></kwd>
<kwd lng="en"><![CDATA[Markov decision processes]]></kwd>
<kwd lng="en"><![CDATA[abstractions]]></kwd>
<kwd lng="es"><![CDATA[Sistemas de recomendaciones]]></kwd>
<kwd lng="es"><![CDATA[plantas generadoras]]></kwd>
<kwd lng="es"><![CDATA[procesos de decisión de Markov]]></kwd>
<kwd lng="es"><![CDATA[abstracciones]]></kwd>
</kwd-group>
</article-meta>
</front><body><![CDATA[ <p align="justify"><font face="verdana" size="4">Art&iacute;culos</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="4"><b>AsistO: A Qualitative MDP&#150;based Recommender System for Power Plant Operation</b></font></p>     <p align="center"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="3"><b><i>AsistO: Un Sistema de Recomendaciones basado en MDPs Cualitativos para la Operaci&oacute;n de Plantas Generadoras</i></b></font></p>     <p align="center"><font face="verdana" size="2">&nbsp;</font></p>     <p align="center"><font face="verdana" size="2"><b>Alberto Reyes<sup>1</sup>, L. Enrique Sucar<sup>2</sup> and Eduardo F. Morales<sup>2</sup></b></font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><i><sup>1 </sup>Instituto de Investigaciones El&eacute;ctricas; Av. Reforma 113, Palmira, Cuernavaca, Morelos, 62490, M&eacute;xico; <a href="mailto:areyes@iie.org.mx">areyes@iie.org.mx</a> </i></font></p>     <p align="justify"><font face="verdana" size="2"><i><sup>2 </sup>INAOE; Luis Enrique Erro 1, Sta. Ma. Tonantzintla, Puebla 72840, M&eacute;xico; <a href="mailto:esucar@inaoep.mx">esucar@inaoep.mx</a> ,   <a href="mailto:emorales@inaoep.mx">emorales@inaoep.mx</a></i></font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2">Article received on July 15, 2008    <br>   Accepted on April 03, 2009</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>Abstract</b></font></p>     <p align="justify"><font face="verdana" size="2">This paper proposes a novel and practical model&#150;based learning approach with iterative refinement for solving continuous (and hybrid) Markov decision processes. Initially, an approximate model is learned using conventional sampling methods and solved to obtain a policy. Iteratively, the approximate model is refined using variance in the utility values as partition criterion. In the learning phase, initial reward and transition functions are obtained by sampling the state&#150;action space. The samples are used to induce a decision tree predicting reward values from which an initial partition of the state space is built. The samples are also used to induce a factored MDP. The state abstraction is then refined by splitting states only where the split is locally important. The main contributions of this paper are the use of sampling to construct an abstraction, and a local refinement process of the state abstraction based on utility variance. The proposed technique was tested in <i>AsistO, </i>an intelligent recommender system for power plant operation, where we solved two versions of a complex hybrid continuous&#150;discrete problem. We show how our technique approximates a solution even in cases where standard methods explode computationally.</font></p>     <p align="justify"><font face="verdana" size="2"><b>Keywords: </b>Recommender systems, power plants, Markov decision processes, abstractions.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>Resumen</b></font></p>     <p align="justify"><font face="verdana" size="2">Este art&iacute;culo propone una t&eacute;cnica novedosa y pr&aacute;ctica de aprendizaje basada en modelos con refinamiento iterativo para resolver procesos de decisi&oacute;n de Markov (MDPs) continuos. Inicialmente, se aprende un modelo aproximado usando m&eacute;todos de muestreo convencionales, el cual se resuelve para obtener una pol&iacute;tica. Iterativamente, el modelo aproximado se refina con base en la varianza de los valores de la utilidad esperada. En la fase de aprendizaje, se obtienen las funciones de recompensa inmediata y de transici&oacute;n mediante muestras del tipo estado&#150;acci&oacute;n. &Eacute;stas primero se usan para inducir un &aacute;rbol de decisi&oacute;n que predice los valores de recompensa y a partir del cual se construye una partici&oacute;n inicial del espacio de estados. Posteriormente, las muestras tambi&eacute;n se usan para inducir un MDP factorizado. Finalmente, la abstracci&oacute;n de espacio de estados resultante se refina dividiendo aquellos estados donde pueda haber cambios en la pol&iacute;tica. Las contribuciones principales de este trabajo son el uso de datos para construir una abstracci&oacute;n inicial, y el proceso de refinamiento local basado en la varianza de la utilidad. La t&eacute;cnica propuesta fue probada en AsistO, un sistema inteligente de recomendaciones para la operaci&oacute;n de plantas generadoras de electricidad, donde resolvimos dos versiones de un problema complejo con variables h&iacute;bridas continuas y discretas. Aqu&iacute; mostramos como nuestra t&eacute;cnica aproxima una soluci&oacute;n aun en casos donde los m&eacute;todos est&aacute;ndar explotan computacionalmente.</font></p>     ]]></body>
<body><![CDATA[<p align="justify"><font face="verdana" size="2"><b>Palabras clave: </b>Sistemas de recomendaciones, plantas generadoras, procesos de decisi&oacute;n de Markov, abstracciones.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><a href="/pdf/cys/v13n1/v13n1a2.pdf" target="_blank">DESCARGAR ART&Iacute;CULO EN FORMATO PDF</a></font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>Acknowledgments</b></font></p>     <p align="justify"><font face="verdana" size="2">This work was supported jointly by the <i>Instituto de Investigaciones El&eacute;ctricas, </i>Mexico and CONACYT Project No. 47968.</font></p>     <p align="justify"><font face="verdana" size="2">&nbsp;</font></p>     <p align="justify"><font face="verdana" size="2"><b>References</b></font></p>     <!-- ref --><p align="justify"><font face="verdana" size="2">1. <b>J. Baum and A. E. Nicholson. </b>Dynamic non&#150;uniform abstractions for approximate planning in large structured stochastic domains. In <i>PRICAI'98 &#150; Proceedings of the 5th Pacific Rim International Conference on Artificial Intelligence, </i>pages 587&#150;598, Singapore, 1998.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047601&pid=S1405-5546200900030000200001&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">2. <b>R.E. Bellman. </b><i>Dynamic Programming. </i>Princeton U. Press, Princeton, N.J., 1957.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047602&pid=S1405-5546200900030000200002&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">3. <b>D. P. Bertsekas. </b><i>A counter&#150;example to temporal difference learning. </i>Neural Computation, 1994.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047603&pid=S1405-5546200900030000200003&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">4. <b>D. P. Bertsekas and J.N. Tsitsiklis</b>. <i>Neuro&#150;dynamic programming. </i>Athena Sciences, 1996.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047604&pid=S1405-5546200900030000200004&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">5. <b>B. Bonet and J. Pearl. </b>Qualitative MDPs and POMDPs: An order&#150;of&#150;magnitude approach. In <i>Proceedings of the 18th Conf. on Uncertainty in AI, UAI&#150;02, </i>pages 61&#150;68, Edmonton, Canada, 2002.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047605&pid=S1405-5546200900030000200005&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">6. <b>C. Boutilier, T. Dean, and S. Hanks. </b>Decision&#150;theoretic planning: structural assumptions and computational leverage. <i>Journal of AI Research, </i>11:1&#150;94, 1999.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047606&pid=S1405-5546200900030000200006&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">7. <b>C. Boutilier, M. Goldszmidt, and B. Sabata. </b>Continuous value function approximation for sequential bidding policies. In Kathryn Laskey and Henri Prade, editors, <i>Proceedings of the 15th Conference on Uncertainty in Artificial Intelligence (UAI&#150;99), </i>pages 81&#150;90. Morgan Kaufmann Publishers, San Francisco, California, USA, 1999.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047607&pid=S1405-5546200900030000200007&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">8. <b>Elvira Consortium. </b>Elvira: an environment for creating and using probabilistic graphical models. Technical report, U. de Granada, Spain, 2002.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047608&pid=S1405-5546200900030000200008&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">9. <b>G. F. Cooper and E. Herskovits. </b>A bayesian method for the induction of probabilistic networks from data. <i>Machine Learning, </i>1992.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047609&pid=S1405-5546200900030000200009&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">10. <b>T. Dean   and   K. Kanazawa.   </b>A  model  for reasoning  about persistence  and  causation.   <i>Computational Intelligence, </i>5:142&#150;150, 1989.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047610&pid=S1405-5546200900030000200010&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">11. <b>Z. Feng, R. Dearden, N. Meuleau, and R. Washington. </b>Dynamic programming for structured continuous Markov decision problems. <i>In Proc. of the 20th Conf. on Uncertainty in AI (UAI&#150;2004). Banff, Canada, </i>2004.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047611&pid=S1405-5546200900030000200011&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">12. <b>C. Guestrin, M. Hauskrecht, and B. Kveton. </b>Solving factored MDPs with continuous and discrete variables. In <i>Twentieth Conference on Uncertainty in Artificial Intelligence (UAI 2004), </i>Banff, Canada, 2004.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047612&pid=S1405-5546200900030000200012&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">13. <b>M. Hauskrecht and B. Kveton. </b>Linear program approximation for factored continuous&#150;state Markov decision processes. <i>In In Advances in Neural Information Processing Systems NIPS(03), </i>pages 895&#150; 902, 2003.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047613&pid=S1405-5546200900030000200013&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">14. <b>J. Hoey, R. St&#150;Aubin, A. Hu, and C. Boutilier. </b>SPUDD: Stochastic planning using decision diagrams. <i>In Proc. of the 15th Conf. on Uncertainty in AI, UAI&#150;99, </i>pages 279&#150;288, 1999.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047614&pid=S1405-5546200900030000200014&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">15. <b>L. Li and M. L. Littman. </b>Lazy approximation for solving continuous finite&#150;horizon MDPs. In <i>AAAI&#150;05, </i>pages 1175&#150;1180, Pittsburgh, PA, 2005.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047615&pid=S1405-5546200900030000200015&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">16. <b>R. Munos and A. Moore. </b>Variable resolution discretization for high&#150;accuracy solutions of optimal control problems. In Thomas Dean, editor, <i>Proceedings of the 16th International Joint Conference on Artificial Intelligence (IJCAI&#150;99), </i>pages 1348&#150;1355. Morgan Kaufmann Publishers, San Francisco, California, USA, August 1999.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047616&pid=S1405-5546200900030000200016&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">17. <b>J. Pineau, G. Gordon, and S. Thrun. </b>Policy&#150;contingent abstraction for robust control. In <i>Proc. of the 19th Conf. on Uncertainty in AI, UAI&#150;03, </i>pages 477&#150;484, 2003.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047617&pid=S1405-5546200900030000200017&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">18. <b>M. L. Puterman. </b><i>Markov Decision Processes. </i>Wiley, New York, 1994.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047618&pid=S1405-5546200900030000200018&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">19. <b>J.R. Quinlan. </b>Induction of decision trees. <i>Machine Learning, </i>1(1):81<b>&#151;</b>106, 1986.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047619&pid=S1405-5546200900030000200019&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">20. <b>J.R. Quinlan. </b><i>C4.5: Programs for machine learning. </i>Morgan Kaufmann, San Francisco, Calif., USA., 1993.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047620&pid=S1405-5546200900030000200020&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">21. <b>R. S. Sutton and A.G. Barto. </b><i>Reinforcement Learning: An Introduction. </i>MIT Press, 1998.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047621&pid=S1405-5546200900030000200021&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --><!-- ref --><p align="justify"><font face="verdana" size="2">22. <b>I.H. Witten. </b><i>Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations, 2nd Ed. </i>Morgan Kaufmann, USA, 2005.</font>&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;[&#160;<a href="javascript:void(0);" onclick="javascript: window.open('/scielo.php?script=sci_nlinks&ref=2047622&pid=S1405-5546200900030000200022&lng=','','width=640,height=500,resizable=yes,scrollbars=1,menubar=yes,');">Links</a>&#160;]<!-- end-ref --> ]]></body><back>
<ref-list>
<ref id="B1">
<label>1</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Baum]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Nicholson]]></surname>
<given-names><![CDATA[A. E.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Dynamic non-uniform abstractions for approximate planning in large structured stochastic domains]]></article-title>
<source><![CDATA[]]></source>
<year>1998</year>
<conf-name><![CDATA[5th Pacific Rim International Conference on Artificial Intelligence]]></conf-name>
<conf-loc> </conf-loc>
<page-range>587-598</page-range></nlm-citation>
</ref>
<ref id="B2">
<label>2</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Bellman]]></surname>
<given-names><![CDATA[R.E.]]></given-names>
</name>
</person-group>
<source><![CDATA[Dynamic Programming]]></source>
<year>1957</year>
<publisher-loc><![CDATA[Princeton^eN.J. N.J.]]></publisher-loc>
<publisher-name><![CDATA[Princeton U. Press]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B3">
<label>3</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Bertsekas]]></surname>
<given-names><![CDATA[D. P.]]></given-names>
</name>
</person-group>
<source><![CDATA[A counter-example to temporal difference learning]]></source>
<year>1994</year>
<publisher-name><![CDATA[Neural Computation]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B4">
<label>4</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Bertsekas]]></surname>
<given-names><![CDATA[D. P.]]></given-names>
</name>
<name>
<surname><![CDATA[Tsitsiklis]]></surname>
<given-names><![CDATA[J.N.]]></given-names>
</name>
</person-group>
<source><![CDATA[Neuro-dynamic programming: Athena Sciences]]></source>
<year>1996</year>
</nlm-citation>
</ref>
<ref id="B5">
<label>5</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Bonet]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
<name>
<surname><![CDATA[Pearl]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Qualitative MDPs and POMDPs: An order-of-magnitude approach]]></article-title>
<source><![CDATA[]]></source>
<year>2002</year>
<conf-name><![CDATA[18th Conf. on Uncertainty in AI]]></conf-name>
<conf-loc> </conf-loc>
<page-range>61-68</page-range><publisher-loc><![CDATA[Edmonton ]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B6">
<label>6</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Boutilier]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
<name>
<surname><![CDATA[Dean]]></surname>
<given-names><![CDATA[T.]]></given-names>
</name>
<name>
<surname><![CDATA[Hanks]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Decision-theoretic planning: structural assumptions and computational leverage]]></article-title>
<source><![CDATA[Journal of AI Research]]></source>
<year>1999</year>
<volume>11</volume>
<page-range>1-94</page-range></nlm-citation>
</ref>
<ref id="B7">
<label>7</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Boutilier]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
<name>
<surname><![CDATA[Goldszmidt]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Sabata]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Continuous value function approximation for sequential bidding policies]]></article-title>
<person-group person-group-type="editor">
<name>
<surname><![CDATA[Laskey]]></surname>
<given-names><![CDATA[Kathryn]]></given-names>
</name>
<name>
<surname><![CDATA[Prade]]></surname>
<given-names><![CDATA[Henri]]></given-names>
</name>
</person-group>
<source><![CDATA[]]></source>
<year>1999</year>
<conf-name><![CDATA[15th Conference on Uncertainty in Artificial Intelligence]]></conf-name>
<conf-loc> </conf-loc>
<page-range>81-90</page-range><publisher-loc><![CDATA[San Francisco^eCalifornia California]]></publisher-loc>
<publisher-name><![CDATA[Morgan Kaufmann Publishers]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B8">
<label>8</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Consortium]]></surname>
<given-names><![CDATA[Elvira]]></given-names>
</name>
</person-group>
<source><![CDATA[Elvira: an environment for creating and using probabilistic graphical models]]></source>
<year>2002</year>
<publisher-name><![CDATA[Technical reportU. de Granada]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B9">
<label>9</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Cooper]]></surname>
<given-names><![CDATA[G. F.]]></given-names>
</name>
<name>
<surname><![CDATA[Herskovits]]></surname>
<given-names><![CDATA[E.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[A bayesian method for the induction of probabilistic networks from data]]></article-title>
<source><![CDATA[Machine Learning]]></source>
<year>1992</year>
</nlm-citation>
</ref>
<ref id="B10">
<label>10</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Dean]]></surname>
<given-names><![CDATA[T.]]></given-names>
</name>
<name>
<surname><![CDATA[Kanazawa]]></surname>
<given-names><![CDATA[K.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[A model for reasoning about persistence and causation]]></article-title>
<source><![CDATA[Computational Intelligence]]></source>
<year>1989</year>
<volume>5</volume>
<page-range>142-150</page-range></nlm-citation>
</ref>
<ref id="B11">
<label>11</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Feng]]></surname>
<given-names><![CDATA[Z.]]></given-names>
</name>
<name>
<surname><![CDATA[Dearden]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
<name>
<surname><![CDATA[Meuleau]]></surname>
<given-names><![CDATA[N.]]></given-names>
</name>
<name>
<surname><![CDATA[Washington]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Dynamic programming for structured continuous Markov decision problems]]></article-title>
<source><![CDATA[]]></source>
<year></year>
<conf-name><![CDATA[20th Conf. on Uncertainty in AI]]></conf-name>
<conf-date>2004</conf-date>
<conf-loc>Banff </conf-loc>
</nlm-citation>
</ref>
<ref id="B12">
<label>12</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Guestrin]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
<name>
<surname><![CDATA[Hauskrecht]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Kveton]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Solving factored MDPs with continuous and discrete variables]]></article-title>
<source><![CDATA[]]></source>
<year></year>
<conf-name><![CDATA[Twentieth Conference on Uncertainty in Artificial Intelligence]]></conf-name>
<conf-date>2004</conf-date>
<conf-loc>Banff </conf-loc>
</nlm-citation>
</ref>
<ref id="B13">
<label>13</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Hauskrecht]]></surname>
<given-names><![CDATA[M.]]></given-names>
</name>
<name>
<surname><![CDATA[Kveton]]></surname>
<given-names><![CDATA[B.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Linear program approximation for factored continuous-state Markov decision processes]]></article-title>
<source><![CDATA[In Advances in Neural Information Processing Systems NIPS(03)]]></source>
<year>2003</year>
<page-range>895- 902</page-range></nlm-citation>
</ref>
<ref id="B14">
<label>14</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Hoey]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[St-Aubin]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
<name>
<surname><![CDATA[Hu]]></surname>
<given-names><![CDATA[A.]]></given-names>
</name>
<name>
<surname><![CDATA[Boutilier]]></surname>
<given-names><![CDATA[C.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[SPUDD: Stochastic planning using decision diagrams]]></article-title>
<source><![CDATA[]]></source>
<year>1999</year>
<conf-name><![CDATA[15th Conf. on Uncertainty in AI]]></conf-name>
<conf-loc> </conf-loc>
<page-range>279-288</page-range></nlm-citation>
</ref>
<ref id="B15">
<label>15</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Li]]></surname>
<given-names><![CDATA[L.]]></given-names>
</name>
<name>
<surname><![CDATA[Littman]]></surname>
<given-names><![CDATA[M. L.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Lazy approximation for solving continuous finite-horizon MDPs]]></article-title>
<source><![CDATA[]]></source>
<year>2005</year>
<conf-name><![CDATA[ AAAI-05]]></conf-name>
<conf-loc> </conf-loc>
<page-range>1175-1180</page-range><publisher-loc><![CDATA[Pittsburgh^ePA PA]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B16">
<label>16</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Munos]]></surname>
<given-names><![CDATA[R.]]></given-names>
</name>
<name>
<surname><![CDATA[Moore]]></surname>
<given-names><![CDATA[A.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Variable resolution discretization for high-accuracy solutions of optimal control problems]]></article-title>
<person-group person-group-type="editor">
<name>
<surname><![CDATA[Dean]]></surname>
<given-names><![CDATA[Thomas]]></given-names>
</name>
</person-group>
<source><![CDATA[]]></source>
<year>Augu</year>
<month>st</month>
<day> 1</day>
<conf-name><![CDATA[16th International Joint Conference on Artificial Intelligence]]></conf-name>
<conf-loc> </conf-loc>
<page-range>1348-1355</page-range><publisher-loc><![CDATA[San Francisco^eCalifornia California]]></publisher-loc>
<publisher-name><![CDATA[Morgan Kaufmann Publishers]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B17">
<label>17</label><nlm-citation citation-type="confpro">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Pineau]]></surname>
<given-names><![CDATA[J.]]></given-names>
</name>
<name>
<surname><![CDATA[Gordon]]></surname>
<given-names><![CDATA[G.]]></given-names>
</name>
<name>
<surname><![CDATA[Thrun]]></surname>
<given-names><![CDATA[S.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Policy-contingent abstraction for robust control]]></article-title>
<source><![CDATA[]]></source>
<year>2003</year>
<conf-name><![CDATA[19th Conf. on Uncertainty in AI]]></conf-name>
<conf-loc> </conf-loc>
<page-range>477-484</page-range></nlm-citation>
</ref>
<ref id="B18">
<label>18</label><nlm-citation citation-type="">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Puterman]]></surname>
<given-names><![CDATA[M. L.]]></given-names>
</name>
</person-group>
<source><![CDATA[Markov Decision Processes: Wiley]]></source>
<year>1994</year>
<publisher-loc><![CDATA[New York ]]></publisher-loc>
</nlm-citation>
</ref>
<ref id="B19">
<label>19</label><nlm-citation citation-type="journal">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Quinlan]]></surname>
<given-names><![CDATA[J.R.]]></given-names>
</name>
</person-group>
<article-title xml:lang="en"><![CDATA[Induction of decision trees]]></article-title>
<source><![CDATA[Machine Learning]]></source>
<year>1986</year>
<volume>1</volume>
<numero>1</numero>
<issue>1</issue>
<page-range>81-106</page-range></nlm-citation>
</ref>
<ref id="B20">
<label>20</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Quinlan]]></surname>
<given-names><![CDATA[J.R.]]></given-names>
</name>
</person-group>
<source><![CDATA[C4.5: Programs for machine learning]]></source>
<year>1993</year>
<publisher-loc><![CDATA[San Francisco^eCalif. Calif.]]></publisher-loc>
<publisher-name><![CDATA[Morgan Kaufmann]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B21">
<label>21</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Sutton]]></surname>
<given-names><![CDATA[R. S.]]></given-names>
</name>
<name>
<surname><![CDATA[Barto]]></surname>
<given-names><![CDATA[A.G.]]></given-names>
</name>
</person-group>
<source><![CDATA[Reinforcement Learning: An Introduction]]></source>
<year>1998</year>
<publisher-name><![CDATA[MIT Press]]></publisher-name>
</nlm-citation>
</ref>
<ref id="B22">
<label>22</label><nlm-citation citation-type="book">
<person-group person-group-type="author">
<name>
<surname><![CDATA[Witten]]></surname>
<given-names><![CDATA[I.H.]]></given-names>
</name>
</person-group>
<source><![CDATA[Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations]]></source>
<year>2005</year>
<edition>2nd</edition>
<publisher-name><![CDATA[Morgan Kaufmann]]></publisher-name>
</nlm-citation>
</ref>
</ref-list>
</back>
</article>
