{"id":162,"date":"2014-12-06T11:32:58","date_gmt":"2014-12-06T11:32:58","guid":{"rendered":"http:\/\/meyn.coron.us\/?page_id=162"},"modified":"2014-12-06T11:32:58","modified_gmt":"2014-12-06T11:32:58","slug":"feature-selection-for-neuro-dynamic-programming","status":"publish","type":"page","link":"https:\/\/faculty.eng.ufl.edu\/meyn\/publications\/feature-selection-for-neuro-dynamic-programming\/","title":{"rendered":"Feature Selection for Neuro-Dynamic Programming"},"content":{"rendered":"<table border=\"0\" cellspacing=\"1\" cellpadding=\"1\" align=\"center\">\n<tbody>\n<tr>\n<td><em>Neuro-Dynamic Programming<\/em> encompasses techniques from both reinforcement learning and approximate dynamic programming. Feature selection refers to the choice of basis that defines the function class that is required in the application of these techniques. This chapter reviews two popular approaches to neuro-dynamic programming, TD-learning and Q-learning. The main goal of the chapter is to demonstrate how insight from idealized models can be used as a guide for feature selection for these algorithms. Several approaches are surveyed, including fluid and diffusion models, and the application of idealized models arising from mean-field game approximations. The theory is illustrated with several examples.<\/td>\n<td><em>Book chapter,\u00a0<\/em> D. Huang, W. Chen, P. Mehta, S. Meyn, and A. Surana. <a href=\"\/meyn\/assets\/uploads\/2014\/12\/RLADP_Wiley_Features_online.pdf\">Feature selection for neuro-dynamic programming<\/a>. In F. Lewis, editor, Reinforcement Learning and Approximate Dynamic Programming for Feedback Control. Wiley, 2011<\/p>\n<p>&nbsp;<\/td>\n<\/tr>\n<tr>\n<td><\/td>\n<td>\n<div><strong><a title=\"Approximate dynamic programming using fluid and diffusion approximations with applications to power management\" href=\"http:\/\/www.slideshare.net\/spmeyn\/approximate-dynamic-programming-using-fluid-and-diffusion-approximations-with-applications-to-power-management\" target=\"_blank\" rel=\"noopener\">Related Presentation<\/a><\/strong><\/div>\n<\/td>\n<\/tr>\n<tr>\n<td>2013 journal submission<\/p>\n<ul>\n<li><a href=\"\/meyn\/assets\/uploads\/2014\/12\/CDC555_journalJuly2_afterSubmission.pdf\">Approximate Dynamic Programming using Fluid and Diffusion Approximations with Applications to Power Management<\/a><\/li>\n<\/ul>\n<p>Related references from the Illinois Archive:<\/p>\n<ul>\n<li><a href=\"\/meyn\/archive\/spm_files\/TD%20with%20Exploration\/TD2011.html\">TD-Learning with Exploration<\/a><\/li>\n<li><a href=\"\/meyn\/archive\/spm_files\/Q2009\/Q09.html\">\u00a0Q-learning and Pontryagin&#8217;s Minimum Principle<\/a><\/li>\n<li><a href=\"\/meyn\/archive\/spm_files\/TD5552009\/TD555.html\">Approximate Dynamic Programming using Fluid and Diffusion Approximations with Applications to Power Management<\/a><\/li>\n<\/ul>\n<\/td>\n<td>\n<div>@incollection{huachemehmeysur11,<\/div>\n<div>Author = {Huang, D. and Chen, W. and Mehta, P. and Meyn, S. and Surana, A.},<\/div>\n<div>Booktitle = {Reinforcement Learning and Approximate Dynamic Programming for Feedback Control},<\/div>\n<div>Editor = {Lewis, F.},<\/div>\n<div>Publisher = {Wiley},<\/div>\n<div>Title = {Feature Selection for Neuro-Dynamic Programming},<\/div>\n<div>Year = {2011}}<\/div>\n<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n","protected":false},"excerpt":{"rendered":"<p>Neuro-Dynamic Programming encompasses techniques from both reinforcement learning and approximate dynamic programming. Feature selection refers to the choice of basis that defines the function class that is required in the application of these techniques. This chapter reviews two popular approaches to neuro-dynamic programming, TD-learning and Q-learning. The main goal of the chapter is to demonstrate [&hellip;]<\/p>\n","protected":false},"author":1347,"featured_media":0,"parent":27,"menu_order":4,"comment_status":"closed","ping_status":"closed","template":"page-templates\/page-section-nav.php","meta":{"_acf_changed":false,"inline_featured_image":false,"featured_post":"","footnotes":"","_links_to":"","_links_to_target":""},"class_list":["post-162","page","type-page","status-publish","hentry"],"acf":[],"_links":{"self":[{"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/pages\/162","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/pages"}],"about":[{"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/types\/page"}],"author":[{"embeddable":true,"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/users\/1347"}],"replies":[{"embeddable":true,"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/comments?post=162"}],"version-history":[{"count":0,"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/pages\/162\/revisions"}],"up":[{"embeddable":true,"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/pages\/27"}],"wp:attachment":[{"href":"https:\/\/faculty.eng.ufl.edu\/meyn\/wp-json\/wp\/v2\/media?parent=162"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}