ADPRL 2011 - ADPRL 2011 2011 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning
View: 1417
Website www.ieee-ssci.org |
Edit Freely
Category ADPRL 2011
Deadline: October 31, 2010 | Date: April 11, 2011-April 15, 2011
Venue/Country: Paris, France
Updated: 2010-08-02 13:25:21 (GMT+9)
Call For Papers - CFP
ADPRL 20112011 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement LearningAdaptive (or Approximate) dynamic programming (ADP) is a general and effective approach for solving optimal control problems by adapting to uncertain environments over time. ADP optimizes a user-defined cost function with respect to an adaptive control law, conditioned on prior knowledge of the system, and its state, in the presence of system uncertainties. A numerical search over the present value of the control minimizes a nonlinear cost function forward-in-time providing a basis for real-time, approximate optimal control. The ability to improve performance over time subject to new or unexplored objectives or dynamics has made ADP an attractive approach in a number of application domains including optimal control and estimation, operation research, and computational intelligence. ADP is viewed as a form of reinforcement learning based on an actor-critic architecture that optimizes a user-prescribed value online and obtains the resulting optimal control policy. Reinforcement learning (RL) algorithms learn to optimize an agent by letting it interact with an environment and learn from its received feedback. The goal of the agent is to optimize its accumulated reward over time, and for this it estimates value functions that predict its future reward intake when executing a particular policy. Reinforcement learning techniques can be combined with many different function approximators and do not assume any a priori knowledge about the environment. An important aspect in RL is that an agent has to explore parts of the environment it does not know well, while at the same time it has to exploit its knowledge to maximize its reward intake. RL techniques have already been applied successfully for many problems such as controlling robots, game playing, elevator control, network routing, and traffic light optimization.TopicsThe symposium topics include, but are not limited to:Convergence and performance bounds of ADPComplexity issues in RL and ADPStatistical learning and RL, PAC bounds for RLMonte-Carlo and quasi Monte-Carlo methodsDirect policy search, actor-critic methodsParsimoneous function representationAdaptive feature discoveryLearning rules and architectures for RLSensitivity analysis for policy gradient estimationNeuroscience and biologically inspired controlPartially observable Markov decision processesDistributed intelligent systemsMulti-agent RL systemsMulti-level multi-objective optimization for ADPRLKernel methods and value function representationApplications of ADP and RLSymposium ChairJagannathan Sarangapani, Missouri University of Science and Technology, USASymposium Co-ChairsHuaguang Zhang, Northeastern University, ChinaMarco Wiering, University of Groningen, NetherlandsProgram CommitteeCharles W. Anderson, Colorado State University, USAS. N. Balakrishnan, Missouri University of Science and Technology, USATamer Basar, University of Illinois, USADimitri P. Bertsekas, Massachusetts Institute of Technology, USAMingcong Deng, Okayama University, JapanHai-Bin Duan, Beihang University, ChinaEl-Sayed El-Alfy, King Fahd University of Petroleum and Minerals, Saudi ArabiaXiao Hu, GE Global Research, USADerong Liu, University of Illinois Chicago, USAAbhijit Gosavi, Missouri University of Science and Technology, USAZeng-Guang Hou, Chinese Academy of Sciences, ChinaHossein Javaherian, General Motors, USAGeorge G. Lendaris, Portland State University, USAFrank L. Lewis, University of Texas at Arlington, USAEduardo Morales, INAOE, MexicoRemi Munos, INRIA Lille - Nord Europe, FranceKumpati S. Narendra, Yale University, USAHector D. Patino, Universidad Nacional de San Juan, ArgentinaJan Peters, Max Planck Institute for Biological Cybernetics, GermanyWarren Powell, Princeton University, USAPhilippe Preux, INRIA & CNRS (LIFL), FranceDanil Prokhorov, Toyota Technical Center, USAJennie Si, Arizona State University, USAL. Enrique Sucar, INAOE, MexicoCsaba Szepesvari, University of Alberta, CanadaAntonios Tsourdos, Cranfield University (DCMT), UKG. Kumar Venayagamoorthy, Missouri University of Science and Technology, USADraguna Vrabie, University of Texas at Arlington, USAPaul Werbos, National Science Foundation, USABernard Widrow, Stanford University, USADonald C. Wunsch, Missouri University of Science and Technology, USAGary G. Yen, Oklahoma State University, USA
Keywords: Accepted papers list. Acceptance Rate. EI Compendex. Engineering Index. ISTP index. ISI index. Impact Factor.
Disclaimer: ourGlocal is an open academical resource system, which anyone can edit or update. Usually, journal information updated by us, journal managers or others. So the information is old or wrong now. Specially, impact factor is changing every year. Even it was correct when updated, it may have been changed now. So please go to Thomson Reuters to confirm latest value about Journal impact factor.