JSM 2005 - Toronto

Abstract #302724

This is the preliminary program for the 2005 Joint Statistical Meetings in Minneapolis, Minnesota. Currently included in this program is the "technical" program, schedule of invited, topic contributed, regular contributed and poster sessions; Continuing Education courses (August 7-10, 2005); and Committee and Business Meetings. This on-line program will be updated frequently to reflect the most current revisions.

To View the Program:
You may choose to view all activities of the program or just parts of it at any one time. All activities are arranged by date and time.



The views expressed here are those of the individual authors
and not necessarily those of the ASA or its board, officers, or staff.


The Program has labeled the meeting rooms with "letters" preceding the name of the room, designating in which facility the room is located:

Minneapolis Convention Center = “MCC” Hilton Minneapolis Hotel = “H” Hyatt Regency Minneapolis = “HY”

Back to main JSM 2005 Program page



Legend: = Applied Session, = Theme Session, = Presenter
Activity Number: 378
Type: Invited
Date/Time: Wednesday, August 10, 2005 : 10:30 AM to 12:20 PM
Sponsor: ENAR
Abstract - #302724
Title: Using Machine Learning to Approximate Dynamic Programming for Large-scale Sequential Decision Problems
Author(s): Andrew Barto*+
Companies: University of Massachusetts
Address: Department of Computer Science, Amherst, MA, 01003, USA
Keywords: machine learning ; sequential decision problems ; reinforcement learning ; approximate dynamic programming
Abstract:

An active subarea of machine learning addresses methods for efficiently approximating solutions to large-scale stochastic sequential decision problems. Although originally motivated by the goal of building artificial intelligent agents whose activities can be modeled in terms of sequential decisionmaking, research has expanded to encompass a range of methods and applications that make contact with control engineering, operations research, psychology, and neuroscience. In this approach, called Reinforcement Learning, the goal is for an autonomous agent to learn how to increase the amount of reward it attains while interacting with an incompletely-known stochastic environment. The focus is on methods that compute, or approximate, value functions that provide predictions of future reward and can be used in a variety of ways to direct action. The relationship between three basic classes of methods (stochastic dynamic programming, Monte Carlo methods, and temporal-difference methods) provides the framework for understanding what the Machine Learning perspective has added to our collection of methods for finding high-performance policies.


  • The address information is for the authors that have a + after their name.
  • Authors who are presenting talks have a * after their name.

Back to the full JSM 2005 program

JSM 2005 For information, contact jsm@amstat.org or phone (888) 231-3473. If you have questions about the Continuing Education program, please contact the Education Department.
Revised March 2005