Show simple item record

dc.contributor.advisorFagg, Andrew H.
dc.contributor.authorPalmer, Thomas J.
dc.date.accessioned2015-05-07T20:00:26Z
dc.date.available2015-05-07T20:00:26Z
dc.date.issued2015-05-08
dc.identifier.urihttps://hdl.handle.net/11244/14592
dc.description.abstractReal world tasks, in homes or other unstructured environments, require interacting with objects (including people) and understanding the variety of physical relationships between them. For example, choosing where to place a fork at a table requires knowing the correct position and orientation relative to the plate. Further, the quantity of objects and the roles they play might change from one occasion to the next; the variables are not fixed and predefined. For an intelligent agent to navigate this complex space, it needs to be able to identify and focus on just those variables that are relevant. Also, if a robot or other artificial agent can learn such physical relations from its own experience in a task, it can save manual engineering effort and automatically adapt to new situations. Relational learning, while often focused on discrete domains, applies to situations with arbitrary numbers of objects by using existential and/or universal quantifiers from first-order logic. The field of reinforcement learning (RL) addresses learning task execution from scalar rewards based on agent state and action. Relational reinforcement learning (RRL) combines these two fields. In this dissertation, I present an RRL technique emphasizing relations that are merely implicit in multidimensional, continuous object attributes, such as position, color, and size. This technique requires analyzing permutations of possible object comparisons while simultaneously working in the multidimensional spaces defined by their attributes. Existing similar RRL methods query only one dimension at a time, which limits effectiveness when multiple dimensions are correlated. Specifically, I present a representation policy iteration (RPI) method using the spatiotemporal multidimensional relational framework (SMRF) for learning relational decision trees from object attributes. This SMRF-RPI algorithm interleaves the learning of relational representations and of policies for agent action. Further, SMRF-RPI includes support for continuous actions. As a component of the SMRF framework, I also present a novel multiple instance learning (MIL) algorithm, which is able to learn parametric, existential decision volumes within a feature space in a robust manner. Finally, I demonstrate SMRF-RPI on a variety of developmentally motivated blocks world tasks, as well as effective transfer and sample efficient learning in a standard keepaway soccer benchmark task. Both domains involve complicated, simulated world dynamics in continuous space. These experiments demonstrate SMRF-RPI as a promising method for applying RRL techniques in multidimensional, continuous domains.en_US
dc.languageen_USen_US
dc.subjectRelational Reinforcement Learning, Multiple Instance Learning, Symbol Groundingen_US
dc.titleLearning Action-State Representation Forests for Implicitly Relational Worldsen_US
dc.contributor.committeeMemberHougen, Dean F.
dc.contributor.committeeMemberLakshmivarahan, S.
dc.contributor.committeeMemberMcGovern, Amy
dc.contributor.committeeMemberMiller, David P.
dc.contributor.committeeMemberThomas, Rick P.
dc.date.manuscript2015-05-07
dc.thesis.degreePh.D.en_US
ou.groupCollege of Engineering::School of Computer Scienceen_US
shareok.orcid0000-0002-1593-0011en_US
shareok.nativefileaccessrestricteden_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record