Title page for etd-0906115-151230


[Back to Results | New Search]

URN etd-0906115-151230
Author Hsuan-yi Chiang
Author's Email Address No Public.
Statistics This thesis had been viewed 5354 times. Download 446 times.
Department Electrical Engineering
Year 2015
Semester 1
Degree Master
Type of Document
Language English
Title Action Segmentation and Learning by Inverse Reinforcement Learning
Date of Defense 2015-10-02
Page Count 70
Keyword
  • Upper Confidence Bounds
  • Adaboost classifier
  • reward function
  • Inverse Reinforcement learning
  • Reinforcement learning
  • Abstract Reinforcement learning allows agents to learn behaviors through trial and error. However, as the level of difficulty increases, the reward function of the mission also becomes harder to be defined. By combining the concepts of Adaboost classifier and Upper Confidence Bounds (UCB), a method based on inverse reinforcement learning is proposed to construct the reward function of a complex mission. Inverse reinforcement learning allows the agent to rebuild a reward function that imitates the process of interaction between the expert and the environment. During the imitation, the agent continuously compares the difference between the expert and itself, and then the proposed methods determines a specific weight for each state via Adaboost. The weight is then combined with the state confidence from UCB to construct an approximate reward function. This thesis uses a state encoding method and action segmentation to simplify the problem, then utilize the proposed method to determine the optimal reward function. Finally, a maze environment and a soccer robot environment simulation are used to validate the proposed method, further to decreasing the computational time.
    Advisory Committee
  • Jin-Ling Lin - chair
  • Ming-Yi Ju - co-chair
  • Yu-Jen Chen - co-chair
  • Kao-Shing Hwang - advisor
  • Files
  • etd-0906115-151230.pdf
  • indicate access worldwide
    Date of Submission 2015-10-06

    [Back to Results | New Search]


    Browse | Search All Available ETDs

    If you have more questions or technical problems, please contact eThesys