Presentation is loading. Please wait.

Presentation is loading. Please wait.

 By Ashwinkumar Ganesan CMSC 601.  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions.

Similar presentations


Presentation on theme: " By Ashwinkumar Ganesan CMSC 601.  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions."— Presentation transcript:

1  By Ashwinkumar Ganesan CMSC 601

2  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions

3  Method for learning by experience  There agent or bot learns by interacting with the environment.  There is reward attached for action taken in a particular state.  GOAL : MAXIMIZE THE REWARD

4

5

6

7  Co-ordinated Group Action is the situation a set of agents perform a single task.  GOAL: To maximize the output or the reward globally.

8  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions

9  Communication i.e. what should the agent communicate and how much should it communicate with other agents.  Optimal Policy i.e. defining an optimal policy of the entire group. Is an optimal policy a set of optimal individual policies for each agent?  How much of the individual policy information of a certain agent is available to the entire group.

10  Create a method for implementing reinforcement learning on co-ordinated group activity efficiently  Modify Reinforcement Learning algorithm to implement group action  Implement the proposed method and measure its efficiency in World of Warcraft.

11  The proposal is to research co-operative group learning under the following conditions: 1. The environment is assumed to partially observable. 2. Each agent in the system is knows the final action taken by the other agent. 3. Agents do not have access the to state information generated by other agents while selecting an action. 4. Agents do not have access to the policies of other agents, to make the decision. 5. The rewards are not linearly separable.

12  World of Warcraft is a large multi-player online game by Blizzard Entertainment  It is game where every player has his own and roams the virtual world, fighting demons, observing the landscape, buy and selling items and interacting with other players  In short, it is a large game with lots of options for tools, skills and levels, making it challenging for Bots.

13 http://world-of-warcraft.en.softonic.com/

14  Today games especially like World Of Warcraft, have a large number of human players who play in groups.  Single as well as multiplayer games have AI engines, but attacks and actions in these games by opponents is still one at a time.  Reinforcement Learning can be implemented real time in these games, to improve the AI over a period of time and customize the games for users  Other applications – Robotics,defense

15  QUICR Method: This method calculates the counterfactual action, which is the action an agent did not take at a time[2].  Least Square Policy Iteration (LSPI) can implemented. The method performs policy iterations using samples instead of an actual policy[3].  FMQ Algorithm. The algorithm is helpful for environments where agents have partial or no observability[4].

16  Inverse reinforcement learning is the exact opposite of reinforcement learning.  Input is the optimal policy or behavior that is expected from the agent  The agent learns to find the reward function based on observed values in the environment

17  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions

18  Implement Reinforcement Learning in 2 parts: 1. Implement Inverse Reinforcement Learning 2. Implement a modified Reinforcement Learning on the rewards learned in step1. Observe Expert Calculate Reward Calculate Policy based on reward Calculate the new Q(a,s) value Observe other agents

19  Calculating the iteration when the agent is known to have optimal policy  Finding the point when to switch from Inverse Learning to Reinforcement Learning  Working with the reward function got from observing the expert and rewards obtained from the environment  Finding a method to observe other agents and “experts”.

20  The metrics will be evaluated against the known methods  Metrics are: 1. Number of states generated 2. Number of iterations required to reach optimal policy 3. Rewards in terms of points earned (in the game) 4. Rate of Convergence to optimal Policy

21  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions

22  We can use Inverse reinforcement learning with reinforcement learning methods to speed up the learning time required for bots.  Improve bot reward functions over time.

23 1. R Bellman, On the Theory of Dynamic Programming, Proceedings of the National Academy of Sciences, 1952 2. Adrian K. Agogino and Kagan Turner, QUICKER Q-Learning in Multi-agent systems. 3. Lihong Li, Micheal L. Littman, Christopher R. Mansley, Online Exploration in Least – Squares policy Iteration. 4. La¨etitia Matignon, Guillaume J. Laurent and Nadine Le Fort- Piat, A study of FMQ heuristic in cooperative multi-agent games 5. Acknowledgement to Prof. Tim Oates for helping with the literature survey.

24


Download ppt " By Ashwinkumar Ganesan CMSC 601.  Reinforcement Learning  Problem Statement  Proposed Method  Conclusions."

Similar presentations


Ads by Google