April 10, 20081 Simplifying solar harvesting model- development in situated agents using pre-deployment learning and information sharing Huzaifa Zafar.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

Running a model's adjoint to obtain derivatives, while more efficient and accurate than other methods, such as the finite difference method, is a computationally.
Design of the fast-pick area Based on Bartholdi & Hackman, Chpt. 7.
A 2 -MAC: An Adaptive, Anycast MAC Protocol for Wireless Sensor Networks Hwee-Xian TAN and Mun Choon CHAN Department of Computer Science, School of Computing.
Design Guidelines for Maximizing Lifetime and Avoiding Energy Holes in Sensor Networks with Uniform Distribution and Uniform Reporting Stephan Olariu Department.
1 Dynamic Programming Week #4. 2 Introduction Dynamic Programming (DP) –refers to a collection of algorithms –has a high computational complexity –assumes.
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Solving POMDPs Using Quadratically Constrained Linear Programs Christopher Amato.
Infocom'04Ossama Younis, Purdue University1 Distributed Clustering in Ad-hoc Sensor Networks: A Hybrid, Energy-Efficient Approach Ossama Younis and Sonia.
Target Tracking Algorithm based on Minimal Contour in Wireless Sensor Networks Jaehoon Jeong, Taehyun Hwang, Tian He, and David Du Department of Computer.
Presenter: Yufan Liu November 17th,
Adaptive Data Collection Strategies for Lifetime-Constrained Wireless Sensor Networks Xueyan Tang Jianliang Xu Sch. of Comput. Eng., Nanyang Technol. Univ.,
A Heuristic Bidding Strategy for Multiple Heterogeneous Auctions Patricia Anthony & Nicholas R. Jennings Dept. of Electronics and Computer Science University.
Three heuristics for transmission scheduling in sensor networks with multiple mobile sinks Damla Turgut and Lotzi Bölöni University of Central Florida.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
May 14, Organization Design and Dynamic Resources Huzaifa Zafar Computer Science Department University of Massachusetts, Amherst.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
Discriminative Training of Kalman Filters P. Abbeel, A. Coates, M
14-1. Copyright © 2006 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin 14 Capacity Planning and Queuing Models.
Maximum Network lifetime in Wireless Sensor Networks with Adjustable Sensing Ranges Mihaela Cardei, Jie Wu, Mingming Lu, and Mohammad O. Pervaiz Department.
June 6, 2002D.H.J. Epema/PDS/TUD1 Processor Co-Allocation in Multicluster Systems DAS-2 Workshop Amsterdam June 6, 2002 Anca Bucur and Dick Epema Parallel.
A Multi-Agent Learning Approach to Online Distributed Resource Allocation Chongjie Zhang Victor Lesser Prashant Shenoy Computer Science Department University.
1 Algorithms for Bandwidth Efficient Multicast Routing in Multi-channel Multi-radio Wireless Mesh Networks Hoang Lan Nguyen and Uyen Trang Nguyen Presenter:
Preliminary Analysis of the SEE Future Infrastructure Development Plan and REM Benefits.
Lecture 12 Monte Carlo Simulations Useful web sites:
COGNITIVE RADIO FOR NEXT-GENERATION WIRELESS NETWORKS: AN APPROACH TO OPPORTUNISTIC CHANNEL SELECTION IN IEEE BASED WIRELESS MESH Dusit Niyato,
Conference Paper by: Bikramjit Banerjee University of Southern Mississippi From the Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence.
Steady and Fair Rate Allocation for Rechargeable Sensors in Perpetual Sensor Networks Zizhan Zheng Authors: Kai-Wei Fan, Zizhan Zheng and Prasun Sinha.
EShare: A Capacitor-Driven Energy Storage and Sharing Network for Long-Term Operation(Sensys 2010) Ting Zhu, Yu Gu, Tian He, Zhi-Li Zhang Department of.
07/21/2005 Senmetrics1 Xin Liu Computer Science Department University of California, Davis Joint work with P. Mohapatra On the Deployment of Wireless Sensor.
SoftCOM 2005: 13 th International Conference on Software, Telecommunications and Computer Networks September 15-17, 2005, Marina Frapa - Split, Croatia.
Adaptive Tree-based Convergecast Protocol CS 234 Project - Anirudh Ramesh Iyer, Swaroop Kashyap Tiptur Srinivasa, Tameem Anwar Guide - Prof. Nalini Venkatasubramanian,
Resource mediation internals The mediator can decide which resource provider(s) to allocate for a resource request This can be based on  resource cost.
Energy-Aware Scheduling with Quality of Surveillance Guarantee in Wireless Sensor Networks Jaehoon Jeong, Sarah Sharafkandi and David H.C. Du Dept. of.
1 Near-Optimal Play in a Social Learning Game Ryan Carr, Eric Raboin, Austin Parker, and Dana Nau Department of Computer Science, University of Maryland.
1 EnviroStore: A Cooperative Storage System for Disconnected Operation in Sensor Networks Liqian Luo, Chengdu Huang, Tarek Abdelzaher John Stankovic INFOCOM.
Optimal Selection of Power Saving Classes in IEEE e Lei Kong, Danny H.K. Tsang Department of Electronic and Computer Engineering Hong Kong University.
An Energy-Efficient Voting-Based Clustering Algorithm for Sensor Networks Min Qin and Roger Zimmermann Computer Science Department, Integrated Media Systems.
Lecture 4 TTH 03:30AM-04:45PM Dr. Jianjun Hu CSCE569 Parallel Computing University of South Carolina Department of.
An Energy-efficient Task Scheduler for Multi-core Platforms with per-core DVFS Based on Task Characteristics Ching-Chi Lin Institute of Information Science,
L. Bertazzi, B. Golden, and X. Wang Route 2014 Denmark June
1 Real-Time Parking Information on Parking-Related Travel Cost TRIP Internship Presentation 2014 Kory Harb July 24, 2014 Advisor: Dr. Yafeng Yin Coordinator:
Node Reclamation and Replacement for Long-lived Sensor Networks Bin Tong, Wensheng Zhang, and Chuang Wang Department of Computer Science, Iowa State University.
Evaluation of harvest control rules (HCRs): simple vs. complex strategies Dorothy Housholder Harvest Control Rules Workshop Bergen, Norway September 14,
Decision Making Under Uncertainty CMSC 471 – Spring 2041 Class #25– Tuesday, April 29 R&N, material from Lise Getoor, Jean-Claude Latombe, and.
A Dynamic Query-tree Energy Balancing Protocol for Sensor Networks H. Yang, F. Ye, and B. Sikdar Department of Electrical, Computer and systems Engineering.
Po-Yu Chen, Zan-Feng Kao, Wen-Tsuen Chen, Chi-Han Lin Department of Computer Science National Tsing Hua University IEEE ICPP 2011 A Distributed Flow-Based.
Supply Assumptions for Investment Planning Transmission Planning Code Workshop 1 3rd April 2008.
Data Consolidation: A Task Scheduling and Data Migration Technique for Grid Networks Author: P. Kokkinos, K. Christodoulopoulos, A. Kretsis, and E. Varvarigos.
Essential Questions What is biology? What are possible benefits of studying biology? What are the characteristics of living things? Introduction to Biology.
On the Difficulty of Achieving Equilibrium in Interactive POMDPs Prashant Doshi Dept. of Computer Science University of Georgia Athens, GA Twenty.
R. Brafman and M. Tennenholtz Presented by Daniel Rasmussen.
Computacion Inteligente Least-Square Methods for System Identification.
Lead from the front Texas Nodal 1 TDWG Nodal Update – June 6, Texas Nodal Market Implementation Server.
Multiscale energy models for designing energy systems with electric vehicles André Pina 16/06/2010.
Markov Decision Process (MDP)
Reinforcement Learning in POMDPs Without Resets
System Control based Renewable Energy Resources in Smart Grid Consumer
Parallel Programming in C with MPI and OpenMP
Wei Li, Flávia C. Delicato Paulo F. Pires, Young Choon Lee
Effective Social Network Quarantine with Minimal Isolation Costs
Multi-hop Coflow Routing and Scheduling in Data Centers
An Adaptive Middleware for Supporting Time-Critical Event Response
An Agent that plays Pacman
October 6, 2011 Dr. Itamar Arel College of Engineering
Using the queuing-theoretic approximations for the performance of “push” and “pull” production lines to address Design Problems.
Adaptive Choice of Information Sources
Reinforcement Learning Dealing with Partial Observability
Enabling Prediction of Performance
Parallel Programming in C with MPI and OpenMP
Presentation transcript:

April 10, Simplifying solar harvesting model- development in situated agents using pre-deployment learning and information sharing Huzaifa Zafar & Dan Corkill Computer Science Department University of Massachusetts, Amherst

Introduction

Problem Definition: How much energy is this agent going to be able to harvest? Clouds Shading and tilting How can an agent use its neighboring agents in developing its local models? Two agents see the same (if not very close) cloud attenuation Two agents have different shade attenuation at any given time (unless in the deserts of dubai) {30%,20%} {30%,40%} {30%,0%}

Related Work Multi-Agent Reinforcement Learning Extends traditional reinforcement learning to multiple agents Each agent learns local policies given policies of neighboring agents Requires a large observation set and time to converge to optimal policies. Multi-Agent Inductive Learning Learning models by interacting with other agents in the network Each agent shares information with other agents in the network in order to better learn local models. Again, requires a large observation set and time to converge to usable models

Observations Agent performance is reduced while models are learned Is it possible to reduce the time taken in developing local models once an agent has been deployed? How can an agent better take advantage of the observations of its neighboring agents in developing its local models?

PLASMA (Pre-deployment Learning And Situated Model-development in Agents) Two phase strategy Phase 1: A pre-deployment learning phase Define and develop a parameterized model of the environment The parameters of the model - environmental effects Phase 2: Post-deployment model-completion phase Complete the local parameterized model by sharing information among agents

Advantages of the approach Transfer a majority of the, time consuming, learning part to pre-deployment phase Dramatic reduction in the time and observations required to complete models post deployment If information is shared, requires two days of observations to learn complete models

Input: Current time, location (GPS) Energy Harvested depends on: The maximum energy provided no attenuation Cloud attenuation Shade attenuation Tilt of the solar panel Assume geographical location and angle of solar panel to be constant for the lifetime of the agent Combine them into site attenuation Solar Harvesting Model

Observations Two agents have the same (or very close) cloud attenuation at any time of the day Very small chance of two agents having the exact the same shade attenuation at any given time (unless you are in the deserts of Dubai) Maximum energy does not depend on the exact location of the agent (approximate location is enough) The relationship between cloud attenuation and energy harvested does not depend on the environment of the agent Same with site attenuation

PLASMA: Pre-deployment learning phase Learn the maximum observable energy and the relation between attenuations and observed energy Model for the maximum observable energy - a.sin(time) Model for the relation between attenuations and observed energy - a.log -1 (C(t))

PLASMA: Post-deployment model completion Agent 1, Day 2 we have the following equations: Equations from Day = 1000 * (1 - (f(C(t1)) + k(S(t1,e1))) 600 = 1000 * (1 - (f(C(t1)) + k(S(t1,e2))) Equations from Day = 1000 * (1 - (f(C(t2)) + k(S(t2,e1))) 670 = 1000 * (1 - (f(C(t2)) + k(S(t2,e2))) {??,??} {30%,20%} {??,??} {30%,40%} {??,??} {30%,0%}

PLASMA: Diversity - The deserts of Dubai phenomena Cloud attenuations remains exactly the same for consecutive days (in general low likelihood) Site attenuation remains exactly the same across agents (generally low likelihood in most areas) Take away - diversity is important. Probability of there being no diversity is very very low

PLASMA: The know-it-all agent Converged agent shares values with all neighboring agents Neighboring agents can use meaningful values to converge themselves Take away - If one agent converges, all agents will converge {30%,20%} {30%,40%} {30%,0%}

Experiment - I Evaluate PLASMA in a simulated environment 2 Agents, both learning their respective local models. For one of the agents : Shaded for 4 hours Result: PLASMA is able to accurately predict the solar radiation collected for day 3

Experiment II - Load Balancing Benefits of PLASMA in energy dependent load balancing (Kansal et.al.) Each agent can undertake certain task load depending on available energy Agents make load balancing decisions depending on predicted energy levels for the near future 10 Agents; 20 Days; Mean Cloud Attenuation is 20%

Experiment II - Load Balancing Overall utility given no storage capacity and infinite energy storage capacity Min utility = 2; Max utility = 5 -1 utility for unaccomplished task Result: Can maximize utility with and without residual energy storage (compared with Kansal et.al.)

Conclusions Developed a two phase model-development strategy called PLASMA Minimize the time and number of observations required in developing models post-deployment by transferring all the learning to the pre- deployment phase Its all about the diversity (in agent observations) Agents converge On the first day if there exists a converged agent that shares meaningful observations On the second day if there exists an agent that shares two meaningful observations

April 10, Questions??