1 J.Y Greff, L. Idoumghar and R. Schott TDF && IECN / LORIA - INRIA July 2002 Using Markov Decision Processes to the Frequency Assignment Problem.

Slides:



Advertisements
Similar presentations
Algorithm Design Methods (I) Fall 2003 CSE, POSTECH.
Advertisements

Algorithm Design Methods Spring 2007 CSE, POSTECH.
Approximation Algorithms Chapter 14: Rounding Applied to Set Cover.
C4.5 algorithm Let the classes be denoted {C1, C2,…, Ck}. There are three possibilities for the content of the set of training samples T in the given node.
Agent-based sensor-mission assignment for tasks sharing assets Thao Le Timothy J Norman WambertoVasconcelos
U NIVERSITY OF M ASSACHUSETTS, A MHERST Department of Computer Science Solving POMDPs Using Quadratically Constrained Linear Programs Christopher Amato.
1 Appendix B: Solving TSP by Dynamic Programming Course: Algorithm Design and Analysis.
Decision Theoretic Planning
Parallel Scheduling of Complex DAGs under Uncertainty Grzegorz Malewicz.
Hidden Markov Models Adapted from Dr Catherine Sweeney-Reed’s slides.
Markov Decision Processes
Planning under Uncertainty
Searching for the Minimal Bézout Number Lin Zhenjiang, Allen Dept. of CSE, CUHK 3-Oct-2005
Computability and Complexity 23-1 Computability and Complexity Andrei Bulatov Search and Optimization.
ISEN 601 Location Logistics Dr. Gary M. Gaukler Fall 2011.
SA-1 1 Probabilistic Robotics Planning and Control: Markov Decision Processes.
KI Kunstmatige Intelligentie / RuG Markov Decision Processes AIMA, Chapter 17.
*Sponsored in part by the DARPA IT-MANET Program, NSF OCE Opportunistic Scheduling with Reliability Guarantees in Cognitive Radio Networks Rahul.
[1][1][1][1] Lecture 5-7: Cell Planning of Cellular Networks June 22 + July 6, Introduction to Algorithmic Wireless Communications David Amzallag.
Michael Heusch - IntCP 2006 Modeling and solving of a radio antennas deployment support application with discrete and interval constraints.
Convergence Analysis of Canonical Genetic Algorithm ChungHsiang, Hsueh.
1 Cell Planning of 4G Cellular Networks David Amzallag Computer Science Department, Technion Joint work with Roee Engelberg (Technion), Seffi Naor (Microsoft.
Utrecht, february 22, 2002 Applications of Tree Decompositions Stan van Hoesel KE-FdEWB Universiteit Maastricht
Markov Decision Processes
More RL. MDPs defined A Markov decision process (MDP), M, is a model of a stochastic, dynamic, controllable, rewarding process given by: M = 〈 S, A,T,R.
Sampling Combinatorial Space Using Biased Random Walks Jordan Erenrich, Wei Wei and Bart Selman Dept. of Computer Science Cornell University.
Elements of the Heuristic Approach
MAKING COMPLEX DEClSlONS
Planning and Verification for Stochastic Processes with Asynchronous Events Håkan L. S. Younes Carnegie Mellon University.
Building “ Problem Solving Engines ” for Combinatorial Optimization Toshi Ibaraki Kwansei Gakuin University (+ M. Yagiura, K. Nonobe and students, Kyoto.
General Polynomial Time Algorithm for Near-Optimal Reinforcement Learning Duke University Machine Learning Group Discussion Leader: Kai Ni June 17, 2005.
Constraint Satisfaction Problem:  Used to model constrained combinatorial problems  Important real-world applications: hardware & software verification,
1 11 Subcarrier Allocation and Bit Loading Algorithms for OFDMA-Based Wireless Networks Gautam Kulkarni, Sachin Adlakha, Mani Srivastava UCLA IEEE Transactions.
Introduction to Job Shop Scheduling Problem Qianjun Xu Oct. 30, 2001.
1 Operations Research Prepared by: Abed Alhameed Mohammed Alfarra Supervised by: Dr. Sana’a Wafa Al-Sayegh 2 nd Semester ITGD4207 University.
Markov Decision Processes1 Definitions; Stationary policies; Value improvement algorithm, Policy improvement algorithm, and linear programming for discounted.
1 N -Queens via Relaxation Labeling Ilana Koreh ( ) Luba Rashkovsky ( )
Simulated Annealing.
Efficient Solution Algorithms for Factored MDPs by Carlos Guestrin, Daphne Koller, Ronald Parr, Shobha Venkataraman Presented by Arkady Epshteyn.
Introduction to Reinforcement Learning Dr Kathryn Merrick 2008 Spring School on Optimisation, Learning and Complexity Friday 7 th.
Markov Chain Monte Carlo and Gibbs Sampling Vasileios Hatzivassiloglou University of Texas at Dallas.
Carnegie Mellon University
© 2015 McGraw-Hill Education. All rights reserved. Chapter 19 Markov Decision Processes.
Conformant Probabilistic Planning via CSPs ICAPS-2003 Nathanael Hyafil & Fahiem Bacchus University of Toronto.
George F Luger ARTIFICIAL INTELLIGENCE 6th edition Structures and Strategies for Complex Problem Solving Machine Learning: Probabilistic Luger: Artificial.
MDPs (cont) & Reinforcement Learning
Ant Colonies using Arc Consistency Techniques for the Set Partitioning Problem Broderick Crawford Pontificia Universidad Católica de Valparaíso - Chile.
Automated Planning and Decision Making Prof. Ronen Brafman Automated Planning and Decision Making Fully Observable MDP.
A Formalism for Stochastic Decision Processes with Asynchronous Events Håkan L. S. YounesReid G. Simmons Carnegie Mellon University.
Chapter 13 Backtracking Introduction The 3-coloring problem
Markov Decision Process (MDP)
Possible actions: up, down, right, left Rewards: – 0.04 if non-terminal state Environment is observable (i.e., agent knows where it is) MDP = “Markov Decision.
Markov Decision Processes AIMA: 17.1, 17.2 (excluding ), 17.3.
Tabu Search Applications Outlines: 1.Application of Tabu Search 2.Our Project with Tabu Search: EACIIT analytics.
1 Markov Decision Processes Finite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Since all points on the x-axis have a y-coordinate of 0, to find x-intercept, let y = 0 and solve for x Since all points on the y-axis have an x-coordinate.
Making complex decisions
Haim Kaplan and Uri Zwick
B. Jayalakshmi and Alok Singh 2015
Markov Decision Processes
Planning to Maximize Reward: Markov Decision Processes
Markov Decision Processes
Hidden Markov Models Part 2: Algorithms
Reinforcement Learning in MDPs by Lease-Square Policy Iteration
Evaluation of (Deterministic) BT Search Algorithms
13. Acting under Uncertainty Wolfram Burgard and Bernhard Nebel
Evaluation of (Deterministic) BT Search Algorithms
Evaluation of (Deterministic) BT Search Algorithms
Markov Decision Processes
Markov Decision Processes
Presentation transcript:

1 J.Y Greff, L. Idoumghar and R. Schott TDF && IECN / LORIA - INRIA July 2002 Using Markov Decision Processes to the Frequency Assignment Problem

2 Summary Radio network planning Markov Decision Processes Experimental results Conclusion

3 Radio network planning: Network dimensionning

4 Radio network planning: Interference phenomena

5 3 Radio network planning: Frequency Assignment Constraints graph

6 = 0 = 1 = Radio network planning: Frequency Assignment

7 One tool of AI that can be used to get Optimal solution under a stochastic domain. A formal description of MDP is a Tuple (S, A, T, R) : * S: is a finite set of states of the world, * A: is a finite set of actions, * T: S x A  P Transition function of states (T(s,a,s’) represents the probability p s,s’ of reaching state s’ starting in state s’ and given the action a) * R: S x A   reward function. R(s,a) represents the reward of executing action a. Markov Decision Processes

8

9 Our Approach * S: is given by a set of transmitters, * A: each action correspond to the search of a best frequency to assign to a transmitter s. If there does not exist a best frequency, the action consists in keeping the frequency assigned to this transmitter, Markov Decision Processes

10 Let M= (m i,j ) n*n be a symetric matrix, where m i,j with i  j represents the manimum frequency separation required to satisfy the constraint between vertices vi and vj Let P = (p i,j ) n*n be a transition matrix, where p i,j represents a probability of reaching a state j starting in state i, For i:= 1 to n Do sum := 0 For j:= 1 to n Do sum := sum + m i,j End For j:= 1 to n Do p i,j := m i,j /sum End * Transition matrix is calculated by using the algorithm given as following: Markov Decision Processes

11 : represents the minimal gap in frequency required to satisfy the constraint between vertices v s and v s’ weight associate to the constraint m s,s’  s,s’ :  s,s’  R(s, a )  s,s’  * If  s,s’ m  (s’) a  0 s,s’  If  s,s’ m  (s’) a  a  s,s’ m  m Markov Decision Processes * The Reward function is defined as follows:

12

13 Markov Decision Processes: Probabilistic tabu search stop := false tabu_list :=  let A be a best solution While not stop do with a probability p t, chose c in the best neighbouring configuration forbidden by tabu list let d be the best neighbouring configuration which is not forbidden by the tabu list. This configuration is choosen by using a uniform probability choose the best configuration B between c and d wich minimizes an objective function A := B adjust tabu_list stop := stopping criterion done L. Idoumghar and R. Schott and M. Alabau, New Hybrid Genetic Algorithm for the Frequency Assignment Problem, Journal of IEEE Transactions on Broadcasting, Mars 2002, Vol. 48, N. 1,

14 Experimental results Comparison of the results obtained by differents approachs that are used to solve instance1 (more than 970 transmitters and more than constraints) of the frequency assignment problem.

15 Comparison of the results obtained by differents approachs that are used to solve instance2 (more than 970 transmitters and more than constraints) of the frequency assignment problem. Experimental results

16 Comparison of the results obtained by differents approachs that are used to solve instance3 (more than 970 transmitters and more than constraints) of the frequency assignment problem. Experimental results

17 Experimental results

18 Experimental results: Quality plan estimationin regard of radio system characteristics coverage area service area percentage = Ca : Coverage Area Ca Sa E Sa : Service Area E : Transmitter

% of transmitters have a service area at most equal to 65 % of the cover area Or 75 % of transmitters have a service area at less equal to 65 % of the cover area % of transmitters have a service area at less equal to 70 % of the cover area % accumulated of transmitters Ratio service area / coverage area Experimental results: Quality plan estimationin regard of radio system characteristics

20 Instance 1 Experimental results: Quality plan estimationin regard of radio system characteristics

21 Instance 2 Experimental results: Quality plan estimationin regard of radio system characteristics

22 Instance 3 Experimental results: Quality plan estimationin regard of radio system characteristics

23 Conclusion Using Markov decision processes to solve the frequency assignment problem, Improvement of the quality of the solutions and the time complexity

24 Future works Parallelization of our algorithms Markov Decision Processes: Decomposition

25 End

26

27 Instance 3

28 Instance 2