Getting rid of stochasticity (applicable sometimes) Han Hoogeveen Universiteit Utrecht Joint work with Marjan van den Akker.

Slides:



Advertisements
Similar presentations
Algorithm Design Methods Spring 2007 CSE, POSTECH.
Advertisements

Some additional Topics. Distributions of functions of Random Variables Gamma distribution,  2 distribution, Exponential distribution.
Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
ISE480 Sequencing and Scheduling Izmir University of Economics ISE Fall Semestre.
Crew: Pavel Babenkov Kirill Kalinkin Customized Stochastic Lot Scheduling Problem with Due Date Management.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Experimental Design, Response Surface Analysis, and Optimization
Online Scheduling with Known Arrival Times Nicholas G Hall (Ohio State University) Marc E Posner (Ohio State University) Chris N Potts (University of Southampton)
Spring, Scheduling Operations. Spring, Scheduling Problems in Operations Job Shop Scheduling. Personnel Scheduling Facilities Scheduling.
In this handout Stochastic Dynamic Programming
All Hands Meeting, 2006 Title: Grid Workflow Scheduling in WOSE (Workflow Optimisation Services for e- Science Applications) Authors: Yash Patel, Andrew.
Planning under Uncertainty
Complexity 16-1 Complexity Andrei Bulatov Non-Approximability.
Clustering short time series gene expression data Jason Ernst, Gerard J. Nau and Ziv Bar-Joseph BIOINFORMATICS, vol
1 IOE/MFG 543 Chapter 11: Stochastic single machine models with release dates.
Job Release-Time Design in Stochastic Manufacturing Systems Using Perturbation Analysis By: Dongping Song Supervisors: Dr. C.Hicks & Dr. C.F.Earl Department.
Accept or Reject: Can we get the work done in time? Marjan van den Akker Joint work with Han Hoogeveen.
CSE 421 Algorithms Richard Anderson Lecture 6 Greedy Algorithms.
Computability and Complexity 24-1 Computability and Complexity Andrei Bulatov Approximation.
Planning operation start times for the manufacture of capital products with uncertain processing times and resource constraints D.P. Song, Dr. C.Hicks.
1 IOE/MFG 543 Chapter 3: Single machine models (Sections )
1 IOE/MFG 543 Chapter 10: Single machine stochastic models Sections 10.1 and 10.4 You may skip Sections
Maximum likelihood (ML)
MATHEMATICS 3 Operational Analysis Štefan Berežný Applied informatics Košice
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 8 Tests of Hypotheses Based on a Single Sample.
Scheduling of Flexible Resources in Professional Service Firms Arun Singh CS 537- Dr. G.S. Young Dept. of Computer Science Cal Poly Pomona.
ECES 741: Stochastic Decision & Control Processes – Chapter 1: The DP Algorithm 1 Chapter 1: The DP Algorithm To do:  sequential decision-making  state.
© 2009 IBM Corporation 1 Improving Consolidation of Virtual Machines with Risk-aware Bandwidth Oversubscription in Compute Clouds Amir Epstein Joint work.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Use of moment generating functions 1.Using the moment generating functions of X, Y, Z, …determine the moment generating function of W = h(X, Y, Z, …).
Stochastic Models Lecture 2 Poisson Processes
Week 10Complexity of Algorithms1 Hard Computational Problems Some computational problems are hard Despite a numerous attempts we do not know any efficient.
CP Summer School Modelling for Constraint Programming Barbara Smith 2. Implied Constraints, Optimization, Dominance Rules.
CSC 2535 Lecture 8 Products of Experts Geoffrey Hinton.
1 Short Term Scheduling. 2  Planning horizon is short  Multiple unique jobs (tasks) with varying processing times and due dates  Multiple unique jobs.
OR 2004 Tilburg, 1/9/20041 Simulation of a job shop scheduling strategy: Do advanced scheduling algorithms pay off when processing times are disturbed?
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Expectation. Let X denote a discrete random variable with probability function p(x) (probability density function f(x) if X is continuous) then the expected.
Linear Program Set Cover. Given a universe U of n elements, a collection of subsets of U, S = {S 1,…, S k }, and a cost function c: S → Q +. Find a minimum.
Sampling and estimation Petter Mostad
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Earliness and Tardiness Penalties Chapter 5 Elements of Sequencing and Scheduling by Kenneth R. Baker Byung-Hyun Ha R1.
Dynamic Programming.  Decomposes a problem into a series of sub- problems  Builds up correct solutions to larger and larger sub- problems  Examples.
Algorithm Design Methods 황승원 Fall 2011 CSE, POSTECH.
Stochastic Optimization
March 7, Using Pattern Recognition Techniques to Derive a Formal Analysis of Why Heuristic Functions Work B. John Oommen A Joint Work with Luis.
Introduction A probability distribution is obtained when probability values are assigned to all possible numerical values of a random variable. It may.
1 Introduction Optimization: Produce best quality of life with the available resources Engineering design optimization: Find the best system that satisfies.
The NP class. NP-completeness Lecture2. The NP-class The NP class is a class that contains all the problems that can be decided by a Non-Deterministic.
Single Machine Scheduling Problem Lesson 5. Maximum Lateness and Related Criteria Problem 1|r j |L max is NP-hard.
Chapter 6 – Continuous Probability Distribution Introduction A probability distribution is obtained when probability values are assigned to all possible.
OPERATING SYSTEMS CS 3502 Fall 2017
Hard Problems Some problems are hard to solve.
Expectations of Random Variables, Functions of Random Variables
Water Resources Planning and Management Daene McKinney
Algorithm Design Methods
CHAPTER 8 Operations Scheduling
Basic Project Scheduling
CONCEPTS OF HYPOTHESIS TESTING
Chapter 9 Hypothesis Testing.
Hidden Markov Models Part 2: Algorithms
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Algorithm Design Methods
Algorithm Design Methods
Stochastic Simulations
Algorithm Design Methods
Presentation transcript:

Getting rid of stochasticity (applicable sometimes) Han Hoogeveen Universiteit Utrecht Joint work with Marjan van den Akker

Outline of the talk Problem description How to solve the deterministic problem Stochastic processing times –consequences –four classes of instances Stochastic machine Processing times and machine stochastic

Problem description 1 machine n jobs become available at time 0 Known processing time p j Known due date d j Known reward w j for timely completion (currently 1) Decision to make at time 0: accept or reject job j minimize number of tardy jobs on a single machine

Moore-Hodgson 1.Number the jobs in EDD order 2.Let S denote the EDD schedule 3.Find the first job not on time in S (suppose this is job j) 4.Remove from S the largest available job from jobs 1,…,j 5.Continue with Step 3 for this new schedule S until all jobs are on time

Solving the problem from scratch Observations First the on time jobs On time jobs in EDD order Forget about the late jobs (once rejected = lost for ever) Knowing the on time set is sufficient

Dominance rule Let E 1 and E 2 be two subsets of jobs 1,…,j All jobs in E 1 and E 2 are on time (feasible) Cardinality of E 1 and E 2 is equal The total processing time of the jobs in E 2 is more than the total processing time of the jobs in E 1 Then subset E 2 can be discarded.

Proof (sketch) Take an optimal schedule starting with E 2 (remainder: jobs from j+1, …, n) E2E2 remainder E1E1 time 0

Apply Dynamic Programming Define E j * (k): feasible subset of jobs 1,…,j with cardinality k and minimum total processing time Knowledge of the total processing time of E j * (k) is sufficient in the next step ⇒ use state variables f j (k) to represent this, etc.

DP (2) Add job j+1 Recurrence relation: If the number of on time jobs k can be reached without job j+1, then compute f j+1 (k)=min{f j (k), f j (k-1)+p j+1 } Effectively: the longest job of the old set plus job j+1 gets removed.

Further remarks DP computes more state variables than necessary DP can be used for the weighted case: –Use f j (W) with W is the total weight of the on time set (instead of cardinality of the on time set) DP can be used for more problems

Stochastic processing times Completion times are uncertain Decision about accept or reject must be made before running the schedule When do you consider a job on time?

On time stochastically Work with a sequence of on time jobs (instead of a set of completion times) Add a job to this sequence and compute the probability that it is ready on time If this probability is large enough (at least equal to the minimum success probability msp) then accept it as on time

Classes of processing times Gamma distribution Negative binomial distribution Equally disturbed processing times p j Normal distribution Jobs must be independent

Class 1: Gamma distribution Parameters a j and b (common) If X 1 and X 2 follow the gamma distribution and are independent, then X 1 +X 2 is gamma distributed with parameters a 1 +a 2 and b (we call this additive).

More gamma Define S as the set containing job j and all its predecessors in the schedule. Define p(S) as the sum of all processing times p j of jobs in S; a(S) is defined similarly. Then C j =p(S) follows a gamma distribution with parameters a(S) and b.

Even more gamma Denote the msp of job j by y j Job j is on time if P(C j  d j )  y j. The distribution of C j depends on a(S) only Given d j and y j, you can compute the maximum value of a(S) such that P(C j  d j )  y j : call this maximum value D j

Last of Gamma Important: a(S)  D j  P(C j  d j )  y j (S contains job j and its predecessors) Treat D j as ordinary due dates Treat a j as ordinary deterministic processing times You can use Moore-Hodgson!

Negative binomial distribution Parameters s_j and p (common for all jobs) If independent, then C_j=p(S) follows a negative binomial distribution with parameters s(S) and p Solvable like the gamma distribution

Same approach is viable if The distribution of p j depends on one specific parameter a j The probability distribution is additive P(C j  d j )  y j does not increase when a(S) increases, irrespective of y j The negative binomial distribution possesses these characteristics

More complicated problems p j Var dj dj msp j Job Job Normally distributed processing times Optimum: first job 2 and then job 1 Necessary for DP: msp values and due dates are oppositely ordered

Equal disturbances On time probability of job j depends on: –Number of predecessors (on time jobs before j) –Total processing time of its predecessors Dominance rule: given the cardinality of the on time set, take the one with minimum total processing time Use dynamic programming with state variables f j (k) that indicate the minimum total processing time possible (as before) Hence: Moore-Hodgson’s solves it!

Normal distribution (1) Parameters: expected processing time of job j and variance of job j Two parameters ⇒ simple approach fails Reminder: expected value and variances of X 1 +X 2 are equal to the respective sums Necessary for computing the on time probability of job j: –Total processing time of predecessors –Total variance of predecessors

Normal distribution (2) Dominance rule: if cardinality and total processing time are equal, then take the set with minimum total variance (msp > 0.5) Use state variables f j (k,P): –k is cardinality of on time set –P is total processing time of on time set –f j (k,P) is minimum variance possible

Normal distribution: details Running time pseudo-polynomial Problem is NP-hard Role of total variance and total processing time in the dominance rule and in the DP is interchangeable

Unreliable machine Assume deterministic processing times Amount of work done by the machine per period is stochastic Define X(t) as the stochastic variable denoting the amount of work done in [0,t] Assume that the probability distribution of X(t) is known

Solution method Use the minimum success probability y j Again, use S to denote job j and its predecessors in the schedule; the total processing time of S is p(S) Job j is stochastically on time if P(X(d j )  p(S))  y j

Solution method (2) Compute D j as the maximum value of p(S) such that P(X(d j )  p(S))  y j. Treat the D j values as the traditional due dates for a reliable machine. => Traditional deterministic problem Moore-Hodgson solves it

Everything unreliable Same approach as with the unreliable machine: –X(t) denotes the amount of work done in [0,t] –p(S) is total processing time of the jobs in S; this is a stochastic variable now –Job j is stochastically on time if P(X(d j )  p(S))  y j

Gamma distribution Now p(S) is gamma distributed with parameter a(S) Compute D j as the maximum value of a(S) such that P(X(d j )  p(S))  y j. Treat D j as ordinary due dates again Use a j as processing time of job j => Traditional deterministic problem Moore-Hodgson solves it

Conclusion We can sometimes get rid of stochasticity by –using minimum success probability –concept of stochastically on time Necessary to compute the probabilities –either analytically –or numerically (simulation)

Conclusion (2) Unreliable machines are made reliable again Stochastic processing times are represented by their deterministic parameter Then the deterministic algorithm can be applied again!