markov chain analysis manpower data

42
SJEM2252 : STOCHASTIC PROCESSES :

Upload: wmhan-farhan

Post on 25-May-2015

796 views

Category:

Education


0 download

DESCRIPTION

Markov Chain Analysis

TRANSCRIPT

Page 1: MARKOV CHAIN ANALYSIS MANPOWER DATA

SJEM2252 : STOCHASTIC PROCESSES

:

Page 2: MARKOV CHAIN ANALYSIS MANPOWER DATA
Page 3: MARKOV CHAIN ANALYSIS MANPOWER DATA

What is Markov Chain?

A stochastic model that describes the probabilities of

transition among the states of a system

Change of state depends probabilistically only on

the current state of the system

Independent of the past given that the present state

is known.

The behaviour depends on the structure of the

transition matrix P

Page 4: MARKOV CHAIN ANALYSIS MANPOWER DATA

Markov chain modelling is one of the most powerful

tools for analysing complex stochastic systems (Kim

and Smith, 1989)

Markov chain models have become popular in

manpower system planning. Several researchers

have adopted Markov chain models to clarify

manpower policy issues.

Page 5: MARKOV CHAIN ANALYSIS MANPOWER DATA
Page 6: MARKOV CHAIN ANALYSIS MANPOWER DATA

Method of Data Collection

• Based on the journal, the staff movement, in

steps, in this system is considered to be an

absorbing chain. A forty year data of staff

movements, as specified by state space

described above, was obtained from the

organization studied.

• Study Area: University of Benin, Nigeria

Page 7: MARKOV CHAIN ANALYSIS MANPOWER DATA

Table 1 : Original state space

Table 2 : Lumped state space

States 1970-2010 States 1970-2010

Contract 282 Special leave 5

Dismissal 2 Study leave 2

Late 109 Suspension 2

Leave of absence 65 Temporary 617

Left 108 Termination

appointment

3

Permanent 3051 Training leave 63

Resign 12 Visiting appointment 12

Retired 446 Withdrawal 7

Sabatical 49

States Manpower

Retirement 458

Wastage 227

Disciplinary case 463

Leave 191

Staff Stock 3109

Recruitment 911

Total 5359

Page 8: MARKOV CHAIN ANALYSIS MANPOWER DATA

State space was reduced to six from seventeen as follows:

(i) Recruitment (Rc)

(ii) Staff Stock (Ss)

(iii) Training leave (T)

(iv) Suspension (S)

(v) Wastage (W)

(vi) Retirement (R).

We developed the theory of canonical forms of transition probability

matrix of a finite aperiodic recurrent set decomposable Markov chain

of order n with steady state probability row vector.

By states, it is specifically implied the condition, the status, the

position or situation a Markov Chain (object or staff) undergoes in the

transition process towards the final position of retirement, if the staff

can ever get there.

Page 9: MARKOV CHAIN ANALYSIS MANPOWER DATA

• Retirement, (R) : is the point where a person stops

employment completely.

• Wastage(W): reduction in size of a workforce by

voluntary resignation.

• Suspension (S): To bar for a period from a privilege,

office, or position, usually as a punishment.

• Training Leave (T): company provide opportunity to

training or outside study.

• Staff stock(Ss) : official employee.

• Recruitment(Rc) : temporary worker.

Page 10: MARKOV CHAIN ANALYSIS MANPOWER DATA

Markov Transition Matrix:

A square matrix describing the probabilities of moving from one state to

another in a dynamic system.

In each row are the probabilities of moving from the state represented

by that row, to the other states. Thus the rows of a Markov transition

matrix each add to one.

Probabilities were organized in canonical form.

The data obtained appear pure/original although we cannot

guarantee for perfection in the way the records were kept.

The data obtained were used to compute the transition probability

matrix (TPM) which we depict in Table 1. The estimates of the transition

probabilities were based on frequency distributions or tabulations of the

number of transitions from one state to the other in the system

considered. The frequency were converted to TPM by dividing each row

by its total.

Page 11: MARKOV CHAIN ANALYSIS MANPOWER DATA

It is assumed that after this number of movements, the matrix T would have attained stationarity.

Transition matrix, P is

P =

18607.063501.003901.0004636.009355.0

17751.060581.003722.009022.0008924.0

18588.063436.003897.009447.004632.00

000100

000010

000001

Rc

Ss

T

S

W

R

R W S T Ss Rc

Page 12: MARKOV CHAIN ANALYSIS MANPOWER DATA
Page 13: MARKOV CHAIN ANALYSIS MANPOWER DATA

DECOMPOSITION OF A STATE SPACE INTO EQUIVALENCE CLASSES

• We decompose state space into equivalence classes

• States in the same class will have the same characterizing attributes.

• For state i, we let:

C(i) denote the class containing state i.

T(i) be the set of all states that are accessible from i.

F(i) be the set of states from which state I is accessible.

Page 14: MARKOV CHAIN ANALYSIS MANPOWER DATA

Algorithm:

• Let T(i)={i} and F(i)={Φ}

• For each state k in T(i), add to T(i) all states j such that pkj>0 (if k is not already there). Reiterate.

• Add state j to F(i) if state i is in T(j). Reiterate.

• C(i)= F(i) ∩ T(i).

Page 15: MARKOV CHAIN ANALYSIS MANPOWER DATA

Transient and Recurrent Classes

• State i is recurrent if i is accessible from every state that is accessible from i.

• If i communicate with j, but j does not communicate with I, then I is transient.

Canonical Form

• By grouping the states together in their classes and list the recurrent classes first, the transition matrix P can be written the following canonical form:

R 0Q T

Page 16: MARKOV CHAIN ANALYSIS MANPOWER DATA

Absorbing State:

pii=1

Stays forever

Recurrent

Forms a class by itself

Absorbing Chain:

All of its states are either transient or absorbing.

It implies that it is reducible chain

The transition matrix P can be organized into the following canonical form:

𝐈 𝐎𝐐 𝐓

Page 17: MARKOV CHAIN ANALYSIS MANPOWER DATA

PERIODIC STATES AND PERIODIC CLASS

• A state i has period d if the chain can only revisit it a multiple of d steps later.

• d is the greatest common divisor of the set of integers

• If d=1, then the state has period 1 and is said to be aperiodic. Otherwise, if d>1,the state has period more than 1 and is said to be periodic.

Page 18: MARKOV CHAIN ANALYSIS MANPOWER DATA

EXPECTED NUMBER OF VISITS TO A TRANSIENT STATE

• Let vij be the total number of visits the chain makes to state j throughout its life, starting from state i.

• Matrix U is called the fundamental matrix.

Page 19: MARKOV CHAIN ANALYSIS MANPOWER DATA

EXPECTED ABSORPTION TIME

• Let ui be the absorption time or the number of steps before the chain is absorbed into one of the absorbing states, starting from a transient state i.

• The expected absorption :

• In matrix form:

where e is a column vector with all elements equal 1.

Page 20: MARKOV CHAIN ANALYSIS MANPOWER DATA

ABSORBING PROBABILITIES

• Q - transient to absorbing portion

• T - transient to transient portion

• [UQ]i,j - probability that the chain starts from transient state i, and then moves to the absorbing state j rather than any other absorbing state regardless of the number of steps taken.

Page 21: MARKOV CHAIN ANALYSIS MANPOWER DATA

LIMITING DISTRIBUTIONS

• Dependent on initial conditions

• After infinite steps, probability of visiting transient state equals to 0

Page 22: MARKOV CHAIN ANALYSIS MANPOWER DATA
Page 23: MARKOV CHAIN ANALYSIS MANPOWER DATA

• State space, S = { Retirement,(R) , Wastage,(W) , Suspension(S) ,Training Leave,(T), Staff Stock,(Ss) , Recruitment,(Rc) }

• Transition matrix, P is

P =

18607.063501.003901.0004636.009355.0

17751.060581.003722.009022.0008924.0

18588.063436.003897.009447.004632.00

000100

000010

000001

Rc

Ss

T

S

W

R

R W S T Ss Rc

Page 24: MARKOV CHAIN ANALYSIS MANPOWER DATA

DECOMPOSITION OF STATE SPACE

State, i T ( i ) F ( i ) C ( i )

R R R, T, Ss, Rc R

W W W, T, Ss, Rc W

S S S T, Ss, Rc S

T T, W, S, Ss, Rc, R T, Ss, Rc T, Ss, Rc

Ss Ss, R, S, T, Rc, W T, Ss, Rc T, Ss, Rc

Rc Rc, R, W, T, Ss, S T, Ss, Rc T, Ss, Rc

Page 25: MARKOV CHAIN ANALYSIS MANPOWER DATA

• Closed class

I. E1 = { R }

II. E2 = { W }

III. E3 = { S }

• Non-closed class

T = { T, Ss, Rc }

R,W ,S are recurrent states

T, Ss, Rc are transient states

Page 26: MARKOV CHAIN ANALYSIS MANPOWER DATA

Absorbing state

{ R, W, S }

Transient state

{ T, Ss, Rc }

Absorbing chain

Page 27: MARKOV CHAIN ANALYSIS MANPOWER DATA

P =

P2 =

18607.063501.003901.0004636.009355.0

17751.060581.003722.009022.0008924.0

18588.063436.003897.009447.004632.00

000100

000010

000001

Rc

Ss

T

S

W

R

15459.015459.052760.006098.005679.016763.0

14749.050334.003092.014839.000995.015991.0

15444.052706.003238.015538.005674.007400.0

000100

000010

000001

R W S T Ss Rc

Page 28: MARKOV CHAIN ANALYSIS MANPOWER DATA

P3 =

P4 =

12844.043835.002693.011164.006546.022917.0

12254.041820.002569.019672.001822.021862.0

12831.043791.002690.020599.006540.013548.0

000100

000010

000001

10672.036421.002238.015373.007266.028031.0

10181.034746.002135.023688.002509.026741.0

10661.036383.002235.024804.007260.018656.0

000100

000010

000001

Page 29: MARKOV CHAIN ANALYSIS MANPOWER DATA

P5 =

P6 =

08867.030260.001859.018870.007865.032279.0

08459.028869.001774.027025.003080.030794.0

08858.030229.001857.028298.007857.022901.0

000100

000010

000001

1 0 0 0 0 0

0 1 0 0 0 0

0 0 1 0 0 0

0.26427 0.08354 0.31201 0.01543 0.25116 0.07359

0.34161 0.03555 0.29797 0.01474 0.23986 0.07028

0.35809 0.08362 0.21776 0.01545 0.25142 0.07367

Page 30: MARKOV CHAIN ANALYSIS MANPOWER DATA

1 0 0 0 0 0

0 1 0 0 0 0

0 0 1 0 0 0

0.29357 0.08767 0.33612 0.01282 0.20868 0.06114

0.36959 0.03949 0.32100 0.01224 0.19928 0.05839

0.38742 0.08775 0.24190 0.01283 0.20889 0.06121

P7 =

P8 =

1 0 0 0 0 0

0 1 0 0 0 0

0 0 1 0 0 0

0.31791 0.09110 0.35616 0.01065 0.17338 0.05080

0.39284 0.04276 0.34013 0.01017 0.16558 0.04852

0.41178 0.09118 0.26196 0.01066 0.17356 0.05085

Page 31: MARKOV CHAIN ANALYSIS MANPOWER DATA

P9 =

P10 =

1 0 0 0 0 0

0 1 0 0 0 0

0 0 1 0 0 0

0.33813 0.09394 0.37281 0.00885 0.14405 0.04221

0.41216 0.04548 0.35603 0.00845 0.13757 0.04031

0.43203 0.09403 0.27863 0.00886 0.14420 0.04225

1 0 0 0 0 0

0 1 0 0 0 0

0 0 1 0 0 0

0.35494 0.09631 0.38664 0.00735 0.11969 0.03507

0.42820 0.04774 0.36924 0.00702 0.11430 0.03349

0.44885 0.09640 0.29247 0.00736 0.11981 0.03511

Page 32: MARKOV CHAIN ANALYSIS MANPOWER DATA

For i = 1,2,3,4,5,and 6

Pii(n) > 0 ,

when n = 1,2,…,10,…

Greatest common divisor d is 1. Hence d=1.

That is, all the states are aperiodic state.

Hence, the chain is aperiodic.

Page 33: MARKOV CHAIN ANALYSIS MANPOWER DATA

EXPECTED NUMBER OF VISIT TO TRANSIENT STATES

U = ( I – T )-1

= 1 0 00 1 00 0 1

−0.03897 0.63436 0.185880.03722 0.60581 0.177510.03901 0.63501 0.18607

-1

𝑇 𝑆𝑠 𝑅𝑐

= 𝑇𝑆𝑠𝑅𝑐

1.23040 3.75028 1.098890.22004 4.58150 1.049430.23064 3.75413 2.10002

Page 34: MARKOV CHAIN ANALYSIS MANPOWER DATA

If we are currently on training leave,T : We are expected to go for the next training leave 1.23040 steps

later.

We are expected to be official employee 3.75028 steps later

We are expected to be recruited as a temporary worker 1.09889

steps later

If we are now an official employee , Ss : We are expected to go for next training leave 0.22004 steps later.

We are expected to stay as an official employee 4.58150 steps

later

We are expected to be recruited to be as a temporary worker

1.04943 steps later

If we are now recruited, Rc : We are expected to go for next training leave 0.23064 steps later.

We are expected to be an official employee 3.75413 steps later

We are expected to be recruited again 2.10002 steps later

Page 35: MARKOV CHAIN ANALYSIS MANPOWER DATA

EXPECTED ABSORPTION TIME

ui = Ue

= 1.23040 3.75028 1.098890.22004 4.58150 1.049430.23064 3.75413 2.10002

111

= 𝑇𝑆𝑠𝑅𝑐

6.0795775.8509626.084779

Page 36: MARKOV CHAIN ANALYSIS MANPOWER DATA

If we are now on training leave (T), we will move

6.079577 number of steps before getting

absorbed into retirement (R),wastage (W) or

suspension (S)

If we are now an official employee (Ss), we will

move 5.850962 number of steps before getting

absorbed into retirement (R),wastage (W) or

suspension (S)

If we are now recruited (Rc), we will move

6.084779 number of steps before getting

absorbed into retirement (R),wastage (W) or

suspension (S)

Page 37: MARKOV CHAIN ANALYSIS MANPOWER DATA

ABSORBING PROBABILITIES

A= UQ

= 1.23040 3.75028 1.098890.22004 4.58150 1.049430.23064 3.75413 2.10002

0 0.04632 0.094470.08924 0 0.090220.09355 0.04636 0

𝑅 𝑊 𝑆

= 𝑇𝑆𝑠𝑅𝑐

0.43748 0.10794 0.454590.50703 0.05884 0.434130.53148 0.10804 0.36049

Page 38: MARKOV CHAIN ANALYSIS MANPOWER DATA

If we are currently on training leave,T : We will end up in retirement with probability of 0.43748

We will end up in wastage with probability of 0.10794

We will end up being suspended with probability of 0.45459

If we are now an official employee , Ss : We will end up in retirement with probability of 0.50703

We will end up in wastage with probability of 0.05884

We will end up being suspended with probability of 0.43413

If we are now recruited, Rc : We will end up in retirement with probability of 0.53148

We will end up in wastage with probability of 0.10804

We will end up being suspended with probability of 0.36049

Page 39: MARKOV CHAIN ANALYSIS MANPOWER DATA

LIMITING

DISTRIBUTIONS

𝑃(𝑘) = 𝐼 0𝐴 𝑇𝑘

𝑃(∞)= 𝐼 0𝐴 0

Page 40: MARKOV CHAIN ANALYSIS MANPOWER DATA

The limiting probabilities, 𝑃(∞)

𝑃(∞) =

After infinite years, employees must go to Retirement(R ) ,Wastage(W), and Suspension(S)

00036049.010804.053148.0

00043413.005884.050703.0

00045459.010794.043748.0

000100

000010

000001

Rc

Ss

T

S

W

R

R W S T Ss Rc

Page 41: MARKOV CHAIN ANALYSIS MANPOWER DATA

Arising from the foregoing analysis and discussion, it is evident that the Markov Chain model applied has been able to provide useful insight into the behaviour pattern for a long run manpower policy for the organization studied.

The transition probability matrix, after several sets of transitions, shows the properties of stochastic regularities and aperiodic.The fundamental matrix, U when multiplied with the Q, do, gave us a dependable probability matrix that provides reliable manpower forecast outcomes in line with the theory of Markov Chain.

The results of this study appear useful for understanding the long run manpower policy for a university system.

Page 42: MARKOV CHAIN ANALYSIS MANPOWER DATA

• A.C. Igboanugo & M.K. Onifade(2011). Markov Chain Analysis of Manpower Data of a Nigerian University. Journal of Innovative Research in Engineering and Science 2, 2(2), 107-123.

• Adriana Irawati Nur Ibrahim. (2013). SJEM2252: Stochastic Processes, week 5-11 notes [pdf].

• Retrieved from http://www.thefreedictionary.com