Agenda Setting and Improvement Monitoring in a University Department

21
Agenda Setting and Improvement Monitoring in a University Department Igor Dubina School of Economics and Management Altai State University Barnaul, Russia Email: [email protected] Stuart Umpleby Research Program in Social and Organizational Learning The George Washington University Washington, DC USA Email: [email protected] Deming Conference New York City, February 2006

description

Agenda Setting and Improvement Monitoring in a University Department. Stuart Umpleby Research Program in Social and Organizational Learning The George Washington University Washington, DC USA Email: [email protected]. Igor Dubina School of Economics and Management Altai State University - PowerPoint PPT Presentation

Transcript of Agenda Setting and Improvement Monitoring in a University Department

Page 1: Agenda Setting  and Improvement Monitoring in a University Department

Agenda Setting and Improvement Monitoringin a University Department

Igor Dubina School of Economics and

ManagementAltai State University

Barnaul, RussiaEmail: [email protected]

Stuart UmplebyResearch Program in Social and Organizational LearningThe George Washington UniversityWashington, DC USAEmail: [email protected]

Deming Conference

New York City, February 2006

Page 2: Agenda Setting  and Improvement Monitoring in a University Department

The research was conducted in the Department of Management Science

George Washington University

Igor Dubina was a visiting scholar at GWU in 2004-2005 under the Junior Faculty Development Program (JFDP) of the

U.S. Department of State

Page 3: Agenda Setting  and Improvement Monitoring in a University Department

Quality Improvement Priority Matrix (QIPM)

• 1995, 1996 Baldrige Award Conferences • A method for achieving data-driven decision-making • QIPM is a way of focusing management attention on

high priority tasks. It can be seen as an alternative to control charts

• Features of an organization (or product or service) are rated on two scales – importance and performance

• Scales range from 1 to 9 • The measures that result are averaged Importance (I),

Performance (P), and Importance/ Performance Ratio (IPR)

Page 4: Agenda Setting  and Improvement Monitoring in a University Department

QIPM

1

5

9

1 5 9

Importance

Per

form

ance

1

52 50

45

4

32

13

6

49

2

16

41

25

21

46

47

35

2234

43

48

51

53

12

11

36

19 17

42

27

10

Page 5: Agenda Setting  and Improvement Monitoring in a University Department

Data was collected from members of the GWU Department of Management Science in 2001, 2002, 2003, and 2005

They evaluated features of the Department (a total of 52 features):

• Funds to support research• Salaries• Coordination with other depts.• Computer labs• Classroom facilities• Classroom scheduling• Office space for faculty• Travel support• Dept. and School websites• Library book and journal collection• Office security• English skills of students• Course evaluations• Teaching assistants• Faculty annual reports• Conference room and other space• Computer hardware and software• Course catalogue• Copiers• Secretarial support• Dept. strategic plan

Page 6: Agenda Setting  and Improvement Monitoring in a University Department

The most stable high importance features (always in the first 15) from 2001 to 2005

Feature Ave. Imp.

1. Health care benefits 8.72

2. Computer software 8.65

3. Classroom facilities 8.65

4. A supportive climate in

the dept. 8.60

5. Salaries 8.58

6. Projection equipment 8.48

7. Computer labs 8.47

Page 7: Agenda Setting  and Improvement Monitoring in a University Department

The most stable low importance features (always in the last 15) from 2001 to 2005

Feature Ave. Imp.

1. Recreational activities 4.19

2. Social activities 4.94

3. Faculty annual reports 5.31

4. SBPM working papers

series 5.92

5. Faculty websites 5.94

6. Annual retreat 6.11

Page 8: Agenda Setting  and Improvement Monitoring in a University Department

The most stable low Performance features (always in the last 15) from 2001 to 2005

Feature

Ave.

Perf.

Help with writing research proposals 3.34

Dept. organization to implement its strategic plan 3.54

Use of continuous improvement methods in the Dept. 3.74

Conference room and other space 3.81

Dept. strategic plan 3.89

Building/ physical environment 3.94

Recreational activities 4.06

Page 9: Agenda Setting  and Improvement Monitoring in a University Department

The most stable high Performance features (always in the first 15) from 2001 to 2005

Feature Ave. Perf.

Dept. head protects faculty from admin.

interference 7.76

Computer hardware 7.00

A supportive climate in the dept. 6.93

Interlibrary loan 6.85

Computer software 6.84

Copiers 6.72

Fax machines 6.62

Course catalogue 6.39

Campus grounds 6.17

Page 10: Agenda Setting  and Improvement Monitoring in a University Department

The features always in the SE quadrant from 2001 to 2005

Feature Ave. IPR

1. Dept. organization to implement its

strategic plan 2.06

2. Help with writing research proposals 1.96

3. Dept. strategic plan 1.95

4. Building/ physical environment 1.95

5. Conference room and other space 1.93

6. Classroom facilities 1.89

7. Salaries 1.88

8. Promotion of contract faculty 1.87

9. Parking for students 1.75

10. Funds to support research 1.74

11. Computer labs 1.72

12. Use of continuous improvement methods in

the Dept. 1.69

13. Coordination with other depts. 1.65

14. SBPM working papers series 1.62

Page 11: Agenda Setting  and Improvement Monitoring in a University Department

– From 1/3 to 1/2 of all features routinely fall into the SE quadrant

(e.g., 19 of 51 features in 2001, 17 of 52 in 2002, 23 of 52 in 2003, and 26 of 52 in 2005

– The “border effect”– The problem of automatic

clustering of factors by their priorities

1

5

9

1 5 9

ImportanceP

erfo

rman

ce

1

52 50

45

4

32

13

6

49

2

16

41

25

21

46

47

35

2234

43

48

51

53

12

11

36

19 17

42

27

10

A classical approach: features in the SE quadrant are considered to have a high priority

Visual analysis of QIPM does not discriminate features’ priorities sufficiently

Page 12: Agenda Setting  and Improvement Monitoring in a University Department

Using average Importance and Performance as a midpoint rather than the scale midpoint

1

5

9

1 5 9

Importance

Per

form

ance

1

52 50

45

4

32

13

6

49

2

16

41

25

21

46

47

35

2234

43

48

51

53

12

11

36

19 17

42

27

10

1

5

9

3 7

Importance

Per

form

ance

1

52 50

45

4

32

13

6

49

2

16

41

25

21

46

47

35

2234

43

48

51

53

12

11

36

19 17

42

27

10

Page 13: Agenda Setting  and Improvement Monitoring in a University Department

Clustering features by the IPR interval

1

5

9

1 5 9

importance

per

form

ance

2005

1

5

9

1 5 9

importancep

erfo

rman

ce

2003

Cluster 0 (urgent) – IPR>2 Cluster 1 (high priority) – [1.5 – 2] Cluster 2 (medium priority) – [1.25 – 1.5)Cluster 3 (low priority) – IPR<1.25

rIP = 0.96 (0), 0.88 (1), 0.85 (2), 0.90 (3)rIP = 0.18 (unclustered)

A way to automatically cluster features with different priorities is to choose intervals that create clusters with the highest correlation coefficient

Page 14: Agenda Setting  and Improvement Monitoring in a University Department

An approach to automatically cluster features with different priorities

P=a0+a1I+b1C1+b2C2+b3C3 , r2

P – Performance I – Importance C1, C2, and C3 – dummy variables corresponding to clusters (These

variables have values 1 or 0 depending on whether a point is or is not in the corresponding cluster: 1, 2, or 3)

Coefficients b1, b2, b3 represent the increased performance for each cluster compared with the cluster 0

r2 – coefficient of determination

The higher r2 is in this regression equation, the more precise the clustering

Page 15: Agenda Setting  and Improvement Monitoring in a University Department

Clusters with different priorities(2005 г., r2 = 0.90 )

1

5

9

1 5 9

importance

per

form

ance

2005

Page 16: Agenda Setting  and Improvement Monitoring in a University Department

A simplified approach

P=a0+a1I+a2X

P – Performance I – Importance X – a dummy variable corresponding to the number of the

cluster. It may have values 0, 1, 2, or 3 if a point falls into the corresponding cluster.

The coefficient a2 represents the average shift in performance between clusters.

2005: r2 = 0.89

Page 17: Agenda Setting  and Improvement Monitoring in a University Department

An integrated approach

The parameters for automatically clustering features with different priorities

– Number of clusters– IPR intervals– Number of features in clusters– Correlation of features in clusters– The coefficient of determination – Average shift in performance between clusters

Page 18: Agenda Setting  and Improvement Monitoring in a University Department

Analysis of year-to-year dynamics

dIPR = IPRt2 – IPRt1 represents the direction of movement (becoming more urgent or less urgent)

represents the amount of movement

3 clusters with different levels and directions of change:

• DI >= DIt и dIPR>0 (regress and greater urgency)• DI >= DIt и dIPR<0 (progress and less urgency)• DI < DIt (change is not significant)

212

212 tttt PPIIDI

Page 19: Agenda Setting  and Improvement Monitoring in a University Department

Multiyear analysis of feature dynamics

dIPR = abs(IPRt1 – IPRt2) + abs(IPRt2 – IPRt3) + abs(IPRt3 – IPRt4)

reflects changes between clusters with different priorities

represents the amount of movement

The more important movement was between clusters, even if the distance moved was not as great. IPR significantly changes when a feature moves in a perpendicular direction (from cluster to cluster). Movement between clusters means a change in priority. Therefore, the indicator dIPR is more important for analyzing changes in priorities.

243

243

232

232

221

221 tttttttttttt PPIIPPIIPPIIDI

Page 20: Agenda Setting  and Improvement Monitoring in a University Department

QIPM

• Is easy to understand• Is efficient in terms of time and resources• Provides enough precision for monitoring

changes in priorities and performance• Is based on subjective data, so can be used to

extend process improvement methods beyond manufacturing into service-oriented activities

Page 21: Agenda Setting  and Improvement Monitoring in a University Department

• Presented by Stuart Umpleby at the

• Deming Conference

• Fordham University

• New York City

• February 13-14, 2006