Using graphs and Laplacian eigenvalues to evaluate block ... · Laplacian eigenvalues of one of the...

Post on 21-Aug-2020

2 views 0 download

Transcript of Using graphs and Laplacian eigenvalues to evaluate block ... · Laplacian eigenvalues of one of the...

Using graphs and Laplacian eigenvaluesto evaluate block designs

R. A. BaileyUniversity of St Andrews / QMUL (emerita)

Ongoing joint work with Peter J. Cameron

Modern Trends in Algebraic Graph Theory,Villanova, June 2014

1/52

Abstract

Consider an experiment to compare v treatments in b blocks ofsize k.

Statisticians use various criteria to decide which design is best.

It turns out that most of these criteria are defined by theLaplacian eigenvalues of one of the two graphs defined by theblock design:I the Levi graph, which has v + b vertices,I or the concurrence graph, which has v vertices.

The algebraic approach shows that sometimes all the criteriaprefer highly symmetric designs but sometimes they favourvery different ones.

2/52

Abstract

Consider an experiment to compare v treatments in b blocks ofsize k.

Statisticians use various criteria to decide which design is best.

It turns out that most of these criteria are defined by theLaplacian eigenvalues of one of the two graphs defined by theblock design:I the Levi graph, which has v + b vertices,I or the concurrence graph, which has v vertices.

The algebraic approach shows that sometimes all the criteriaprefer highly symmetric designs but sometimes they favourvery different ones.

2/52

Abstract

Consider an experiment to compare v treatments in b blocks ofsize k.

Statisticians use various criteria to decide which design is best.

It turns out that most of these criteria are defined by theLaplacian eigenvalues of one of the two graphs defined by theblock design:I the Levi graph, which has v + b vertices,I or the concurrence graph, which has v vertices.

The algebraic approach shows that sometimes all the criteriaprefer highly symmetric designs but sometimes they favourvery different ones.

2/52

Abstract

Consider an experiment to compare v treatments in b blocks ofsize k.

Statisticians use various criteria to decide which design is best.

It turns out that most of these criteria are defined by theLaplacian eigenvalues of one of the two graphs defined by theblock design:I the Levi graph, which has v + b vertices,I or the concurrence graph, which has v vertices.

The algebraic approach shows that sometimes all the criteriaprefer highly symmetric designs but sometimes they favourvery different ones.

2/52

An experiment on detergents

In a consumer experiment, twelve housewives volunteer to testnew detergents. There are 16 new detergents to compare, but itis not realistic to ask any one volunteer to compare this manydetergents.

Each housewife tests one detergent per washload for each offour washloads, and assesses the cleanliness of each washload.

The experimental units are the 48 washloads.The housewives form 12 blocks of size 4.

The treatments are the 16 new detergents.

3/52

An experiment on detergents

In a consumer experiment, twelve housewives volunteer to testnew detergents. There are 16 new detergents to compare, but itis not realistic to ask any one volunteer to compare this manydetergents.

Each housewife tests one detergent per washload for each offour washloads, and assesses the cleanliness of each washload.

The experimental units are the 48 washloads.The housewives form 12 blocks of size 4.

The treatments are the 16 new detergents.

3/52

An experiment on detergents

In a consumer experiment, twelve housewives volunteer to testnew detergents. There are 16 new detergents to compare, but itis not realistic to ask any one volunteer to compare this manydetergents.

Each housewife tests one detergent per washload for each offour washloads, and assesses the cleanliness of each washload.

The experimental units are the 48 washloads.The housewives form 12 blocks of size 4.

The treatments are the 16 new detergents.

3/52

Experiments in blocks

I have v treatments that I want to compare.I have b blocks, with k plots in each block.

blocks b k treatments vhousewives 12 4 detergents 16

How should I choose a block designfor these values of b, v and k?

What makes a block design good?

4/52

Experiments in blocks

I have v treatments that I want to compare.I have b blocks, with k plots in each block.

blocks b k treatments vhousewives 12 4 detergents 16

How should I choose a block designfor these values of b, v and k?

What makes a block design good?

4/52

Experiments in blocks

I have v treatments that I want to compare.I have b blocks, with k plots in each block.

blocks b k treatments vhousewives 12 4 detergents 16

How should I choose a block designfor these values of b, v and k?

What makes a block design good?

4/52

Experiments in blocks

I have v treatments that I want to compare.I have b blocks, with k plots in each block.

blocks b k treatments vhousewives 12 4 detergents 16

How should I choose a block designfor these values of b, v and k?

What makes a block design good?

4/52

Two designs with v = 5, b = 7, k = 3: which is better?

Conventions: columns are blocks;order of treatments within each block is irrelevant;order of blocks is irrelevant.

1 1 1 1 2 2 22 3 3 4 3 3 43 4 5 5 4 5 5

1 1 1 1 2 2 21 3 3 4 3 3 42 4 5 5 4 5 5

binary non-binary

A design is binary if no treatment occurs more than once in anyblock.

5/52

Two designs with v = 15, b = 7, k = 3: which is better?

1 1 2 3 4 5 62 4 5 6 10 11 123 7 8 9 13 14 15

1 1 1 1 1 1 12 4 6 8 10 12 143 5 7 9 11 13 15

replications differ by ≤ 1 queen-bee design

The replication of a treatment is its number of occurrences.

A design is a queen-bee design if there is a treatment thatoccurs in every block.

6/52

Two designs with v = 7, b = 7, k = 3: which is better?

1 2 3 4 5 6 72 3 4 5 6 7 14 5 6 7 1 2 3

1 2 3 4 5 6 72 3 4 5 6 7 13 4 5 6 7 1 2

balanced (2-design) non-balanced

A binary design is balanced if every pair of distinct treamentsoccurs together in the same number of blocks.

7/52

Experimental units and incidence matrix

There are bk experimental units.

If ω is an experimental unit, put

f (ω) = treatment on ω

g(ω) = block containing ω.

For i = 1, . . . , v put

ri = |{ω : f (ω) = i}| = replication of treatment i.

For i = 1, . . . , v and j = 1, . . . , b, let

nij = |{ω : f (ω) = i and g(ω) = j}|

= number of experimental units in block j which havetreatment i.

The v× b incidence matrix N has entries nij.

8/52

Experimental units and incidence matrix

There are bk experimental units.

If ω is an experimental unit, put

f (ω) = treatment on ω

g(ω) = block containing ω.

For i = 1, . . . , v put

ri = |{ω : f (ω) = i}| = replication of treatment i.

For i = 1, . . . , v and j = 1, . . . , b, let

nij = |{ω : f (ω) = i and g(ω) = j}|

= number of experimental units in block j which havetreatment i.

The v× b incidence matrix N has entries nij.

8/52

Experimental units and incidence matrix

There are bk experimental units.

If ω is an experimental unit, put

f (ω) = treatment on ω

g(ω) = block containing ω.

For i = 1, . . . , v put

ri = |{ω : f (ω) = i}| = replication of treatment i.

For i = 1, . . . , v and j = 1, . . . , b, let

nij = |{ω : f (ω) = i and g(ω) = j}|

= number of experimental units in block j which havetreatment i.

The v× b incidence matrix N has entries nij.

8/52

Experimental units and incidence matrix

There are bk experimental units.

If ω is an experimental unit, put

f (ω) = treatment on ω

g(ω) = block containing ω.

For i = 1, . . . , v put

ri = |{ω : f (ω) = i}| = replication of treatment i.

For i = 1, . . . , v and j = 1, . . . , b, let

nij = |{ω : f (ω) = i and g(ω) = j}|

= number of experimental units in block j which havetreatment i.

The v× b incidence matrix N has entries nij.

8/52

Experimental units and incidence matrix

There are bk experimental units.

If ω is an experimental unit, put

f (ω) = treatment on ω

g(ω) = block containing ω.

For i = 1, . . . , v put

ri = |{ω : f (ω) = i}| = replication of treatment i.

For i = 1, . . . , v and j = 1, . . . , b, let

nij = |{ω : f (ω) = i and g(ω) = j}|

= number of experimental units in block j which havetreatment i.

The v× b incidence matrix N has entries nij.8/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,I one vertex for each block,I one edge for each experimental unit, with edge ω joining

vertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,

I one vertex for each block,I one edge for each experimental unit, with edge ω joining

vertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,I one vertex for each block,

I one edge for each experimental unit, with edge ω joiningvertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,I one vertex for each block,I one edge for each experimental unit, with edge ω joining

vertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,I one vertex for each block,I one edge for each experimental unit, with edge ω joining

vertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Levi graph

The Levi graph G̃ of a block design ∆ has

I one vertex for each treatment,I one vertex for each block,I one edge for each experimental unit, with edge ω joining

vertex f (ω) (the treatment on ω)to vertex g(ω) (the block containing ω).

It is a bipartite graph,with nij edges between treatment-vertex i and block-vertex j.

9/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2

HHH

HHH

������

���

���

HHHHHH

10/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2

HHH

HHH

������

���

���

HHHHHH

10/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2H

HHH

HH

������

���

���

HHHHHH

10/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2H

HHH

HH

������

���

���

HHHHHH

10/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2H

HHH

HH

������

���

���

HHHHHH

10/52

Example 2: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

11/52

Example 2: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

11/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,

I one edge for each unordered pair α, ω, with α 6= ω,g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Concurrence graph

The concurrence graph G of a block design ∆ has

I one vertex for each treatment,I one edge for each unordered pair α, ω, with α 6= ω,

g(α) = g(ω) (in the same block) and f (α) 6= f (ω):this edge joins vertices f (α) and f (ω).

There are no loops.

If i 6= j then the number of edges between vertices i and j is

λij =b

∑s=1

nisnjs;

this is called the concurrence of i and j,and is the (i, j)-entry of Λ = NN>.

12/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2HH

HHHH

����

��

������

HHHH

HH

x

x

x

x

������

@@

@@

@@

3 4

1 2

Levi graph concurrence graph

can recover design may have more symmetrymore vertices

more edges if k = 2 more edges if k ≥ 4

13/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2HH

HHHH

����

��

������

HHHH

HH

x

x

x

x

������

@@

@@

@@

3 4

1 2

Levi graph concurrence graph

can recover design may have more symmetrymore vertices

more edges if k = 2 more edges if k ≥ 4

13/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2HH

HHHH

����

��

������

HHHH

HH

x

x

x

x

������

@@

@@

@@

3 4

1 2

Levi graph concurrence graphcan recover design may have more symmetry

more verticesmore edges if k = 2 more edges if k ≥ 4

13/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2HH

HHHH

����

��

������

HHHH

HH

x

x

x

x

������

@@

@@

@@

3 4

1 2

Levi graph concurrence graphcan recover design may have more symmetry

more vertices

more edges if k = 2 more edges if k ≥ 4

13/52

Example 1: v = 4, b = k = 3

1 2 13 3 24 4 2

x� � �

xx x

3

4

1 2HH

HHHH

����

��

������

HHHH

HH

x

x

x

x

������

@@

@@

@@

3 4

1 2

Levi graph concurrence graphcan recover design may have more symmetry

more verticesmore edges if k = 2 more edges if k ≥ 4

13/52

Example 2: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

14/52

Example 2: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

14/52

Example 3: v = 15, b = 7, k = 3

1 1 2 3 4 5 62 4 5 6 10 11 123 7 8 9 13 14 15

1 1 1 1 1 1 12 4 6 8 10 12 143 5 7 9 11 13 15

x13 x7 x8 x11

x10

x1

x2

x14

x12 x6

x9x15

x4 x5x3

�����

TTT

�����

TTTTT

bb

bbb

"""""

bbb

bb

"""""

x1

�����

������

`````̀

QQ

QQQ

JJJJJ

DDDDDD

��

���

`````̀QQQQQ

JJJJJ

x12

x11

x10x9x5

x6

x7 x8

x14x15

x13x4 x3 x2

��

@@

@@

��

15/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:

I if i 6= j thenLij = −(number of edges between i and j) = −λij;

I Lii = valency of i = ∑j 6=i

λij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:

I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;

I Lii = valency of i = ∑j 6=i

λij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:

I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;I Lii = valency of i = ∑

j 6=iλij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:

I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;I Lii = valency of i = ∑

j 6=iλij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:

I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;I Lii = valency of i = ∑

j 6=iλij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:

I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;I Lii = valency of i = ∑

j 6=iλij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Laplacian matrices

The Laplacian matrix L of the concurrence graph G is av× v matrix with (i, j)-entry as follows:I if i 6= j then

Lij = −(number of edges between i and j) = −λij;I Lii = valency of i = ∑

j 6=iλij.

The Laplacian matrix L̃ of the Levi graph G̃ is a(v + b)× (v + b) matrix with (i, j)-entry as follows:I L̃ii = valency of i

=

{k if i is a blockreplication ri of i if i is a treatment

I if i 6= j then Lij = −(number of edges between i and j)

=

0 if i and j are both treatments0 if i and j are both blocks−nij if i is a treatment and j is a block, or vice versa.

16/52

Connectivity

All row-sums of L and of L̃ are zero,so both matrices have 0 as eigenvalueon the appropriate all-1 vector.

TheoremThe following are equivalent.

1. 0 is a simple eigenvalue of L;2. G is a connected graph;3. G̃ is a connected graph;4. 0 is a simple eigenvalue of L̃;5. the design ∆ is connected in the sense that all differences between

treatments can be estimated.

From now on, assume connectivity.

Call the remaining eigenvalues non-trivial.They are all non-negative.

17/52

Connectivity

All row-sums of L and of L̃ are zero,so both matrices have 0 as eigenvalueon the appropriate all-1 vector.

TheoremThe following are equivalent.

1. 0 is a simple eigenvalue of L;2. G is a connected graph;3. G̃ is a connected graph;4. 0 is a simple eigenvalue of L̃;5. the design ∆ is connected in the sense that all differences between

treatments can be estimated.

From now on, assume connectivity.

Call the remaining eigenvalues non-trivial.They are all non-negative.

17/52

Connectivity

All row-sums of L and of L̃ are zero,so both matrices have 0 as eigenvalueon the appropriate all-1 vector.

TheoremThe following are equivalent.

1. 0 is a simple eigenvalue of L;2. G is a connected graph;3. G̃ is a connected graph;4. 0 is a simple eigenvalue of L̃;5. the design ∆ is connected in the sense that all differences between

treatments can be estimated.

From now on, assume connectivity.

Call the remaining eigenvalues non-trivial.They are all non-negative.

17/52

Connectivity

All row-sums of L and of L̃ are zero,so both matrices have 0 as eigenvalueon the appropriate all-1 vector.

TheoremThe following are equivalent.

1. 0 is a simple eigenvalue of L;2. G is a connected graph;3. G̃ is a connected graph;4. 0 is a simple eigenvalue of L̃;5. the design ∆ is connected in the sense that all differences between

treatments can be estimated.

From now on, assume connectivity.

Call the remaining eigenvalues non-trivial.They are all non-negative.

17/52

Generalized inverse

Under the assumption of connectivity,the Moore–Penrose generalized inverse L− of L is defined by

L− =(

L +1v

Jv

)−1

− 1v

Jv,

where Jv is the v× v all-1 matrix.

(The matrix1v

Jv is the orthogonal projector onto the null spaceof L.)

The Moore–Penrose generalized inverse L̃− of L̃ is definedsimilarly.

18/52

Generalized inverse

Under the assumption of connectivity,the Moore–Penrose generalized inverse L− of L is defined by

L− =(

L +1v

Jv

)−1

− 1v

Jv,

where Jv is the v× v all-1 matrix.

(The matrix1v

Jv is the orthogonal projector onto the null spaceof L.)

The Moore–Penrose generalized inverse L̃− of L̃ is definedsimilarly.

18/52

Estimation

We measure the response Yω on each experimenal unit ω.

If experimental unit ω has treatment i and is in block m(f (ω) = i and g(ω) = m), then we assume that

Yω = τi + βm + random noise.

We will do an experiment, collect data yω on each experimentalunit ω, then want to estimate certain functions of the treatmentparameters using functions of the data.

We want to estimate contrasts ∑i xiτi with ∑i xi = 0.

In particular, we want to estimate all the simple differencesτi − τj.

19/52

Estimation

We measure the response Yω on each experimenal unit ω.

If experimental unit ω has treatment i and is in block m(f (ω) = i and g(ω) = m), then we assume that

Yω = τi + βm + random noise.

We will do an experiment, collect data yω on each experimentalunit ω, then want to estimate certain functions of the treatmentparameters using functions of the data.

We want to estimate contrasts ∑i xiτi with ∑i xi = 0.

In particular, we want to estimate all the simple differencesτi − τj.

19/52

Estimation

We measure the response Yω on each experimenal unit ω.

If experimental unit ω has treatment i and is in block m(f (ω) = i and g(ω) = m), then we assume that

Yω = τi + βm + random noise.

We will do an experiment, collect data yω on each experimentalunit ω, then want to estimate certain functions of the treatmentparameters using functions of the data.

We want to estimate contrasts ∑i xiτi with ∑i xi = 0.

In particular, we want to estimate all the simple differencesτi − τj.

19/52

Estimation

We measure the response Yω on each experimenal unit ω.

If experimental unit ω has treatment i and is in block m(f (ω) = i and g(ω) = m), then we assume that

Yω = τi + βm + random noise.

We will do an experiment, collect data yω on each experimentalunit ω, then want to estimate certain functions of the treatmentparameters using functions of the data.

We want to estimate contrasts ∑i xiτi with ∑i xi = 0.

In particular, we want to estimate all the simple differencesτi − τj.

19/52

Estimation

We measure the response Yω on each experimenal unit ω.

If experimental unit ω has treatment i and is in block m(f (ω) = i and g(ω) = m), then we assume that

Yω = τi + βm + random noise.

We will do an experiment, collect data yω on each experimentalunit ω, then want to estimate certain functions of the treatmentparameters using functions of the data.

We want to estimate contrasts ∑i xiτi with ∑i xi = 0.

In particular, we want to estimate all the simple differencesτi − τj.

19/52

Variance: why does it matter?

We want to estimate all the simple differences τi − τj.

Put Vij = variance of the best linear unbiased estimator forτi − τj.

The length of the 95% confidence interval for τi − τj isproportional to

√Vij.

(If we always present results using a95% confidence interval, then our interval will contain the truevalue in 19 cases out of 20.)

The smaller the value of Vij, the smaller is the confidenceinterval, the closer is the estimate to the true value(on average), and the more likely are we to detect correctlywhich of τi and τj is bigger.

We can make better decisions about new drugs, about newvarieties of wheat, about new engineering materials . . . if wemake all the Vij small.

20/52

Variance: why does it matter?

We want to estimate all the simple differences τi − τj.

Put Vij = variance of the best linear unbiased estimator forτi − τj.

The length of the 95% confidence interval for τi − τj isproportional to

√Vij. (If we always present results using a

95% confidence interval, then our interval will contain the truevalue in 19 cases out of 20.)

The smaller the value of Vij, the smaller is the confidenceinterval, the closer is the estimate to the true value(on average), and the more likely are we to detect correctlywhich of τi and τj is bigger.

We can make better decisions about new drugs, about newvarieties of wheat, about new engineering materials . . . if wemake all the Vij small.

20/52

Variance: why does it matter?

We want to estimate all the simple differences τi − τj.

Put Vij = variance of the best linear unbiased estimator forτi − τj.

The length of the 95% confidence interval for τi − τj isproportional to

√Vij. (If we always present results using a

95% confidence interval, then our interval will contain the truevalue in 19 cases out of 20.)

The smaller the value of Vij, the smaller is the confidenceinterval, the closer is the estimate to the true value(on average), and the more likely are we to detect correctlywhich of τi and τj is bigger.

We can make better decisions about new drugs, about newvarieties of wheat, about new engineering materials . . . if wemake all the Vij small.

20/52

Variance: why does it matter?

We want to estimate all the simple differences τi − τj.

Put Vij = variance of the best linear unbiased estimator forτi − τj.

The length of the 95% confidence interval for τi − τj isproportional to

√Vij. (If we always present results using a

95% confidence interval, then our interval will contain the truevalue in 19 cases out of 20.)

The smaller the value of Vij, the smaller is the confidenceinterval, the closer is the estimate to the true value(on average), and the more likely are we to detect correctlywhich of τi and τj is bigger.

We can make better decisions about new drugs, about newvarieties of wheat, about new engineering materials . . . if wemake all the Vij small.

20/52

How do we calculate variance?

TheoremAssume that all the noise is independent, with variance σ2.If ∑i xi = 0, then the variance of the best linear unbiased estimator of∑i xiτi is equal to

(x>L−x)kσ2.

In particular, the variance of the best linear unbiased estimator of thesimple difference τi − τj is

Vij =(

L−ii + L−jj − 2L−ij)

kσ2.

(This follows from assumption

Yω = τi + βm + random noise.

by using standard theory of linear models.)

21/52

How do we calculate variance?

TheoremAssume that all the noise is independent, with variance σ2.If ∑i xi = 0, then the variance of the best linear unbiased estimator of∑i xiτi is equal to

(x>L−x)kσ2.

In particular, the variance of the best linear unbiased estimator of thesimple difference τi − τj is

Vij =(

L−ii + L−jj − 2L−ij)

kσ2.

(This follows from assumption

Yω = τi + βm + random noise.

by using standard theory of linear models.)21/52

. . . Or we can use the Levi graph

TheoremThe variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L̃−ii + L̃−jj − 2L̃−ij)

σ2.

(Or βi − βj, appropriately labelled.)

(This follows from assumption

Yω = τi − β̃m + random noise.

by using standard theory of linear models.)

22/52

. . . Or we can use the Levi graph

TheoremThe variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L̃−ii + L̃−jj − 2L̃−ij)

σ2.

(Or βi − βj, appropriately labelled.)

(This follows from assumption

Yω = τi − β̃m + random noise.

by using standard theory of linear models.)

22/52

. . . Or we can use the Levi graph

TheoremThe variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L̃−ii + L̃−jj − 2L̃−ij)

σ2.

(Or βi − βj, appropriately labelled.)

(This follows from assumption

Yω = τi − β̃m + random noise.

by using standard theory of linear models.)

22/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

How do we calculate these generalized inverses?

We need L− or L̃−.

I Add a suitable multiple of J,use GAP to find the inverse with exact rational coefficients,subtract that multiple of J.

I If the matrix is highly patterned,guess the eigenspaces,then invert each non-zero eigenvalue.

I Direct use of the graph: coming up.

Not all of these methods are suitable for generic designs with avariable number of treatments.

23/52

Electrical networks

We can consider the concurrence graph G as an electricalnetwork with a 1-ohm resistance in each edge.Connect a 1-volt battery between vertices i and j.Current flows in the network, according to these rules.

1. Ohm’s Law:In every edge, voltage drop = current × resistance =current.

2. Kirchhoff’s Voltage Law:The total voltage drop from one vertex to any other vertexis the same no matter which path we take from one to theother.

3. Kirchhoff’s Current Law:At every vertex which is not connected to the battery, thetotal current coming in is equal to the total current goingout.

Find the total current I from i to j, then use Ohm’s Law todefine the effective resistance Rij between i and j as 1/I.

24/52

Electrical networks: variance

TheoremThe effective resistance Rij between vertices i and j in G is

Rij =(

L−ii + L−jj − 2L−ij)

.

SoVij = Rij × kσ2.

Effective resistances are easy to calculate withoutmatrix inversion if the graph is sparse.

25/52

Electrical networks: variance

TheoremThe effective resistance Rij between vertices i and j in G is

Rij =(

L−ii + L−jj − 2L−ij)

.

SoVij = Rij × kσ2.

Effective resistances are easy to calculate withoutmatrix inversion if the graph is sparse.

25/52

Electrical networks: variance

TheoremThe effective resistance Rij between vertices i and j in G is

Rij =(

L−ii + L−jj − 2L−ij)

.

SoVij = Rij × kσ2.

Effective resistances are easy to calculate withoutmatrix inversion if the graph is sparse.

25/52

Example 2 calculation: v = 8, b = 4, k = 3

V = 23 I = 24 R =2324

z

z

z

zz z

z

z

����

��

HHHH

HH

HHHH

HH

����

��

������

AAAAAA

AAAAAA

������

5

7

8 6

4

1 2

3

[0]U5

�5

[10]-

10

[5]

Y3

[3]

*

3[6]

66

�3

[9]

U 3[12]

-6

62*13

[23]

j11

26/52

Example 2 calculation: v = 8, b = 4, k = 3

V = 23 I = 24 R =2324

z

z

z

zz z

z

z

����

��

HHHH

HH

HHHH

HH

����

��

������

AAAAAA

AAAAAA

������

5

7

8 6

4

1 2

3

[0]U5

�5

[10]-

10

[5]

Y3

[3]

*

3[6]

66

�3

[9]

U 3[12]

-6

62*13

[23]

j11

26/52

Example 2 calculation: v = 8, b = 4, k = 3

V = 23 I = 24 R =2324

z

z

z

zz z

z

z

����

��

HHHH

HH

HHHH

HH

����

��

������

AAAAAA

AAAAAA

������

5

7

8 6

4

1 2

3

[0]U5

�5

[10]-

10

[5]

Y3

[3]

*

3[6]

66

�3

[9]

U 3[12]

-6

62*13

[23]

j11

26/52

Example 2 calculation: v = 8, b = 4, k = 3

V = 23 I = 24 R =2324

z

z

z

zz z

z

z

����

��

HHHH

HH

HHHH

HH

����

��

������

AAAAAA

AAAAAA

������

5

7

8 6

4

1 2

3

[0]U5

�5

[10]-

10

[5]

Y3

[3]

*

3[6]

66

�3

[9]

U 3[12]

-6

62*13

[23]

j11

26/52

Example 2 calculation: v = 8, b = 4, k = 3

V = 23 I = 24 R =2324

z

z

z

zz z

z

z

����

��

HHHH

HH

HHHH

HH

����

��

������

AAAAAA

AAAAAA

������

5

7

8 6

4

1 2

3

[0]U5

�5

[10]-

10

[5]

Y3

[3]

*

3[6]

66

�3

[9]

U 3[12]

-6

62*13

[23]

j11

26/52

. . . Or we can use the Levi graph

If i and j are treatment vertices in the Levi graph G̃and R̃ij is the effective resistance between them in G̃ then

Vij = R̃ij × σ2.

27/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?3

33

3

3[15][0]

R- �

55

5 R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?

3

33

3

3[15][0]

R- �

55

5 R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?3

33

3

3[15][0]

R- �

55

5 R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?3

33

3

3[15][0]

R- �

55

5

R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?3

33

3

3[15][0]

R- �

55

5 R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Example 2 yet again: v = 8, b = 4, k = 3

V = 23 I = 8 R̃ =238

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

6

-R

?3

33

3

3[15][0]

R- �

55

5 R8 [23]

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

28/52

Optimality: Average pairwise variance

The variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L−ii + L−jj − 2L−ij)

kσ2 = Rijkσ2.

We want all of the Vij to be small.

Put V̄ = average value of the Vij. Then

V̄ =2kσ2 Tr(L−)

v− 1= 2kσ2 × 1

harmonic mean of θ1, . . . , θv−1,

where θ1, . . . , θv−1 are the nontrivial eigenvalues of L.

29/52

Optimality: Average pairwise variance

The variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L−ii + L−jj − 2L−ij)

kσ2 = Rijkσ2.

We want all of the Vij to be small.

Put V̄ = average value of the Vij. Then

V̄ =2kσ2 Tr(L−)

v− 1= 2kσ2 × 1

harmonic mean of θ1, . . . , θv−1,

where θ1, . . . , θv−1 are the nontrivial eigenvalues of L.

29/52

Optimality: Average pairwise variance

The variance of the best linear unbiased estimator of the simpledifference τi − τj is

Vij =(

L−ii + L−jj − 2L−ij)

kσ2 = Rijkσ2.

We want all of the Vij to be small.

Put V̄ = average value of the Vij. Then

V̄ =2kσ2 Tr(L−)

v− 1= 2kσ2 × 1

harmonic mean of θ1, . . . , θv−1,

where θ1, . . . , θv−1 are the nontrivial eigenvalues of L.

29/52

A-Optimality

A block design is called A-optimal if it minimizes the averageof the variances Vij;

—equivalently, it maximizes the harmonic mean of thenon-trivial eigenvalues of the Laplacian matrix L;over all block designs with block size k and the given v and b.

30/52

A-Optimality

A block design is called A-optimal if it minimizes the averageof the variances Vij;—equivalently, it maximizes the harmonic mean of thenon-trivial eigenvalues of the Laplacian matrix L;

over all block designs with block size k and the given v and b.

30/52

A-Optimality

A block design is called A-optimal if it minimizes the averageof the variances Vij;—equivalently, it maximizes the harmonic mean of thenon-trivial eigenvalues of the Laplacian matrix L;over all block designs with block size k and the given v and b.

30/52

Optimality: Confidence region

When v > 2 the generalization of confidence interval is theconfidence ellipsoid around the point (τ̂1, . . . , τ̂v) in thehyperplane in Rv with ∑i τi = 0. The volume of this confidenceellipsoid is proportional to√√√√v−1

∏i=1

1θi

= (geometric mean of θ1, . . . , θv−1)−(v−1)/2

=1√

v× number of spanning trees for G.

31/52

Optimality: Confidence region

When v > 2 the generalization of confidence interval is theconfidence ellipsoid around the point (τ̂1, . . . , τ̂v) in thehyperplane in Rv with ∑i τi = 0. The volume of this confidenceellipsoid is proportional to√√√√v−1

∏i=1

1θi

= (geometric mean of θ1, . . . , θv−1)−(v−1)/2

=1√

v× number of spanning trees for G.

31/52

Optimality: Confidence region

When v > 2 the generalization of confidence interval is theconfidence ellipsoid around the point (τ̂1, . . . , τ̂v) in thehyperplane in Rv with ∑i τi = 0. The volume of this confidenceellipsoid is proportional to√√√√v−1

∏i=1

1θi

= (geometric mean of θ1, . . . , θv−1)−(v−1)/2

=1√

v× number of spanning trees for G.

31/52

D-Optimality

A block design is called D-optimal if it minimizes the volumeof the confidence ellipsoid for (τ̂1, . . . , τ̂v) ;

—equivalently, it maximizes the geometric mean of thenon-trivial eigenvalues of the Laplacian matrix L;—equivalently, it maximizes the number of spanning trees forthe concurrence graph G;—equivalently, it maximizes the number of spanning trees forthe Levi graph G̃;over all block designs with block size k and the given v and b.

32/52

D-Optimality

A block design is called D-optimal if it minimizes the volumeof the confidence ellipsoid for (τ̂1, . . . , τ̂v) ;—equivalently, it maximizes the geometric mean of thenon-trivial eigenvalues of the Laplacian matrix L;

—equivalently, it maximizes the number of spanning trees forthe concurrence graph G;—equivalently, it maximizes the number of spanning trees forthe Levi graph G̃;over all block designs with block size k and the given v and b.

32/52

D-Optimality

A block design is called D-optimal if it minimizes the volumeof the confidence ellipsoid for (τ̂1, . . . , τ̂v) ;—equivalently, it maximizes the geometric mean of thenon-trivial eigenvalues of the Laplacian matrix L;—equivalently, it maximizes the number of spanning trees forthe concurrence graph G;

—equivalently, it maximizes the number of spanning trees forthe Levi graph G̃;over all block designs with block size k and the given v and b.

32/52

D-Optimality

A block design is called D-optimal if it minimizes the volumeof the confidence ellipsoid for (τ̂1, . . . , τ̂v) ;—equivalently, it maximizes the geometric mean of thenon-trivial eigenvalues of the Laplacian matrix L;—equivalently, it maximizes the number of spanning trees forthe concurrence graph G;—equivalently, it maximizes the number of spanning trees forthe Levi graph G̃;

over all block designs with block size k and the given v and b.

32/52

D-Optimality

A block design is called D-optimal if it minimizes the volumeof the confidence ellipsoid for (τ̂1, . . . , τ̂v) ;—equivalently, it maximizes the geometric mean of thenon-trivial eigenvalues of the Laplacian matrix L;—equivalently, it maximizes the number of spanning trees forthe concurrence graph G;—equivalently, it maximizes the number of spanning trees forthe Levi graph G̃;over all block designs with block size k and the given v and b.

32/52

Optimality: Worst case

If x is a contrast in Rv then the variance of the estimator of x>τis (x>L−x)kσ2.

If we multiply every entry in x by a constant c then thisvariance is multiplied by c2; and so is x>x.

The worst case is for contrasts x giving the maximum value of

x>L−xx>x

.

These are precisely the eigenvectors corresponding to θ1,where θ1 is the smallest non-trivial eigenvalue of L.

33/52

Optimality: Worst case

If x is a contrast in Rv then the variance of the estimator of x>τis (x>L−x)kσ2.

If we multiply every entry in x by a constant c then thisvariance is multiplied by c2;

and so is x>x.

The worst case is for contrasts x giving the maximum value of

x>L−xx>x

.

These are precisely the eigenvectors corresponding to θ1,where θ1 is the smallest non-trivial eigenvalue of L.

33/52

Optimality: Worst case

If x is a contrast in Rv then the variance of the estimator of x>τis (x>L−x)kσ2.

If we multiply every entry in x by a constant c then thisvariance is multiplied by c2; and so is x>x.

The worst case is for contrasts x giving the maximum value of

x>L−xx>x

.

These are precisely the eigenvectors corresponding to θ1,where θ1 is the smallest non-trivial eigenvalue of L.

33/52

Optimality: Worst case

If x is a contrast in Rv then the variance of the estimator of x>τis (x>L−x)kσ2.

If we multiply every entry in x by a constant c then thisvariance is multiplied by c2; and so is x>x.

The worst case is for contrasts x giving the maximum value of

x>L−xx>x

.

These are precisely the eigenvectors corresponding to θ1,where θ1 is the smallest non-trivial eigenvalue of L.

33/52

Optimality: Worst case

If x is a contrast in Rv then the variance of the estimator of x>τis (x>L−x)kσ2.

If we multiply every entry in x by a constant c then thisvariance is multiplied by c2; and so is x>x.

The worst case is for contrasts x giving the maximum value of

x>L−xx>x

.

These are precisely the eigenvectors corresponding to θ1,where θ1 is the smallest non-trivial eigenvalue of L.

33/52

E-Optimality

A block design is called E-optimal if it maximizes thesmallest non-trivial eigenvalue of the Laplacian matrix L;

over all block designs with block size k and the given v and b.

34/52

E-Optimality

A block design is called E-optimal if it maximizes thesmallest non-trivial eigenvalue of the Laplacian matrix L;over all block designs with block size k and the given v and b.

34/52

BIBDs are optimal

Theorem (Kshirsagar, 1958; Kiefer, 1975)

If there is a balanced incomplete-block design (BIBD) (2-design)for v treatments in b blocks of size k,then it is A-, D- and E-optimal.Moreover, no non-BIBD is A-, D- or E-optimal.

35/52

D-optimality: spanning trees

A spanning tree for the graph is a collection of edges of thegraph which form a tree (connected graph with no cycles)and which include every vertex.

Cheng (1981), after Gaffke (1978), after Kirchhoff (1847):

product of non-trivial eigenvalues of L= v× number of spanning trees.

So a design is D-optimal if and only if its concurrence graph Ghas the maximal number of spanning trees.

This is easy to calculate by hand when the graph is sparse.

36/52

D-optimality: spanning trees

A spanning tree for the graph is a collection of edges of thegraph which form a tree (connected graph with no cycles)and which include every vertex.

Cheng (1981), after Gaffke (1978), after Kirchhoff (1847):

product of non-trivial eigenvalues of L= v× number of spanning trees.

So a design is D-optimal if and only if its concurrence graph Ghas the maximal number of spanning trees.

This is easy to calculate by hand when the graph is sparse.

36/52

D-optimality: spanning trees

A spanning tree for the graph is a collection of edges of thegraph which form a tree (connected graph with no cycles)and which include every vertex.

Cheng (1981), after Gaffke (1978), after Kirchhoff (1847):

product of non-trivial eigenvalues of L= v× number of spanning trees.

So a design is D-optimal if and only if its concurrence graph Ghas the maximal number of spanning trees.

This is easy to calculate by hand when the graph is sparse.

36/52

What about the Levi graph?

Theorem (Gaffke, 1982)

Let G and G̃ be the concurrence graph and Levi graph for a connectedincomplete-block design for v treatments in b blocks of size k.Then the number of spanning trees for G̃ is equal tokb−v+1 times the number of spanning trees for G.

So a block design is D-optimal if and only ifits Levi graph maximizes the number of spanning trees.

If v ≥ b + 2 it is easier to count spanning trees in the Levi graphthan in the concurrence graph.

37/52

What about the Levi graph?

Theorem (Gaffke, 1982)

Let G and G̃ be the concurrence graph and Levi graph for a connectedincomplete-block design for v treatments in b blocks of size k.Then the number of spanning trees for G̃ is equal tokb−v+1 times the number of spanning trees for G.

So a block design is D-optimal if and only ifits Levi graph maximizes the number of spanning trees.

If v ≥ b + 2 it is easier to count spanning trees in the Levi graphthan in the concurrence graph.

37/52

What about the Levi graph?

Theorem (Gaffke, 1982)

Let G and G̃ be the concurrence graph and Levi graph for a connectedincomplete-block design for v treatments in b blocks of size k.Then the number of spanning trees for G̃ is equal tokb−v+1 times the number of spanning trees for G.

So a block design is D-optimal if and only ifits Levi graph maximizes the number of spanning trees.

If v ≥ b + 2 it is easier to count spanning trees in the Levi graphthan in the concurrence graph.

37/52

Example 2 one last time: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

8 spanning trees 216 spanning trees

38/52

Example 2 one last time: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph

8 spanning trees 216 spanning trees

38/52

Example 2 one last time: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph8 spanning trees

216 spanning trees

38/52

Example 2 one last time: v = 8, b = 4, k = 3

1 2 3 42 3 4 15 6 7 8

x x xxx x x

x

��

��

@@

@@

��

��

@@

@@

15

2

637

4

8

xx

xxx x

x

x���

HHH

HHH

���

���

AAA

AAA

���

5

7

8 6

4

1 2

3

Levi graph concurrence graph8 spanning trees 216 spanning trees

38/52

E-optimality: the edge-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have an edge-cutset of size c(set of c edges whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n.Then

θ1 ≤ c(

1m

+1n

).

If c is small but m and n are both large, then θ1 is small.

39/52

E-optimality: the edge-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have an edge-cutset of size c(set of c edges whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n.Then

θ1 ≤ c(

1m

+1n

).

If c is small but m and n are both large, then θ1 is small.

39/52

E-optimality: the edge-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have an edge-cutset of size c(set of c edges whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n.Then

θ1 ≤ c(

1m

+1n

).

If c is small but m and n are both large, then θ1 is small.

39/52

E-optimality: the vertex-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have a vertex-cutset of size c(set of c vertices whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n,with m′ and n′ edges between them and the vertices in the cutset.Then

θ1 ≤m′n2 + n′m2

mn(n + m),

which is at most c if no multiple edges are involved.

If m′ << m and n′ << n then θ1 is small.

40/52

E-optimality: the vertex-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have a vertex-cutset of size c(set of c vertices whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n,with m′ and n′ edges between them and the vertices in the cutset.Then

θ1 ≤m′n2 + n′m2

mn(n + m),

which is at most c if no multiple edges are involved.

If m′ << m and n′ << n then θ1 is small.

40/52

E-optimality: the vertex-cutset lemma

A design is E-optimal if it maximizes the smallest non-trivialeigenvalue θ1 of the Laplacian L of the concurrence graph G.

LemmaLet G have a vertex-cutset of size c(set of c vertices whose removal disconnects the graph)whose removal separates the graph into components of sizes m and n,with m′ and n′ edges between them and the vertices in the cutset.Then

θ1 ≤m′n2 + n′m2

mn(n + m),

which is at most c if no multiple edges are involved.

If m′ << m and n′ << n then θ1 is small.

40/52

Minimal connectivity

If the block design is connected then bk ≥ b + v− 1.

If the block design is connected and b(k− 1) = v− 1 thenthe Levi graph is a tree andthe concurrence graph is a b-tree of k-cliques.

x x x xx x x x

x x xx

x xx

�����

TTT

���

TTTTTT

bbb

bb

"""""

bb

bbb

"""""

x �����

������

`````̀

QQQ

QQ

JJJJJ

DDDDDD

���

��

`````̀QQQQQ

JJJJJ

xxxx

xx x x

xxxx

x x ��

@@

@@

��

41/52

Minimal connectivity

If the block design is connected then bk ≥ b + v− 1.

If the block design is connected and b(k− 1) = v− 1 thenthe Levi graph is a tree andthe concurrence graph is a b-tree of k-cliques.

x x x xx x x x

x x xx

x xx

�����

TTT

���

TTTTTT

bbb

bb

"""""

bb

bbb

"""""

x �����

������

`````̀

QQQ

QQ

JJJJJ

DDDDDD

���

��

`````̀QQQQQ

JJJJJ

xxxx

xx x x

xxxx

x x ��

@@

@@

��

41/52

Minimal connectivity

If the block design is connected then bk ≥ b + v− 1.

If the block design is connected and b(k− 1) = v− 1 thenthe Levi graph is a tree andthe concurrence graph is a b-tree of k-cliques.

x x x xx x x x

x x xx

x xx

�����

TTT

���

TTTTTT

bbb

bb

"""""

bb

bbb

"""""

x �����

������

`````̀

QQQ

QQ

JJJJJ

DDDDDD

���

��

`````̀QQQQQ

JJJJJ

xxxx

xx x x

xxxx

x x ��

@@

@@

��

41/52

Optimality of minimally connected designs

The Levi graph is a tree,so all connected designs are equally good under the D-criterion.

The Levi graph is a tree,so effective resistance = graph distance,so the only A-optimal designs are the queen-bee designs.

The concurrence graph is a b-tree of k-cliques,so the Cutset Lemmas show thatthe only E-optimal designs are the queen-bee designs.

42/52

Optimality of minimally connected designs

The Levi graph is a tree,so all connected designs are equally good under the D-criterion.

The Levi graph is a tree,so effective resistance = graph distance,so the only A-optimal designs are the queen-bee designs.

The concurrence graph is a b-tree of k-cliques,so the Cutset Lemmas show thatthe only E-optimal designs are the queen-bee designs.

42/52

Optimality of minimally connected designs

The Levi graph is a tree,so all connected designs are equally good under the D-criterion.

The Levi graph is a tree,so effective resistance = graph distance,so the only A-optimal designs are the queen-bee designs.

The concurrence graph is a b-tree of k-cliques,so the Cutset Lemmas show thatthe only E-optimal designs are the queen-bee designs.

42/52

Can we use the Levi graph to find E-optimal designs?

For binary designs with equal replication,θ1(L) is a monotonic increasing function of θ1(L̃).

But, queen-bee designs are E-optimal under minimalconnectivity,and some non-binary designs are E-optimal.

For general block designs, we do not know if we can use theLevi graph to investigate E-optimality.

43/52

Can we use the Levi graph to find E-optimal designs?

For binary designs with equal replication,θ1(L) is a monotonic increasing function of θ1(L̃).

But, queen-bee designs are E-optimal under minimalconnectivity,

and some non-binary designs are E-optimal.

For general block designs, we do not know if we can use theLevi graph to investigate E-optimality.

43/52

Can we use the Levi graph to find E-optimal designs?

For binary designs with equal replication,θ1(L) is a monotonic increasing function of θ1(L̃).

But, queen-bee designs are E-optimal under minimalconnectivity,and some non-binary designs are E-optimal.

For general block designs, we do not know if we can use theLevi graph to investigate E-optimality.

43/52

Can we use the Levi graph to find E-optimal designs?

For binary designs with equal replication,θ1(L) is a monotonic increasing function of θ1(L̃).

But, queen-bee designs are E-optimal under minimalconnectivity,and some non-binary designs are E-optimal.

For general block designs, we do not know if we can use theLevi graph to investigate E-optimality.

43/52

Large blocks; many unreplicated treatments

Suppose that r̄ = ∑i ri

v< 2.

New conventions: blocks are rows, and block size = k + n.

b blocks

k plots n plots

......

v treatments bn treatmentsall single replication

whole design ∆

Whole design ∆ has v + bn treatments in b blocks of size k + n;the subdesign Γ has v core treatments in b blocks of size k;call the remaining treatments orphans.

44/52

Large blocks; many unreplicated treatments

Suppose that r̄ = ∑i ri

v< 2.

New conventions: blocks are rows, and block size = k + n.

b blocks

k plots n plots

......

v treatments bn treatmentsall single replication

whole design ∆

Whole design ∆ has v + bn treatments in b blocks of size k + n;

the subdesign Γ has v core treatments in b blocks of size k;call the remaining treatments orphans.

44/52

Large blocks; many unreplicated treatments

Suppose that r̄ = ∑i ri

v< 2.

New conventions: blocks are rows, and block size = k + n.

b blocks

k plots n plots

......

v treatments bn treatmentsall single replication

whole design ∆

Whole design ∆ has v + bn treatments in b blocks of size k + n;the subdesign Γ has v core treatments in b blocks of size k;

call the remaining treatments orphans.

44/52

Large blocks; many unreplicated treatments

Suppose that r̄ = ∑i ri

v< 2.

New conventions: blocks are rows, and block size = k + n.

b blocks

k plots n plots

......

v treatments bn treatmentsall single replication

whole design ∆

Whole design ∆ has v + bn treatments in b blocks of size k + n;the subdesign Γ has v core treatments in b blocks of size k;call the remaining treatments orphans.

44/52

Levi graph: 10 + 5n treatments in 5 blocks of 4 + n plots

1 2 3 4 A1 · · · An

1 5 6 7 B1 · · · Bn

2 5 8 9 C1 · · · Cn

3 6 8 0 D1 · · · Dn

4 7 9 0 E1 · · · En

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHH

H x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

45/52

Pairwise resistance

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHHH x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

Resistance(A1, A2) = 2Resistance(A1, B1) = 2 + Resistance(block A, block B) in Γ

Resistance(A1, 8) = 1 + Resistance(block A, 8) in ΓResistance(1, 8) = Resistance(1, 8) in Γ

46/52

Pairwise resistance

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHHH x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

Resistance(A1, A2) = 2

Resistance(A1, B1) = 2 + Resistance(block A, block B) in ΓResistance(A1, 8) = 1 + Resistance(block A, 8) in Γ

Resistance(1, 8) = Resistance(1, 8) in Γ

46/52

Pairwise resistance

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHHH x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

Resistance(A1, A2) = 2Resistance(A1, B1) = 2 + Resistance(block A, block B) in Γ

Resistance(A1, 8) = 1 + Resistance(block A, 8) in ΓResistance(1, 8) = Resistance(1, 8) in Γ

46/52

Pairwise resistance

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHHH x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

Resistance(A1, A2) = 2Resistance(A1, B1) = 2 + Resistance(block A, block B) in Γ

Resistance(A1, 8) = 1 + Resistance(block A, 8) in Γ

Resistance(1, 8) = Resistance(1, 8) in Γ

46/52

Pairwise resistance

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����x1

HHHH x5

�������x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

Resistance(A1, A2) = 2Resistance(A1, B1) = 2 + Resistance(block A, block B) in Γ

Resistance(A1, 8) = 1 + Resistance(block A, 8) in ΓResistance(1, 8) = Resistance(1, 8) in Γ

46/52

Sum of the pairwise variances

Theorem (cf Herzberg and Jarrett, 2007)

The sum of the variances of treatment differences in ∆

= constant + V1 + nV3 + n2V2,where

V1 = the sum of the variances of treatment differences in ΓV2 = the sum of the variances of block differences in ΓV3 = the sum of the variances of sums of

one treatment and one block in Γ.

(If Γ is equi-replicate then V2 and V3 are both increasingfunctions of V1.)

Consequence

For a given choice of k, make Γ as efficient as possible.

47/52

Sum of the pairwise variances

Theorem (cf Herzberg and Jarrett, 2007)

The sum of the variances of treatment differences in ∆

= constant + V1 + nV3 + n2V2,where

V1 = the sum of the variances of treatment differences in ΓV2 = the sum of the variances of block differences in ΓV3 = the sum of the variances of sums of

one treatment and one block in Γ.

(If Γ is equi-replicate then V2 and V3 are both increasingfunctions of V1.)

Consequence

For a given choice of k, make Γ as efficient as possible.47/52

A less obvious consequence

Consequence

If n or b is large,and we want an A-optimal design,it may be best to make Γ a complete block design for k′ controls,even if there is no interest incomparisons between new treatments and controls,or between controls.

48/52

Spanning trees

A spanning tree for the Levi graph is a collection edges whichprovides a unique path between every pair of vertices.

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����

��x1

HHHH

HH x5

�������

����x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

The orphans make no difference to the number of spanningtrees for the Levi graph.

49/52

Spanning trees

A spanning tree for the Levi graph is a collection edges whichprovides a unique path between every pair of vertices.

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

������x1

HHHHHH x5

�������

����x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

The orphans make no difference to the number of spanningtrees for the Levi graph.

49/52

Spanning trees

A spanning tree for the Levi graph is a collection edges whichprovides a unique path between every pair of vertices.

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����

��x1

HHHH

HH x5

�������

����x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

The orphans make no difference to the number of spanningtrees for the Levi graph.

49/52

Spanning trees

A spanning tree for the Levi graph is a collection edges whichprovides a unique path between every pair of vertices.

� �

� �

x0

x2�������x7

AAAA x6

BBBBBBB

x4����x8

����

��x1

HHHH

HH x5

�������

����x

9Q

QQ

QQ

QQ x3

xA1

xAn ... @@ xCn

xC1...��

xD1xDn

...@@

xEn xE1

... ��

xB1HH

xBn��· · ·

The orphans make no difference to the number of spanningtrees for the Levi graph.

49/52

D-optimality under very low replication

Consequence

The whole design ∆ is D-optimalfor v + bn treatments in b blocks of size k + nif and only if the core design Γ is D-optimalfor v treatments in b blocks of size k.

Consequence

Even when n or b is large,D-optimal designs do not include uninteresting controls.

50/52

D-optimality under very low replication

Consequence

The whole design ∆ is D-optimalfor v + bn treatments in b blocks of size k + nif and only if the core design Γ is D-optimalfor v treatments in b blocks of size k.

Consequence

Even when n or b is large,D-optimal designs do not include uninteresting controls.

50/52

Conjectures

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the A-optimal design is very different from the D-optimal design,then the E-optimal design is (almost) the same as the A-optimaldesign.

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the connectivity is more than minimal, then all D-optimal designshave (almost) equal replication.

Conjecture (Underpinned by theoretical work by J. R. Johnsonand M. Walters)

If r̄ > 3.5 then designs optimal under one criterion are (almost)optimal under the other criteria.

51/52

Conjectures

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the A-optimal design is very different from the D-optimal design,then the E-optimal design is (almost) the same as the A-optimaldesign.

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the connectivity is more than minimal, then all D-optimal designshave (almost) equal replication.

Conjecture (Underpinned by theoretical work by J. R. Johnsonand M. Walters)

If r̄ > 3.5 then designs optimal under one criterion are (almost)optimal under the other criteria.

51/52

Conjectures

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the A-optimal design is very different from the D-optimal design,then the E-optimal design is (almost) the same as the A-optimaldesign.

Conjecture (Underpinned by theoretical work by C.-S. Cheng)

If the connectivity is more than minimal, then all D-optimal designshave (almost) equal replication.

Conjecture (Underpinned by theoretical work by J. R. Johnsonand M. Walters)

If r̄ > 3.5 then designs optimal under one criterion are (almost)optimal under the other criteria.

51/52

Main References

I R. A. Bailey and Peter J. Cameron:Combinatorics of optimal designs.In Surveys in Combinatorics 2009(eds. S. Huczynska, J. D. Mitchell andC. M. Roney-Dougal),London Mathematical Society Lecture Note Series, 365,Cambridge University Press, Cambridge, 2009, pp. 19–73.

I R. A. Bailey and Peter J. Cameron:Using graphs to find the best block designs.In Topics in Structural Graph Theory(eds. L. W. Beineke and R. J. Wilson),Cambridge University Press, Cambridge, 2013,pp. 282–317.

52/52