Association scheme

The theory of association schemes arose in statistics, in the theory of experimental design for the analysis of variance.[1][2][3] In mathematics, association schemes belong to both algebra and combinatorics. Indeed, in algebraic combinatorics, association schemes provide a unified approach to many topics, for example combinatorial designs and coding theory.[4][5] In algebra, association schemes generalize groups, and the theory of association schemes generalizes the character theory of linear representations of groups.[6][7][8]


An n-class association scheme consists of a set X together with a partition S of X × X into n + 1 binary relations, R0, R1, ..., Rn which satisfy:

An association scheme is commutative if for all , and . Most authors assume this property.

A symmetric association scheme is one in which each relation is a symmetric relation. That is:

Every symmetric association scheme is commutative.

Note, however, that while the notion of an association scheme generalizes the notion of a group, the notion of a commutative association scheme only generalizes the notion of a commutative group.

Two points x and y are called i th associates if . The definition states that if x and y are i th associates so are y and x. Every pair of points are i th associates for exactly one . Each point is its own zeroth associate while distinct points are never zeroth associates. If x and y are k th associates then the number of points which are both i th associates of and j th associates of is a constant .

Graph interpretation and adjacency matrices

An association scheme can be visualized as a complete graph with labeled edges. The graph has vertices, one for each point of , and the edge joining vertices and is labeled if and are th associates. Each edge has a unique label, and the number of triangles with a fixed base labeled having the other edges labeled and is a constant , depending on but not on the choice of the base. In particular, each vertex is incident with exactly edges labeled ; is the valency of the relation . There are also loops labeled at each vertex , corresponding to .

The relations are described by their adjacency matrices. is the adjacency matrix of for and is a v × v matrix with rows and columns labeled by the points of .

The definition of an association scheme is equivalent to saying that the are v × v (0,1)-matrices which satisfy

I. is symmetric,
II. (the all-ones matrix),
III. ,
IV. .

The (x, y)-th entry of the left side of (IV) is the number of paths of length two between x and y with labels i and j in the graph. Note that the rows and columns of contain 's:



The term association scheme is due to (Bose & Shimamoto 1952) but the concept is already inherent in (Bose & Nair 1939).[9] These authors were studying what statisticians have called partially balanced incomplete block designs (PBIBDs). The subject became an object of algebraic interest with the publication of (Bose & Mesner 1959) and the introduction of the Bose–Mesner algebra. The most important contribution to the theory was the thesis of P. Delsarte (Delsarte 1973) who recognized and fully used the connections with coding theory and design theory.[10] Generalizations have been studied by D. G. Higman (coherent configurations) and B. Weisfeiler (distance regular graphs).

Basic facts

The Bose–Mesner algebra

The adjacency matrices of the graphs generate a commutative and associative algebra (over the real or complex numbers) both for the matrix product and the pointwise product. This associative, commutative algebra is called the Bose–Mesner algebra of the association scheme.

Since the matrices in are symmetric and commute with each other, they can be diagonalized simultaneously. Therefore, is semi-simple and has a unique basis of primitive idempotents .

There is another algebra of matrices which is isomorphic to , and is often easier to work with.


Let A(3) be the following association scheme with three associate classes on the set X = {1,2,3,4,5,6}. The (i,j) entry is s if elements i and j are in relation Rs.
  1 2 3 4 5 6
1  0   1   1   2   3   3 
2  1   0   1   3   2   3 
3  1   1   0   3   3   2 
4  2   3   3   0   1   1 
5  3   2   3   1   0   1 
6  3   3   2   1   1   0 

Coding theory

The Hamming scheme and the Johnson scheme are of major significance in classical coding theory.

In coding theory, association scheme theory is mainly concerned with the distance of a code. The linear programming method produces upper bounds for the size of a code with given minimum distance, and lower bounds for the size of a design with a given strength. The most specific results are obtained in the case where the underlying association scheme satisfies certain polynomial properties; this leads one into the realm of orthogonal polynomials. In particular, some universal bounds are derived for codes and designs in polynomial-type association schemes.

In classical coding theory, dealing with codes in a Hamming scheme, the MacWilliams transform involves a family of orthogonal polynomials known as the Krawtchouk polynomials. These polynomials give the eigenvalues of the distance relation matrices of the Hamming scheme.

See also



This article is issued from Wikipedia - version of the 6/7/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.