series. Such data can be collected either prospectively, following subjects forward in time, or
retrospectively, by extracting measurements on each person from historical records. This
type of data is also often known as repeated measures data, particularly in the social and
behavioural sciences, although in these disciplines such data are more likely to arise from
observing individuals repeatedly under different experimental conditions rather than from a
simple time sequence. Special statistical methods are often needed for the analysis of this
type of data because the set of measurements on one subject tend to be intercorrelated. This
correlation must be taken into account to draw valid scientific inferences. The design of most
such studies specifies that all subjects are to have the same number of repeated measure-
ments made at equivalent time intervals. Such data is generally referred to as balanced
longitudinal data. But although balanced data is generally the aim, unbalanced longitudinal
data in which subjects may have different numbers of repeated measurements made at
differing time intervals, do arise for a variety of reasons. Occasionally the data are unbal-
anced or incomplete by design; an investigator may, for example, choose in advance to take
measurements every hour on one half of the subjects and every two hours on the other half.
In general, however, the main reason for unbalanced data in a longitudinal study is the
occurrence of
missing values
in the sense that intended measurements are not taken, are lost
or are otherwise unavailable. See also Greenhouse and Geisser correction, Huynh–Feldt
correction, compound symmetry, generalized estimating equations, Mauchly test,
response feature analysis, time-by-time ANOVA and split-plot design.[Analysis of
Longitudinal Data, 2nd edition, 2002, P. J. Diggle, P. J. Heagerty, K.-Y. Liang and
S. Zeger, Oxford Scientific Publications, Oxford.]
Lon gitu dinal stu dies: Studies that give rise to
longitudinal data
. The defining characteristic of
such a study is that subjects are measured repeatedly through time.
Long memory processes: A
stationary stochastic process
with slowly decaying or long-range
correlations. See also long-range dependence.[Statistics for Long Memory Processes,
1995, J. Beran, Chapman and Hall/CRC Press, London.]
Long -range dependence: Small but slowly decaying correlations in a stochastic process. Such
correlations are often not detected by standard tests, but their effect can be quite strong.
[Journal of the American Statistical Association, 1997, 92, 881–93.]
Lord, Frederic Mather ( 1 912^2000): Born in Hanover, New Hampshire, Lord graduated from
Dartmouth College in 1936 and received a Ph.D. from Princeton in 1952. In 1944 he joined the
Educational Testing Service and is recognized as the principal developer of the statistical
machinery underlying modern mental testing. Lord died on 5 February 2000 in Naples, Florida.
Lo r d’sparadox: The fact that estimates of treatment effects using
change scores
, where (posttest–
pretest) is regressed on treatment, differ from
analysis of covariance
, where the posttest is
regressed on both treatment and pretest. [ Sociological Methodology, 1990, 20,93–114.]
Lorenz curve: Essentially a graphical representation of the cumulative distribution of a variable,
most often used for income or wealth. If the risks of disease are not monotonically increasing
as the exposure becomes heavier, the data have to be rearranged from the lowest to the highest
risk before the calculation of the cumulative percentages. Associated with such a curve is the
Gini index defined as twice the area between the curve and the diagonal line. This index is
between zero and one, with larger values indicating greater variability while smaller ones
indicate greater uniformity. The further the Lorenz curve lies below the line of equality, the
more unequal is the distribution of, in Figure 91 for example, income. [KA1 Chapter 2.]
Loss function: See decision theory.
261