123
GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED DISTRIBUTION SHEN LIANG NATIONAL UNIVERSITY OF SINGAPORE 2005

GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

GENERALIZED LINEAR AND ADDITIVEMODELS WITH WEIGHTED DISTRIBUTION

SHEN LIANG

NATIONAL UNIVERSITY OF SINGAPORE

2005

Page 2: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

GENERALIZED LINEAR AND ADDITIVEMODELS WITH WEIGHTED DISTRIBUTION

SHEN LIANG

(M.Sc., National University Of Singapore )

A THESIS SUBMITTED

FOR THE DEGREE OF DOCTOR OF PHILOSOPHY

DEPARTMENT OF STATISTICS AND APPLIED PROBABILITY

NATIONAL UNIVERSITY OF SINGAPORE

2005

Page 3: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Acknowledgments

I would like to express my gratitude to my supervisors, Professor Young Kinh-Nhue

Truong and Professor Bai Zhidong, for their patient guidance, invaluable inspiration

and concrete comments and suggestions throughout the research leading to this thesis.

My gratitude also goes to the entire faculty of the Department of Statistics and

Applied Probability at NUS for their constant support and motivation during my

studies. I would like to thank all my friends for their generous help.

Finally I would like to express my appreciation to my family, especially my mother,

without their support and encouragement, it would not be possible for the thesis to

come into being.

Page 4: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Contents

1 Introduction 1

1.1 The background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1

1.2 A literature review . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

1.3 Scope and outline of the thesis . . . . . . . . . . . . . . . . . . . . . . 15

2 Generalized Linear Models with Weighted Exponential Families 17

2.1 Weighted Exponential Families . . . . . . . . . . . . . . . . . . . . . 18

2.2 The components of Generalized Linear Models with Weighted Expo-

nential Families . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21

2.2.1 A brief review on GLIM with ordinary exponential family . . 21

2.2.2 The GLIM with weighted exponential families . . . . . . . . . 22

2.3 Estimation of GLIM with Weighted Exponential Families . . . . . . . 23

2.3.1 The iterative weighted least square procedure for the estimation

of β when φ is known. . . . . . . . . . . . . . . . . . . . . . . 24

2.3.2 The double iterative procedure for the estimation of both β and

φ when φ is unknown. . . . . . . . . . . . . . . . . . . . . . . 29

2.4 Asymptotic Distribution of β . . . . . . . . . . . . . . . . . . . . . . 30

2.4.1 The asymptotic variance-covariance matrix of β in the case of

known φ . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30

2.4.2 The asymptotic variance-covariance matrix of β in the case of

unknown φ . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31

2.5 Deviance and Residuals . . . . . . . . . . . . . . . . . . . . . . . . . . 34

2.5.1 Deviance Analysis . . . . . . . . . . . . . . . . . . . . . . . . . 34

i

Page 5: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

2.5.2 Residuals and Model Diagnostics . . . . . . . . . . . . . . . . 36

3 Generalized Additive Models with Weighted Exponential Families 40

3.1 Specification of a Generalized Additive Model with a Weighted Expo-

nential Family . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40

3.1.1 General Assumptions . . . . . . . . . . . . . . . . . . . . . . . 40

3.1.2 Modeling of the Additive Predictor . . . . . . . . . . . . . . . 41

3.2 Estimation of GAM with Weighted Exponential Families . . . . . . . 45

3.2.1 Backfitting Procedure . . . . . . . . . . . . . . . . . . . . . . 46

3.2.2 Penalized Maximum Likelihood Estimation and Iterated Penal-

ized Weighted Least Square procedure . . . . . . . . . . . . . 47

3.2.3 Relation of PMLE with Backfitting . . . . . . . . . . . . . . . 51

3.2.4 Simplification of IPWLE procedure . . . . . . . . . . . . . . . 53

3.2.5 Algorithms with Fixed Smoothing Parameters . . . . . . . . . 56

3.2.6 The Choice of Smoothing Parameters . . . . . . . . . . . . . . 58

3.2.7 Algorithms with Choice of Smoothing Parameters Incorporated 64

4 Special Models with Weighted Exponential Families 73

4.1 Models for binomial data . . . . . . . . . . . . . . . . . . . . . . . . . 73

4.1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73

4.1.2 Weight functions for models with binomial distributions . . . . 76

4.1.3 Link and response functions for models with binomial distribu-

tions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79

4.1.4 Estimation of weighted generalized linear models and weighted

generalized additive model with binomial data . . . . . . . . . 80

4.2 Models for count data . . . . . . . . . . . . . . . . . . . . . . . . . . 82

4.2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

4.2.2 Weight and link functions for models with count data . . . . . 83

4.2.3 Estimation of weighted generalized linear models with count data 85

4.3 Models for data with constant coefficient of variation . . . . . . . . . 86

4.3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

ii

Page 6: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

4.3.2 Components of models with weighted Gamma distribution . . 87

4.3.3 Estimation of generalized linear additive models with weighted

Gamma distributions . . . . . . . . . . . . . . . . . . . . . . . 88

5 Comparison of Weithed and Unweighted Models by Simulation Stud-

ies 90

5.1 The Effect of Weighted Sampling on Generalized Linear Models . . . 90

5.1.1 Studies on generalized linear models with weighted Binomial

distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91

5.1.2 Studies on generalized linear models with weighted Poisson dis-

tributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94

5.1.3 Studies on generalized linear models with weighted Gamma dis-

tributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98

5.2 The Effect of Weighted Sampling on Generalized Additive Models . . 99

5.2.1 Studies on generalized additive models with length biased Bi-

nomial distribution . . . . . . . . . . . . . . . . . . . . . . . . 99

5.2.2 Studies on generalized additive models with length biased Gamma

Distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 105

6 Concluding Remarks and Open Problems 110

6.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 110

6.2 Further Investigation . . . . . . . . . . . . . . . . . . . . . . . . . . . 111

iii

Page 7: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Abstract

In many practical problems, due to constraints on ascertainment, the chances of

individuals being sampled from the population differ from individual to individual,

which causes what is referred to as the sampling bias. In the existence of sampling

bias, the distribution of the observed variable of interest, say Y , is not the same as the

distribution of Y in nature. This makes a striking difference from the usual simple

random sampling where each individual has the equal chance to be sampled and the

distribution of observed Y and the distribution of Y in nature are the same. Ignoring

the sampling bias in statistical inference can cause serious problems and result in

misleading conclusions. This gives rise to the notion of weighted distribution.

Most research on weighted distributions so far has been devoted to the inference

on the population mean, the density function and the cumulative distribution func-

tion of the variable of interest. But not much attention has been paid to regression

models with weighted distributions. However, such models are important and useful

in practice, especially, in medical studies and genetic analysis. This motivated us to

explore such models and to study their properties. In this thesis, we study general-

ized linear and additive models with weighted response variables that include linear

regression models as special cases.

In this thesis, a systematic treatment is made to the generalized linear and additive

models with weighted exponential families. The general theory on the formulation of

these models and their properties are established. Various aspects of these models

such as the estimation, diagnostics and inference are studied. Computation algo-

rithms are developed. A comprehensive simulation study is carried out to evaluate

the effect of sampling bias on the generalized linear and additive models as well.

iv

Page 8: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 1

Introduction

1.1 The background

In statistical problems, people are interested in the distribution of a particular char-

acteristic, say Y , of a population. To make inference on the distribution of Y , people

take a sample (Y1, . . . , Yn) from the population such that the Yi’s are independent

identically distributed (iid). If, under the mechanism of the sampling, each indi-

vidual of the population has an equal chance being sampled, the distribution of the

observed Yi’s is the same as that of Y . However, in many practical problems, due

to constraints on ascertainment, the chance of being sampled is different for different

individuals. In this case, the distribution of the observed Yi’s is no longer the same

as that of Y . This gives rise to the notion of weighted distribution.

To distinguish between the distribution of Y and the distribution of the observed

Yi’s, the distribution of Y will be referred to as the original distribution. Let the

probability density function (pdf) of the original distribution be denoted by f(y).

Suppose that an individual with characteristic Y = y is sampled with probability

proportional to w(y) ≥ 0. Then the observed Yi’s will have a distribution with pdf

given by

fW (y) =w(y)f(y)

µW,

where µW = E[w(Y )] =∫

w(y)f(y)dy. The distribution of the observed Yi’s will

be referred to as the weighted distribution. The function w(y) is referred to as the

1

Page 9: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

weight function. In different problems, the weight function takes different forms. The

following is a list of the forms of the weight function provided by Patil, Rao and

Ratnaparkhi (1986):

1. w(y) = yα, α > 0.

2. w(y) =(

y2

)α, α = 1, 1/2, 0 < α < 1, for integer y.

3. w(y) = y(y − 1) · · · (y − r), where r is a integer.

4. w(y) = eαy.

5. w(y) = αy + β.

6. w(y) = 1 − (1 − β)y, 0 < β < 1.

7. w(y) = (αy + β)/(δy + γ).

8. w(y) = G(y) = Prob(Z ≤ y) for some random variable Z.

9. w(y) = G(y) = Prob(Z > y) for some random variable Z.

10. w(y) = r(y), where r(y) is the probability of “survival” of observation y.

Note that the parameters involved in the weight function do not depend on the original

distribution though they might be unknown. In the special case that w(y) = yα, the

weighted distribution is called a length-biased (or size-biased) distribution of order

α. In particular, if α = 1, the weighted distribution is simple called length-biased

distribution.

The following are some examples of weighted distributions in practical application

scattered in the literature.

Example 1. In the study of the distribution of the number of children with

certain rare disease (e.g., albino children) in families with proneness to produce such

children, it is impractical to ascertain such families by a simple random sampling,

and a convenient sampling method is first to discover a child with such disease (from

the visiting of the child to a hospital, or through some other means) and then count

2

Page 10: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

the number of his siblings with the disease. If a child with the disease is diagnosed

positive with probability β, then the probability that a family with y diseased children

is ascertained is w(y) = 1− (1− β)y. Thus the observed number of diseased children

follows a weighted binomial distribution with weight function w(y). See Haldane

(1938), Fisher (1934), Rao (1965), and Patil and Rao (1978).

Example 2. In the study of wildlife population density, a sampling scheme

called quadrat sampling has been widely used. Quadrat sampling is carried out by

first selecting at random a number of quadrats of fixed size from the region under

investigation and then obtain the number of animals in each quadrat by an aerial

sighting. Animals occur in groups. The sampling is such that if at least one animal

in a group is sighted then a complete count is made of the group and the number

of animals is ascertained. If each animal is sighted with equal probability β then a

group with y animals is ascertained with probability w(y) = 1 − (1 − β)y. Suppose

the real distribution of the number of animals in groups has a density function f(y).

Then the observed number of animals in groups follow a weighted distribution with

density function w(y)f(y)/∫

w(y)f(y)dy. See Cook and Martin (1974) and Patil and

Rao (1978).

Example 3. Another sampling scheme, the line-transect sampling, has been

used to estimate the abundance of plants or animals of a particular species in a given

region. The line-transect method consists of drawing a baseline across the region

to be surveyed and then drawing a line transect through a randomly selected point

on the baseline. The surveyor views the surroundings while walking along the line

transect and includes the sighted objects of interest in the sample. Usually, individual

objects cluster in groups and it is appropriate to take the clusters as sampling units.

Estimates of cluster abundance can be adjusted to individual abundance using the

recorded cluster sizes. It is obvious that the nearer the cluster and the larger its

size, the more likely the cluster will be sighted. In other words, the probability that

a cluster is sampled is proportional to its size. The size of a sampled cluster then

follows a weighted distribution relative to its real world distribution. See Drummer

and McDonald (1987).

3

Page 11: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Example 4. In the sampling of textile fibers, an assembly of fibers parallel to

a common axis is considered. The position of each fiber along the axis is defined

by the coordinate of its left-end and the length. The numbers of fibers in a typical

cross-section may vary from about 20 for a fine yarn to a thousand or more at earlier

stages of processing. The fibers are well mixed in the early stages of processing and

their left-ends lie approximately in a Poisson process along the axis. For an unbiased

sampling, a very short sampling interval along the axis is chosen and all those fibers

whose left-ends lie in the intervals are taken. However, in practice, because of the

near randomness of the fiber arrangement, the crucial step is the isolation of the

relevant fibers rather than the positioning of the sampling intervals. This can cause

practical difficulties in the sampling. An alternative sampling method is as follows.

The assembly is gripped at a sampling point and those fibers which are not gripped,

i.e. not crossing the sampling point, are gently combed out. The remaining fibers

constitute the sample. Under this sampling scheme, the chance of each fiber being

selected is proportional to its length. The observed fiber length therefore follows a

length-biased distribution. See Cox (1969).

Example 5. In the study of early screening program to detect individuals who

are unaware that they have certain particular diseases, the population of concern

is examined at a particular time point, those persons who have the diseases at the

checking time are picked up. However, the cases of diseases detected by the early

screening program are not a simple random sample from the general distribution of

cases in the screened population. Actually, a long-duration pre-clinical disease has

a higher chance to be detected than does a short-duration pre-clinical disease. The

probability of each disease case being selected is proportional to its length of pre-

clinical state duration. This provides another example of length-biased sampling.

See Zelen (1974).

Example 6. In recent years, the discovery of genes associated with the risk for

common cancers has led to an intense research effort to evaluate the epidemiological

characteristics of germ line abnormalities in these genes. In order to estimate the

lifetime risk associated with genetic abnormalities that predispose individuals to can-

4

Page 12: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

cer, some studies have used data from family members of probands ascertained from

population-based incident cases of cancer. Because mutations in some genes occur in a

small percentage of the population at risk for cancer, genotyping of population-based

control subjects will scarcely identify carriers. The strategy of using case patients to

identify probands with and without mutations and then using the relatives of these

probands to calculate penetrance is appealing. The individuals with higher risks are

more likely to be sampled. Length-biased sampling again comes into play.

Some more examples such as analysis of intervention data, modeling heterogeneity

and extraneous variation, meta-analysis incorporating heterogeneity and publication

bias, statistical analysis incorporating over-dispersion and heterogeneity in teratolog-

ical binary data, etc.. can be found in Patil (2002).

Mathematically, the weighted distribution is nothing special but an ordinary dis-

tribution. However, the statistical inference with weighted distributions differs from

the usual inference. The data available are observations on the weighted distribu-

tion, but it is the original distribution that needs to be inferenced. This poses new

problems in statistical inference.

1.2 A literature review

The origin of weighted distribution can be traced back to Fisher (1934) who studied

the effects of ascertainment methods on the estimation of frequencies. The initial

idea of length biased sampling appeared in Cox (1962). The notion of weighted

distribution in general terms was formulated and developed by Rao (1965).

Rao (1965) identified the following three main sources that give rise to weighted

distributions:

(i) Non-observability of events. Some kinds of events may not be ascertainable

although they occur in nature. A typical example is the study on albino children. If

a family with both parents heterozygous for albinism has no albino children, there is

no chance the family will be observed, since there is no evidence that the parents are

both heterozygous unless at least one albino child is born. The actual frequency of

5

Page 13: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

the event “zero albino children” is unascertainable. Hence, the observed distribution

is the original distribution truncated at one — a special case of weighted distribution.

(ii) Partial destruction of observation. Rao noticed that an observation produced

by nature (as number of eggs, number of accidents etc) may be partially destroyed

or may be only partially ascertained. In such a case, the observed distribution is

a distorted version of the original distribution. If the mechanism underlying the

partial destruction is known, the distribution appropriate to the observed values can

be derived from the assumption on the original distribution. It is typically a weighted

version of the original distribution.

(iii) Sampling with unequal chances of selection. In many practical situations, a

simple random sampling is not feasible. Sampling is carried out by certain “encoun-

tering” approach. For instance, fish are collected via a net, plants are observed when

walking along a transect line, and birds are identified while they are heard in wet

land surveys. Such sampling approach leads naturally to the weighted distribution

which alters the original distribution.

Since the fundamental work of Rao, weighted distributions have attracted much

attention of the statisticians.

The estimation of the mean of the original distribution based on length-biased data

was first considered by Cox (1969). Let Y, f(y) and F (y) denote the random variable

produced by nature and its probability density function and cumulative distribution

function respectively. Denote by Y W , fW (y) and F W (y) the corresponding weighted

versions. If the weight function is w(y) = y, that is, the weighted distribution is in

fact the length-biased distribution, then it is easy to see that

E[(Y W )r] =µr+1

µ,

where µr is the rth moment of Y . It then follows immediately that

E[(Y W )−1] =1

µ,

Var((Y W )−1) =1

µ2[µµ−1 − 1].

6

Page 14: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Therefore, an unbiased estimate of 1/µ can be provided by

1

n

∑(Y W

i )−1.

Based on the above results, Cox proposed the estimate of µ as follows:

µ = n/∑ 1

Y Wi

.

By the central limit theorem, this estimate has an asymptotic normal distribution

with asymptotic mean µ and asymptotic variance

µµ−1 − 1

n.

The estimate of µ given above is biased. In fact,

E(µ) = µ(1 + n−1(µµ−1 − 1)) + o(1

n).

The bias at the first order is given by n−1µ(µµ−1 − 1). Hence, Sen (1987) suggested

to use jackknife method to reduce the bias of the estimate for the statistical inference

on µ.

Cox (1969) also dealt with the estimation of F (y). Let

Ui(y) =

1/Yi, if Yi ≤ y,0, if Yi > y

,

Vi = 1/Yi.

Cox proposed the following estimate of F (y):

Fg(y) =∑

Ui(y)/∑

Vi.

For any fixed y, Fg(y) is asymptotically normally distributed with mean F (y) and

varianceµµ−1(y) − 2µµ−1(y)F (y) + µµ−1(F (y))2

n.

Some general properties of weighted distributions were studied by several authors.

Patil and Ord (1975) studied the form-invariant property of certain distribution fam-

ilies. Patil and Rao (1978) studied the relationship among the means of original and

7

Page 15: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

different weighted distributions. Bayarri and DeGroot (1987, 1989) and Patil and

Taillie (1989) studied the Fisher information of weighted distributions and their rela-

tionship with that of the original distributions. We discuss these issues in more detail

as follows.

A distribution family with probability density function f(y; θ) is called form-

invariant under length-biased sampling of order α if the distribution of the length-

biased variable Y W still belongs to the same family with a different parameter θ′,

i.e.

fW (y; θ) ≡ f(y; θ′).

Patil and Ord (1975) showed that, under certain regularity conditions, any form-

invariant family under length-biased sampling must belong to a more general family

which they called the log-exponential family. They established the following result:

Theorem 1 Let θ(α) denote the θ′induced by order α. Suppose f(y, θ) satisfies the

following regularity conditions:

(a) 0 < limα→0(θ(α)−θ

α) < ∞ and 0 < limα→0(

αθ(α)−θ

) < ∞.

(b) limα→0 |(ln(µW (α))/α)| = |E[lnY ]| < ∞, where µW (α) is the mean of Y W cor-

responding to order α.

Then a necessary and sufficient condition for f(y, θ) to be form-invariant under

length-biased sampling of order α is that the p.d.f. is of the form

f(y; θ) = yθa(y)/m(θ) = expθ ln y + A(y) −B(θ),

where a(y) = exp(A(y)), m(θ) = exp(B(θ)), and E[lnY ] = m′(θ)/m(θ) = B ′(θ).

Patil and Rao (1978) established the following properties about the means of

weighted distributions.

Theorem 2 Let a non-negative random variable Y have p.d.f. f(y) with E(Y ) < ∞.

Let Y W have p.d.f. fW (y) = yf(y)/E(Y ). Then E(Y W ) − E(Y ) = V (Y )/E(Y ),

where V stands for variance, and therefore E(Y W ) > E(Y ) for non-degenerate Y .

8

Page 16: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Theorem 3 Let a random variable Y have p.d.f. f(y). Further let the weight func-

tion w(y) > 0 have E(w(Y )) < ∞. Let Y W be the w-weighted random variable of Y

with p.d.f. fW (y) = w(y)f(y)/E(w(Y )). Then E(Y W ) > E(Y ) if Cov[Y, w(Y )] > 0

and E(Y W ) < E(Y ) if Cov[Y, w(Y )] < 0.

Theorem 4 Let Y and Y w be defined as above. Further let Y be non-negative. Then

E(Y W ) > E(Y ) if w(y) increase in y and E(Y W ) < E(Y ) if w(y) decrease in y.

Theorem 5 Let a non-negative random variable Y have p.d.f. f(y). Let the weight

function wi(y) > 0 have E(wi(Y )) < ∞ for i = 1, 2, defining the corresponding

wi-weighted random variable’s of Y denoted by Y Wi . Then E(Y W2) > E(Y W1) if

r(y) = w2(y)/w1(y) increase in y and E(Y W2) < E(Y W1) if r(y) decrease in y.

It is natural to ask whether or not the observed “weighted” data is more informa-

tive than the corresponding data produced in nature. This question is relevant from

the viewpoint of comparison of experiments when both the length-biased sampling

and simple random sampling are possible options in practice. Let f(y; θ) and fW (y; θ)

be the density functions of the original and the weighted distributions, respectively,

where θ is a vector of unknown parameters. Denote by I(θ) and IW (θ) the respective

Fisher information matrices of θ. An intrinsic comparison between the length-biased

sampling and simple random sampling is possible when IW (θ)−I(θ) is either positive

definite or negative definite. The length-biased sampling is favored or not depending

on whether the difference is positive definite or negative definite. If IW (θ) − I(θ) is

indefinite, the comparison can be made in two ways: (i) in terms of a suitable scalar-

valued measure of the joint information of θ, the generalized variance, det[I(θ)], is

such a natural measure, and (ii) in terms of the standard error of the estimator for a

scalar-valued function of θ that is considered most relevant to the scientific problem

at hand.

Bayarri and DeGroot gave an extensive treatment to the single-parameter families.

Their results are applied to Binomial, Poisson and Negative binomial distributions

and found that length-biased sampling has different effects on Fisher information with

9

Page 17: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

different original distributions. For binomial distribution, a simple random sample

is more informative than a length-biased sample. For Poisson distribution, a simple

random sample and a length-biased sample are equally informative. For negative bi-

nomial distribution, a length-biased sample is more informative than a simple random

sample.

Bayarri and DeGroot derived some specific results for the special case that the

weight function is given below:

w(y) =

1, if y ∈ S0, if y ∈ S

.

The weighted version of the original pdf f is expressed as

fW (y; θ) = f(y; θ)/s(θ),

where s(θ) = Pr(y ∈ S|θ). Bayarri and DeGroot (1987) showed that if Y is restricted

to the set y ≥ τ or τ1 ≤ y ≤ τ2, then the weighted sampling data is more informative

than simple random sampling data. If Y is restricted to the set S = y : y ≤ τ1 or y ≥τ2, where τ1 ≤ τ2, then the converse is true. Further it is shown that, for exponential

families,

IW (θ) = I(θ) +d2

dθ2ln(s(θ)).

Hence,

IW (θ) ≥ I(θ) if ln(s(θ)) is convex,

IW (θ) ≤ I(θ) if ln(s(θ)) is concave.

Patil and Taillie (1989) considered observations from a one-parameter exponential

family

f(y) =c(y)e−θA(y))

M(θ)

and an equal number of observations from the w(y)-weighted version of this family

fW (y) =w(y)c(y)e−θA(y))

MW (θ).

They showed that

10

Page 18: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

1. The weighted version is uniformly more informative for θ if and only if MW (θ)/M(θ)

is log convex.

2. The weighted version is uniformly less informative for θ if and only if MW (θ)/M(θ)

is log concave.

3. The original f and the weighted fW are uniformly equally informative (Fisher

neutral) if and only if MW (θ)/M(θ) is log-linear. For given w, this characterizes

Fisher neutrality by a functional equation involving M .

Patil and Taillie (1989) also studied the two-parameter gamma distribution, neg-

ative binomial distribution, and lognormal distribution with weight function w(y) =

yα. In all these cases, Iw(θ)−I(θ) is indefinite. For the gamma and negative binomial

distributions, the weighted observations are less informative for θ when generalized

variance is used as the criterion. However, the relative efficiency is quite close to

unity unless the shape parameter is small (less than 0.5, say). For the lognormal

distribution, weighted and unweighted observations are equally informative in terms

of the generalized variance, because det[IW (θ)] = det[I(θ)] for all θ.

The estimation of the original density function f using length-biased data was

dealt with by Bhattacharyya, Franklin and Richardson (1988) and Jones (1991).

Bhattacharyya, Franklin and Richardson (1988) proposed a kernel estimator by

making use of the relationship between the weighted density function fW (y) and the

original density function f(y): fW (y) = yf(y)/µ. By the usual kernel estimation, the

weighted density function fW (y) is estimated by

fWn (y) =

1

nhn

n∑j=1

K(y − yW

j

hn),

where K(·) is a known kernel function, hn is a chosen bandwidth parameter and yWj ’s

are observations under length-biased sampling. The estimator of f(y) proposed by

Bhattacharyya, Franklin and Richardson (1988) is given by

fn(y) = µnfWn (y)/y,

11

Page 19: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where µn is an estimate of µ which can be taken as the estimate proposed by Cox

(1969), i.e., µn = n/∑n

i=1(1/yWi ). They established the following results:

Theorem 6 Let µn be any consistent estimator of µ. Then

1. fn(y) is a consistent estimator of f(y).

2. If fW is uniformly continuous and nh2n → ∞, then for each positive α, ε,

Psupy≥α |fn(y) − f(y)| < ε → 1.

Theorem 7 Let µn be any estimator of µ such that√

nhn[µn − µ]P→ 0. Then

1.√

nhn[fn(y) − E(fWn (y))µy

]d→ N(0, a2), where a2 = µf(y)

y

∫K2(z)dz.

2.√

nhn[fn(y) − f(y)]d→ N(0, a2) iff

√nhn[fW

n (y) − fW (y)] → 0.

3. If f ′′(y) exists, then√

nhn[fn(y) − f(y)]d→ N(0, a2) iff nh5

n → 0.

Theorem 8 Let m be a natural number and let fn(y) = nfWn (y)/(x

∑ni=1 Y −1

i ). Sup-

pose that E(Y 2mi ) < ∞. Then E(fn(y) − f(y))m → 0 as n → ∞.

Johns (1991) proposed a new kernel estimate which is derived from smoothing the

nonparametric maximum likelihood estimate. The new kernel estimate is given by

f(y) = n−1µn∑

i=1

Y −1i Kh(y − Y W

i ),

where h is the bandwidth that controls the degree of smoothing and Kh(x) = h−1K(h−1x),

and µ is taken as the same Cox estimate.

An interesting property of f is that , if µ were known, it would have precisely

the same expectation as the kernel estimator with the same bandwidth h based on

a simple random sample, i.e., E[f(y)] = (kh f)(y), where denotes convolution.

However, the variance of f would be different and in fact is given by

var(f (y)) = n−1(K2h γ)(y)− (Kh f)2(y),

12

Page 20: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where γ(y) = µf(y)/y. As n → ∞ and h = h(n) → 0 in such a way that nh(n) → ∞,

an expression for the asymptotic mean squared error of f (y) can be derived as

MSEf(y) 1

4σ4

Kh4f ′′2(y) +γ(y)

nhR(K),

where σ2K =

∫x2K(x)dx and R(K) denotes

∫K2(x)dx. It follows immediately from

the above equation that the integrated mean squared error is given by

IMSE(f ) 1

4σ4

Kh4R(f ′′) +µµ−1

nhR(K),

provided R(f ′′) < ∞ and µ−1 < ∞.

Compared with the estimate of Bhattacharyya, Franklin and Richardson (1988),

the new estimate is necessarily a probability density, it performs better near zero

and has better asymptotic mean integrated squared error properties, and it can be

more readily extended to related problems such as density derivative estimation. The

estimate is easily generalized to general weighted distribution as follows.

ˆfW (y) = n−1µW

n∑i=1

w(Yi)−1Kh(y − Yi),

where µW = n∑

w(Yi)−1−1 which is an estimate of µW .

The Bayesian analogue of the original distribution in the context of weighted dis-

tribution was dealt with by Mahfoud and Patil (1981) and Patil, Rao and Ratnaparkhi

(1986). They obtained the following results.

Theorem 9 (Mahfoud and Patil, 1981): Consider the usual Bayesian inference in

conjunction with (Y, θ) having joint pdf f(y, θ) = fY |θ(y|θ)π(θ) = πθ|Y (θ|y)fY (y).

The posterior πθ|Y (θ|y) = fY |θ(y|θ)π(θ)/fY (y) = l(θ|y)π(θ)/E[l(Θ|y)] is a weighted

version of the prior π(θ). The weight function is the likelihood function of θ for the

observed y.

Theorem 10 (Patil, Rao and Ratnaparkhi, 1986): Consider the usual Bayesian in-

ference in conjunction with (Y, θ) with pdf f(y, θ) = fY |θ(y|θ)π(θ) = πθ|Y (θ|y)fY (y).

13

Page 21: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Let w(y, θ) = w(y) be the weight function for the distribution of Y |θ, so that the pdf of

Y W |θ is w(y)fY |θ(y|θ)/ω(θ), where ω(θ) = E[w(Y )|θ]. The original and the weighted

posteriors are related by

πθ|Y (θ|y) =ω(θ)πW

θ|Y (θ|y)

E[ω(θ)|Y W = y].

Further, the weighted posterior random variable θW |Y W = y is stochastically greater

or smaller than the original posterior random variable θ|Y = y according as ω(θ) is

monotonically decreasing or increasing as a function of θ.

Bivariate weighted distributions have also been introduced and studied, see Patil

and Rao (1978) and Mahfoud and Patil (1982). Let (X, Y ) be a pair of nonnegative

random variables with a joint pdf f(x, y) and let w(x, y) be a nonnegative weight

function such that E[w(X, Y )] exists. The weighted version of f(x, y) is

fW (x, y) =w(x, y)f(x, y)

E[w(X, Y )].

The corresponding weight version of (X, Y ) is denoted by (X, Y )W . The marginal

and conditional distributions of (X, Y )W can be derived as

fW (x) =E[w(x, Y )|x]f(x)

E[w(X, Y )],

fW (y|x) =w(x, y)f(y|x)

E[w(x, Y )|x].

Obviously, both are weighted versions of the corresponding marginal and conditional

distributions of (X, Y ).

Weight functions of special practical interests for bivariate distributions are listed

below:

1. w(x, y) = xα, α > 0.

2. w(x, y) = w(y).

3. w(x, y) = x + y.

14

Page 22: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

4. w(x, y) = xαyβ.

5. w(x, y) = max(x, y).

6. w(x, y) = min(x, y).

The following results are of some interest.

Theorem 11 (Patil and Rao, 1978). Let (X, Y ) be a pair of nonnegative random

variables with pdf f(x, y). Let w(x, y) = w(y), as is the case in sample surveys

involving sampling with probability proportional to size. Then the random variable X

and XW are related by

fW (x) =E[w(Y )|x]f(x)

E[w(Y )].

Note that XW is a weighted version of X, and the regression of w(Y ) on X serves as

the weight function.

Theorem 12 (Mahfoud and Patil, 1982). Let (X, Y ) be a pair of nonnegative in-

dependent random variables with pdf f(x, y) = fX(x)fY (y), let w(x, y) = max(x, y).

Then the random variables (X, Y )W are dependent. Furthermore, the regression of

Y W on XW by E[Y W |XW = x] is a decreasing function of x.

1.3 Scope and outline of the thesis

As briefly reviewed in the last section, most research on weighted distributions has

been devoted to the estimation of the population mean, the density function and

the cumulative distribution function of the weighted variable itself. It seems that so

far not much attention has been paid to regression models with weighted response

variables. However, such models are important and useful in practice, especially, in

medical studies and genetic analysis. This motivated us to explore such models and to

study their properties. In this thesis, we study generalized linear and additive models

with weighted response variables that include regression models as special cases. We

are going to give a systematic treatment to these models. We will develop a general

15

Page 23: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

theory on the formulation of the models and their properties. We will investigate

various aspects of the models such as the estimation, diagnostics and inference of the

models. We will develop algorithms for the computation.

The thesis is organized as follows.

In Chapter 2, the general theory on weighted exponential family and generalized

linear models with weighted exponential families are developed. It includes the def-

inition and properties of the weighed exponential families, the basic components of

generalized linear models with weighted exponential families, the estimation issue,

the asymptotic properties of the estimates, the diagnostics of these models, etc..

In Chapter 3, the theory on generalized linear models with weighed exponential

families is extended to generalized additive models with weighted exponential fami-

lies. Specific aspects of the latter models are studied. It includes the modeling of the

additive predictors, the particular issues associated with the fitting of the general-

ized additive models with weighted exponential families, the choice of the smoothing

parameters, and a host of computation algorithms.

In Chapter 4, special models are treated in detail. It includes models for weighted

binomial responses, models for weighed count data, and models for weighted data

with constant coefficient of variation.

In Chapter 5, we evaluate the effect of sampling bias through the comparison be-

tween weighted and unweighted generalized linear and additive models by simulation

studies.

16

Page 24: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 2

Generalized Linear Models withWeighted Exponential Families

In Chapter 1, we introduced the general notion of a weighted distribution. In this

chapter and subsequent chapters, we concern ourselves with those weighted distribu-

tions whose original distributions are exponential families. For convenience, we refer

to such weighted distributions as weighted exponential families. A class of statistical

models for exponential families, called generalized linear models (GLIM), has been

investigated intensively in the past 20 years or so. A comprehensive treatment of

GLIM is given by McCullagh and Nelder (1989). In a generalized linear model, a cer-

tain feature of the exponential family under consideration (not necessarily the mean

of the distribution) depends on a set of predictor variables through a linear predictor.

In this chapter, we extend GLIM to weighted exponential families. To distinguish,

the original GLIM will be referred to as the GLIM with ordinary exponential family

and the extended GLIM will be referred to as the GLIM with weighted exponential

family. The theory on the GLIM with weighted exponential family is developed in

this chapter. In Section 2.1, we give the definition of a weighted exponential family

and some of its properties. In Section 2.2, we discuss the components of the GLIM

with weighted exponential families. In Section 2.3, we treat the issue of estimation for

the GLIM with weighted exponential families. In Section 2.4, we consider the asymp-

totic properties of the estimates. In Section 2.5, we discuss issues such as residuals,

measures of goodness-of-fit and model diagnostics etc..

17

Page 25: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

2.1 Weighted Exponential Families

A family of distributions is called an exponential family if the probability density

function of the distributions in this family takes the form

f(y; θ, φ) = exp

yθ − b(θ)

a(φ)+ c(y, φ)

(2.1)

for some specific functions a(·), b(·) and c(·), where the support of the distribution

does not depend on the parameter θ. The parameter φ is called the dispersion pa-

rameter. Strictly speaking, the family is an exponential family in ordinary sense only

when φ is an known constant. However, by the convention with generalized linear

models, we still refer to the family as an exponential family when φ is an unknown

parameter.

Assume that the random variable Y in nature follows a distribution with probabil-

ity density function given by (2.1) and that the variable is ascertained with a weight

function w(y). Denote the ascertained variable by Y W . Then the probability density

function of Y W is given by

fW (y; θ, φ) =w(y)f(y; θ, φ)

ω(θ, φ), (2.2)

where ω(θ, φ) = E[w(Y )]. The distribution of Y W with probability density function

given by (2.2) is called a weighted exponential family. If, in particular, w(y) = y then

the weighted exponential family is called a length-biased exponential family.

In the following, we give some properties of the weighted exponential family.

Lemma 1 The weighted exponential family given by (2.2) is still an exponential fam-

ily with specific functions being given by

aW (φ) = a(φ),

bW (θ) = b(θ) + a(φ) lnω(θ, φ),

cW (y) = c(y, φ) + lnw(y).

Note that the function bW depends on φ as well. When φ is known, the weighted

exponential family is an exponential family in ordinary sense. If φ is unknown, it

18

Page 26: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

might not be an exponential family in ordinary sense. But, as a convention, we

still refer to the family as an exponential family. It can be easily obtained that the

cumulant generating function of an exponential family with functions a(φ) and b(θ)

is given by

κ(t) =1

a(φ)[b(θ + a(φ)t)− b(θ)].

Thus, the cumulants of the family are obtained as

κ1 = κ′(0) = b

′(θ),

κ2 = κ′′(0) = a(φ)b

′′(θ),

κm = κ(m)(0) = [a(φ)]m−1b(m)(θ), m ≥ 3.

Note that the first and second cumulants of a distribution are respectively the

mean µ and variance σ2 of the distribution. Thus we have µ = b′(θ) and σ2 =

a(φ)b′′(θ). That is, both µ and σ2 are functions of θ. Since σ2 > 0, b

′′(θ) > 0, which

implies that b′(θ) is an increasing function of θ. Thus we can express θ as a function of

µ which is the inverse of b′(·). Denote this function by θ = θ(µ). Let V (µ) = b

′′(θ(µ)),

which is what is called the variance function in generalized linear models.

By applying the above results to the weighted exponential family, we have

Lemma 2 (i) The cumulant generating function of the weighted exponential family

is given by

κW (t) =1

a(φ)[bW (θ + a(φ)t)− bW (θ)].

Hence, the cumulants are given by

κW1 = κ

′W (0) = b

′(θ) + a(φ)

d lnω(θ, φ)

= κ1 + a(φ)d lnω(θ, φ)

dθ,

κW2 = κ

′′W (0) = a(φ)b

′′(θ) + [a(φ)]2

d2 lnω(θ, φ)

dθ2

= κ2 + [a(φ)]2d2 lnω(θ, φ)

dθ2,

κWm = κ

(m)W (0) = [a(φ)]m−1b(m)(θ) + [a(φ)]m

dm lnω(θ, φ)

dθm

= κm + [a(φ)]mdm lnω(θ, φ)

dθm, m ≥ 3.

19

Page 27: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(ii) The mean of the weighted exponential family µW , as a function of θ, given by

µW = b′(θ) + a(φ)

d lnω(θ, φ)

dθ,

is increasing in θ.

We can see from the above results that a cumulant of the weighted exponential

family is the sum of the corresponding cumulant of the original exponential family

and an additional term. Now, let us consider the special case that w(y) = y, that is,

the case of length-biased exponential family. In this case, we have ω(θ, φ) = b′(θ).

Therefore, the cumulants of the length-biased exponential family are given by

κWm = κm + [a(φ)]m

dm ln b′(θ)

dθm, m ≥ 1.

Sinced ln b

′(θ)

dθ=

b′′(θ)

b′(θ)

andd2 ln b

′(θ)

dθ2=

b(3)(θ)

b′(θ)−

[b′′(θ)

b′(θ)

]2

,

we have, in particular,

κW1 = κ1 +

κ2

κ1

and

κW2 = κ2 +

κ3

κ1

−[κ2

κ1

]2

.

In general, the mth cumulant of the length-biased exponential family can be expressed

in terms of the cumulants of the original exponential family up to order m + 1.

The following lemma is trivial but useful.

Lemma 3 The mean µW of the weighted exponential family and the mean µ of the

original exponential family are one-to-one, in fact, µW is an increasing function of µ

given by

µW = µ + a(φ)∂ lnω(θ(µ), φ)

∂θ.

20

Page 28: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

From (ii) of Lemma 2, µW is an increasing function of θ. Since θ(µ) is the inverse of

an increasing function, θ(µ) is an increasing function of µ. Lemma 3 then follows.

The following lemma is due to Patil and Rao (1978).

Lemma 4 Let Y be a non-negative random variable. Let wi(y), i = 1, 2, be two

positive weight functions with finite expectations. Let Y Wi be the weighted version of

Y determined by weight function wi. Then E(Y W2) > E(Y W1) if r(y) = w2(y)/w1(y)

is increasing in y and E(Y W2) < E(Y W1) if r(y) is decreasing in y. In particular,

if a weight function w(y) is increasing [or decreasing] in y then E(Y w) > E(Y ) [or

E(Y w) < E(Y )].

2.2 The components of Generalized Linear Models

with Weighted Exponential Families

The components of a GLIM with weighted exponential family parallel to those of a

GLIM with ordinary exponential family. We first review briefly the components of

a GLIM with ordinary exponential family, and then describe the components of a

GLIM with weighted exponential family and their relations to their counterparts in

the corresponding GLIM with ordinary exponential family.

2.2.1 A brief review on GLIM with ordinary exponentialfamily

The GLIM with ordinary exponential family generalizes the classical normal linear

regression models in two aspects. First, the error distribution is generalized to any

exponential family. Second, the linear form is detached from the mean of the response

variable and, instead, is associated with a proper function of the mean. Let (yi, x(i)) :

i = 1, . . . , n, denote the observations of the response variable Y and the covariate

vector x on n individuals, where x(i) = (1, xi1, · · · , xip)T . For convenience, we have

included a constant component 1 in the covariate vector. A GLIM with ordinary

exponential families consists of three components: a random part (an assumption

on the distribution of the response variable), a deterministic part (an assumption

21

Page 29: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

on the role of the covariates) and a link function which connects the random and

deterministic parts together.

The random part. The yi’s are assumed to be independent and follow distributions

with probability density functions given by

f(yi; θi, φ) = expyiθi − b(θi)

a(φ)+ c(yi, φ). (2.3)

The deterministic part. The covariate variables enter into the model in a linear

form:

ηi = β0 + β1xi1 + · · · + βpxip = xT(i)β, (2.4)

where β = (β0, β1, · · · , βp)T . The linear form is called the linear predictor.

The link function. A monotone function g which relates the linear predictor ηi to

the mean µi = EYi as follows:

ηi = g(µi).

Since g is monotone, its inverse exists. Let the inverse of g be denoted by h. Then

the third component above can be replaced by

The response function. A monotone function h which relates the linear form ηi to

the mean µi = EYi as follows:

µi = h(ηi).

2.2.2 The GLIM with weighted exponential families

A GLIM with weighted exponential family is also specified by three components

similar to the GLIM with ordinary exponential families. Denote the observations for

a GLIM with weighted exponential family by (yWi , x(i)) : i = 1, . . . , n. The three

components are described as follows:

The random part. The Y Wi ’s are independent and follow distributions with prob-

ability density functions:

fW (yWi ; θi, φ) = expyW

i θi − bW (θi, φ)

a(φ)+ cW (yW

i , φ), (2.5)

22

Page 30: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where

bW (θ, φ) = b(θ) + a(φ) lnω(θ, φ)

cW (y) = c(y, φ) + lnw(y).

The deterministic part. This part remains the same as in the GLIM with ordinary

exponential families.

The response function. The response function hW , which relates the linear pre-

dictor ηi to µWi = EY W

i as

µWi = hW (ηi),

is determined by the response function h in the corresponding GLIM with ordinary

exponential family as follows:

hW (η) = h(η) + a(φ)∂ lnω(θ(h(η)), φ)

∂θ. (2.6)

In particular, in the special length-biased case,

hW (η) = h(η) + a(φ)V (h(η))

h(η),

where V (·) is the variance function.

Lemma 5 The response function hW (η) is monotone in η for any given weight func-

tion.

The monotonicity of hW (η) follows from Lemma 3 and the monotonicity of h(η).

2.3 Estimation of GLIM with Weighted Exponen-

tial Families

The data for a GLIM with weighted exponential family are as follows:

(yWi , x(i)) : i = 1, . . . , n,

23

Page 31: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where yWi is assumed to follow the weighted exponential family with density function

given by

fW (yWi ; θi, φ) = expyW

i θi − bW (θi, φ)

a(φ)+ cW (yW

i , φ),

and x(i) is assumed to affect the distribution of yWi through a linear predictor ηi =

xT(i)β, and ηi is connected to the mean of yW

i , µWi , by the response function hW as

µWi = hW (ηi).

The log likelihood function of (β, φ) based on the data is given by

l(β, φ) =n∑

i=1

[yW

i θi − bW (θi, φ)

a(φ)+ cW (yW

i , φ)

],

where θi is an implicit function of β determined by θi = θ(hW (ηi)). The parameters

(β, φ) are to be estimated by the method of maximum likelihood through the max-

imization of the log likelihood function. In this section, we develop algorithms for

the computation of the maximum likelihood estimates (MLE). We distinguish two

cases: (a) the dispersion parameter φ is known and (b) the dispersion parameter φ

is unknown. In the first case, we develop the algorithm of Newton-Rhapson with

Fisher scoring for the estimation of β. In the second case, we combine together the

Newton-Rhapson algorithm and a coordinator ascent algorithm and develop a double

iterative algorithm for the estimation of β and φ.

2.3.1 The iterative weighted least square procedure for the

estimation of β when φ is known.

In the case that φ is known, the MLE of β can be obtained by solving

∂l

∂β=

⎛⎜⎜⎜⎝

∂l∂β0∂l

∂β1...∂l

∂βp

⎞⎟⎟⎟⎠ = 0.

Let

A = −E

[∂2l

∂β∂βT

],

24

Page 32: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

which is the Fisher information matrix about β. Denote by ∂l

∂β(0)

and A(0), respec-

tively, the ∂l

∂βand A evaluated at β(0). The Newton-Rhapson algorithm with Fisher

scoring solves iteratively for β(1) in the following equation:

A(0)(β(1) − β(0)) =∂l

∂β(0)

. (2.7)

The procedure is essentially the same as that for the GLIM with ordinary exponential

families and is equivalent to an iterative weighted least square (IWLS) procedure. See

McCullagh and Nelder (1989, Section 2.5). The derivation of the IWLS procedure is

sketched as follows.

By chain’s rule, we have

∂l(β, φ)

∂β=

1

a(φ)

n∑i=1

[Y W

i − ∂bW (θi, φ)

∂θi

]∂θi

∂β, (2.8)

where∂θi

∂β=

∂θi

∂µWi

∂µWi

∂ηi

∂ηi

∂β=

[∂2bW (θi, φ)

∂θ2i

]−1∂hW (ηi)

∂ηi

x(i).

Let

w∗i =

[∂hW (ηi)

∂ηi

]2 [∂2bW (θi, φ)

∂θ2

]−1

.

Thus we can write

∂l(β, φ)

∂β=

1

a(φ)

n∑i=1

w∗i x(i)(Y

Wi − µW

i )∂ηi

∂µWi

.

From (2.8), we obtain

∂2l(β, φ)

∂β∂βT=

1

a(φ)

n∑i=1

[Y W

i − ∂bW (θi, φ)

∂θi

]∂2θi

∂β∂βT− ∂2bW (θi, φ)

∂θ2i

∂θi

∂β

∂θi

∂βT

.

Hence

A = −E

[∂2l(β, φ)

∂β∂βT

]

=1

a(φ)

n∑i=1

∂2bW (θi, φ)

∂θ2i

∂θi

∂β

∂θi

∂βT

=1

a(φ)

n∑i=1

x(i)

[∂hW (ηi)

∂ηi

]2 [∂2bW (θi, φ)

∂θ2i

]−1

xT(i)

=1

a(φ)

n∑i=1

w∗i x(i)x

T(i).

25

Page 33: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

For the sake of convenience, we introduce the following notations:

yW = (yw1 , . . . , yW

n )T ,

µW = (µW1 , . . . , µW

n )T ,

η = (η1, . . . , ηn)T ,

W ∗ = diag(w∗1, . . . , w

∗n),

X = (x(1), · · · , x(n))T ,

∂η

∂µWT= diag(

∂η1

∂µW1

, . . . ,∂ηn

∂µWn

).

With the above notations, the∂l(β,φ)

∂βand A can be expressed concisely as follows:

∂l(β, φ)

∂β=

1

a(φ)XTW ∗ ∂η

∂µWT(yW − µW ),

A =1

a(φ)XTW ∗X.

The iterative equation (2.7) of the Newton-Rhapson algorithm with Fisher scoring

can now be written as

XT W ∗(0)X(β(1) − β(0)) = XT W ∗

(0)

∂η

∂µWT(0)

(yW − µW(0)),

where a quantity with a subscript “(0)” indicates that the quantity is evaluated at

β(0). Note that Xβ(0) = η(0). Denote

z = η +∂η

∂µWT(yW − µW ),

which is referred to as the pseudo-response vector. The iterative equation can be

finally written as

XT W ∗(0)Xβ(1) = XT W ∗

(0)z(0),

which is the normal equation of a weighted least square problem with response vector

z(0), design matrix X and weight matrix W ∗(0). It needs to be noted that both W ∗

(0)

and z(0) involve φ, though it does not explicitly appear in the above equation.

The IWLS algorithm for the generalized linear model with weighted exponential

family in the case of known φ is now described as follows.

26

Page 34: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Algorithm 1 Starting with an initial β(0), for k = 0, 1, 2, . . . , do

Step 1. Compute current estimates η(k), µW(k) and W ∗

(k) of η, µW and W ∗, respec-

tively, from β(k) and form the current pseudo-response vector z(k) as follows:

z(k) = η(k) +∂η

∂µWT

∣∣∣∣µW =µW

(k)

(yW −µW(k)).

Step 2. Regress z(k) on X with weight matrix W ∗(k) to obtain a new estimate β(k+1),

i.e., solve for β(k+1) in the following equation:

XT W ∗(k)Xβ(k+1) = XT W ∗

(k)z(k).

The above two steps are repeated until convergence occurs.

Formulae for the computation of η, µW and W ∗. The components of η, µW and

W ∗ needed in the above algorithm are computed as follows:

ηi = β0 + β1xi1 + · · · + βpxip,

µWi = h(ηi) + a(φ)

∂ lnω(θi, φ)

∂θi,

w∗i =

[∂hW (ηi)

∂ηi

]2

/∂2bW (θi, φ)

∂θ2i

=

(∂h(ηi)

∂ηi

)2 (∂2b(θi)

∂θ2i

)−1[1 + a(φ)

(∂2b(θi)

∂θ2i

)−1∂2 log ω(θi, φ)

∂θ2i

],

since

∂hW (ηi)

∂ηi=

∂2bW (θi)

∂θ2i

∂h(ηi)

∂ηi/∂2b(θi)

∂θ2i

,

∂2bW (θi)

∂θ2i

=

[∂2b(θi)

∂θ2i

+ a(φ)∂2 log ω(θi, φ)

∂θ2i

],

where

θi = θ(h(ηi)).

Initial β. The initial value β(0) can be obtained as follows. Let gW be the inverse

function of hW (recall that hW is monotone by Lemma 5). Let

z0i = gW (yWi ), i = 1, . . . , n.

27

Page 35: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Regress z0 = (z01, . . . , z0n)T on X. The resultant least square estimate of the regres-

sion coefficients can be taken as the initial β(0).

Remark I. In the special case of length biased exponential family, we have

ω(θ, φ) = b′(θ),

bW (θ) = b(θ) + a(φ) log b′(θ),

hW (η) = h(η) + a(φ)b′′(θ(h(η)))

h(η).

Therefore, the computational formulae for the quantities µW and w∗ become the

following:

µWi = h(ηi) + a(φ)

d ln b′(θi)

dθi

= h(ηi) + a(φ)b′′(θi)

b′(θi)

= µi + a(φ)b′′(θ(µi))

µi.

w∗i =

[h′(ηi)]

2

b′′(θi)

[1 + a(φ)

µib(3)(θi) − [b

′′(θi)]

2

µ2i b

′′(θi)

].

Remark II. In general, since

ω(θ, φ) =

∫w(y) expyθ − b(θ)

a(φ)+ c(φ, y)dy,

the derivatives of ω can be computed by exchanging the order of the differentiation

and integration, that is,

∂ω(θ, φ)

∂θ=

∫w(y)

∂θexpyθ − b(θ)

a(φ)+ c(φ, y)dy,

∂2ω(θ, φ)

∂θ2=

∫w(y)

∂2

∂θ2expyθ − b(θ)

a(φ)+ c(φ, y)dy.

Let

ω1(θ, φ) =

∫yw(y) expyθ − b(θ)

a(φ)+ c(φ, y)dy,

ω2(θ, φ) =

∫y2w(y) expyθ − b(θ)

a(φ)+ c(φ, y)dy.

28

Page 36: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Then, we have

∂ω(θ, φ)

∂θ= ω1(θ, φ) − b

′(θ)ω(θ, φ),

∂2ω(θ, φ)

∂θ2= [b

′2(θ) − b′′(θ)]ω(θ, φ) − 2b

′(θ)ω1(θ, φ) + ω2(θ, φ).

2.3.2 The double iterative procedure for the estimation ofboth β and φ when φ is unknown.

There is a major difference between the GLIM with weighted exponential families and

the GLIM with ordinary exponential families. With the ordinary exponential families,

the estimation of β in the linear predictor does not involve the dispersion parameter

φ, the estimate of β remains the same no matter φ is known or not. However, with

the weighted exponential families, the estimate of β depends on φ, the estimation of

β and φ can not be separated.

We describe in this subsection a double iterative procedure for the simultaneous

estimation of β and φ. The procedure is a combination of the IWLS procedure

described in Section 2.3.1 and the so-called coordinate ascent procedure. The double

iterative procedure alternates between a β-step — the maximizion of l(β|φ), the

log likelihood function given φ, with respect to β and a φ-step — the maximizion of

l(φ|β), the log likelihood function given β, with respect to φ. In a β-step, Algorithm 1

is implemented with the given φ. In a φ-step, a bisection search procedure is utilized

to search for the maximum of l(φ|β) with the given β. The double iterative procedure

is described in the following algorithm.

Algorithm 2 Starting with an initial value φ(0), for k = 0, 1, 2, . . . , do

β-step: Maximize l(β|φ(k)) with respect to β by Algorithm 1 and obtain the max-

imizer β(k) = β(φ(k)).

φ-step: Maximize l(φ|β(k)) with respect to φ.

Alternate between the above β-step and φ-step until convergence occurs.

29

Page 37: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

2.4 Asymptotic Distribution of β

In order to make inference on β such as constructing confidence intervals or conduct-

ing hypothesis testing on the components of β, we need to know the distribution or

the asymptotic distribution of β. Since it is usually not possible to obtain the exact

distribution of β, we consider in this section the asymptotic distribution of β. The

general asymptotic theory of MLE applies to the generalized linear models with either

ordinary or weighted exponential families, that is, the asymptotic distribution of β

can be approximated by a normal distribution. Specifically, if the sample size is large

then, approximately,

β ∼ N(β, Σ ˆβ),

where Σ ˆβis the asymptotic variance-covariance matrix of β. However, the form of

the asymptotic variance-covariance matrix in the GLIM with weighted exponential

families differs from that in the GLIM with ordinary exponential families when the

dispersion parameter is unknown, because the dispersion parameter φ is involved in

different ways in ordinary and weighted exponential families. This difference arises

from the fact that, for ordinary exponential families, the MLE of φ and the MLE

of β are asymptotically independent, however, this asymptotic independence is not

retained for weighted exponential families. In what follows, we elaborate on these

matters and derive the asymptotic variance-covariance matrix Σ ˆβfor the GLIM with

weighted exponential families.

2.4.1 The asymptotic variance-covariance matrix of β in the

case of known φ

According to the general theory on MLE, the asymptotic variance-covariance matrix

of the MLE of the parameter vector of a distribution family is given by the inverse of

the Fisher information matrix of the parameter vector.

When φ is known, the GLIM with weighted exponential family is parameterized

30

Page 38: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

by β only. The Fisher information matrix of β is given by

IββT = −E

[∂2l(β|yW )

∂β∂βT

],

where l(β|yW ) is the log likelihood function of the GLIM with weighted exponential

family. It was derived in Section 2.3.1 that

Iββ

T =1

a(φ)XT W ∗X.

Therefore, the asymptotic variance-covariance matrix of β is given by

Σ ˆβ= a(φ)(XTW ∗X)−1.

The form of Σ ˆβin the case of known φ is the same as that for the GLIM with ordinary

exponential families except that the contents of W ∗ are different.

2.4.2 The asymptotic variance-covariance matrix of β in the

case of unknown φ

In the case of unknown φ, in addition to IββT , we introduce the following notation:

Iφφ = −E∂2l(β, φ|y)

∂φ2,

Iβφ= −E

∂2l(β, φ|y)

∂β∂φ

IφβT = IT

βφ

where l(β, φ|y) is the log likelihood function of β and φ based on the observation

vector y. The Fisher information matrix of (β, φ) is given by(IββT Iβφ

Iφβ

T Iφφ

).

The joint asymptotic variance-covariance matrix of the MLE (β, φ) is given by the

inverse of the above matrix.

31

Page 39: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

For the GLIM with an ordinary exponential family, the log likelihood function of

β and φ is given by

l(β, φ|y) =

n∑i=1

[yiθi − b(θi)

a(φ)+ c(yi, φ)

].

Hence∂2l(β, φ|y)

∂β∂φ=

n∑i=1

[yi − b′(θi)]

∂θi

∂β

∂[a(φ)]−1

∂φ.

It is easy to see that

Iβφ= −E

∂2l(β, φ|y)

∂β∂φ= 0.

The variance-covariance matrix of the MLE (β, φ) is the inverse of the Fisher infor-

mation matrix of (β, φ) and is given by

(IββT 0

0 Iφφ

)−1

=

(I−1

ββT 0

0 I−1φφ

).

This implies that the MLE of β and the MLE of φ are asymptotically independent

and that the asymptotic variance-covariance matrix of β is given by

Σ ˆβ= I−1

ββT = a(φ)(XTW ∗X)−1,

no matter whether the dispersion parameter φ is known or not.

For the GLIM with a weighted exponential family, the log likelihood function of

β and φ is given by

l(β, φ|yW ) =n∑

i=1

[yW

i θi − bW (θi, φ)

a(φ)+ cW (yW

i , φ)

].

It is straightforward to verify that

Iβφ=

1

a(φ)

n∑i=1

∂2bW (θi, φ)

∂θi∂φ

∂θi

∂β. (2.9)

In general, Iβφdoes not equal to zero. Therefore, the MLE of β and the MLE of φ

are not asymptotically independent. The form of the asymptotic variance-covariance

32

Page 40: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

matrix Σ ˆβis different from that when the dispersion parameter φ is known. By some

matrix algebra, we can obtain

Σ ˆβ= I−1

ββT +

I−1

ββT IβφIφβT I−1

ββT

Iφφ − IφβT I−1

ββT Iβφ

.

In the following, we derive the formulae for Iβφand Iφφ. We have obtained the

expression of ∂θi/∂β as

∂θi

∂β=

[∂2bW (θi, φ)

∂θ2i

]−1∂hW (ηi)

∂ηixT

(i).

Substituting the expression of ∂θi/∂β into (2.9), we obtain

Iβφ=

1

a(φ)

n∑i=1

∂2bW (θi, φ)

∂θi∂φ

[∂2bW (θi, φ)

∂θ2i

]−1∂hW (ηi)

∂ηi

xT(i)

=1

a(φ)

n∑i=1

vi

√wi∗xT

(i)

=1

a(φ)XTW ∗1/2v,

where v = (v1, · · · , vn)T with

vi =∂2bW (θi, φ)

∂θi∂φ

[∂bW (θi, φ)

∂θ2i

]−1/2

.

The quantity Iφφ can be obtained as

Iφφ = Var

(∂

∂φ

n∑i=1

yWi θi − bW (θi, φ)

a(φ)+ c(yW

i , φ)

).

If c(yWi , φ) is free of φ, then Iφφ can be obtained as

Iφφ = Var

(−a

′(φ)

a2(φ)

n∑i=1

θiyWi

)

=[a

′(φ)]2

a3(φ)

n∑i=1

θ2i

∂2bW (θi, φ)

∂θ2i

.

We now summarize the results found in the above two subsections as follows.

33

Page 41: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Theorem 13 The MLE β of the GLIM with a weighted exponential family follows

asymptotically a normal distribution with mean vector β and variance-covariance

matrix Σ ˆβ.

(i) If the dispersion parameter φ is known then

Σ ˆβ= a(φ)(XTW ∗X)−1.

(ii) If the dispersion parameter φ is unknown then

Σ ˆβ= a(φ)

[(XT W ∗X)−1 +

(XT W ∗X)−1XT W ∗1/2vvT W ∗1/2X(XT W ∗X)−1

a(φ)Iφφ − vT W ∗1/2X(XT W ∗X)−1XTW ∗1/2v

],

where v is defined in the previous page.

2.5 Deviance and Residuals

In this section, we discuss some issues on inference with weighted models. In §2.4.1,we consider the deviance measure and its application in model comparison in both

the generalized weighted linear and generalized weighted additive models. In §2.4.2,we discuss the residuals and their use for model diagnostics.

2.5.1 Deviance Analysis

The deviance was introduced by McCullagh and Nelder (1983) as a goodness-of-

fit measure in generalized linear models. It is defined as two times the log of a

likelihood ratio. Here, we give a similar definition of deviance for generalized linear

and additive models with weighted exponential families. Let θ = (θ1, . . . , θn) be the

MLE of θ = (θ1, . . . , θn) under the model of concern. Let θ = (θ1, . . . , θn)T denote

the vector θ corresponding to µW = y, i.e., µWi = yi, i = 1, . . . , n. In the case of

known dispersion parameter φ, the deviance of the model is then defined as

D(y; θ) =1

a(φ)2

n∑i=1

yi(θi − θi) − [bW (θi, φ) − bW (θi, φ)].

When φ is unknown, φ is replaced by its MLE φ in the definition.

34

Page 42: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Let MF and MR be two models such that MR is nested under MF . Denote the

MLE of θ under these two models by θF and θR. Consider the difference of the

deviances of these two models Tn = D(y; θR)−D(y; θF ). If the dispersion parameter

φ is a known constant, the difference Tn is indeed the likelihood ratio test statistic

for testing MR against MF . However, if the dispersion parameter is unknown and

replaced by a consistent estimate, Tn is no longer the likelihood ratio test statistic,

noticing that the estimate of θ depends on the estimate of φ in each model.

In the context of generalized weighted linear models, the difference of deviances

can be used as an inference tool. If the dispersion parameter is a known constant,

it follows from the general theory of likelihood ratio test that, under the assumption

of model MR, the difference Tn has asymptotically a χ2 distribution with degrees of

freedom given by the difference of the numbers of parameters in the two models. This

result can be used to make inferences on various aspects of the model by identifying

appropriate models as MF and MR. A number of examples follow. (a) In the selection

of predictor variables, MF is identified as the model with a larger set of predictor

variables and MR as the model containing a subset of the predictor variables of the

model MF . (b) In the assessment of the adequacy of link function, the link function

currently adopted in the model can be embedded in a family of link functions indexed

by an additional parameter, MF is identified as the model with the link function

containing the additional parameter and MR as the model with the link function

specified by a certain fixed value of the additional parameter. (c) In the checking

of the validity of weight function, the same technique as for the assessment of the

adequacy of link function can be applied, MF is identified as the model with the

weight function specified by a family up to an additional parameter and MR as the

model with weight function which is a special member of the family.

If the dispersion parameter is unknown, it is not clear whether or not Tn still

follows an asymptotic χ2 distribution. The distributional property of Tn is yet to be

investigated.

In the context of generalized weighted additive models, the difference of deviances

is no longer asymptotically χ2 distributed no matter φ is known or not. However,

35

Page 43: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

the deviances can still be used as an informal tool for model assessments. Tn can be

informally considered as a χ2 statistic with degrees of freedom appropriately defined.

Hastie and Tibshirani (1990, §6.8) had some discussion on this aspect when only

the additive predictor is of concern. In the assessment of link function or weight

function, the matter becomes simpler. In those cases, under both MF and MR, the

additive predictors are the same and hence the degrees of freedom associated with

the additive predictors are the same no matter how they are defined, the difference of

the degrees of freedom between the two models only involves the parameters in the

family of link functions or weight functions. Usually, the family of link function or

weight function to be considered is indexed by one parameter. Hence, the degrees of

freedom associated with Tn is 1 when Tn is the difference of deviances of MF and MR

identified in the case of link function or weight function assessment.

2.5.2 Residuals and Model Diagnostics

A statistical model provides, at the best, an approximation to the probabilistic mecha-

nism underlying the data. The validity of such approximation must always be checked.

When the validity is in doubt, specific departures from an adequate approximation

need to be figured out and remedial measures be provided. To this end, diagnostic

tools and methods must be developed. In this section, we discuss issues on the di-

agnostics of the weighted models. First, we define various residuals similar to those

defined in ordinary generalized linear models. Then we discuss the techniques for

checking various departures from the assumed model.

Residuals

We will consider three types of residuals: the deviance residual, the Pearson residual

and the working residual which are defined as follows.

Deviance residual:

rDi = sign(yi − µWi )

√2yi(θi − θi) − [bW (θi, φ) − bW (θi, φ)]1/2.

36

Page 44: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Pearson residual:

rPi =yi − µW (θi, φ)√

V (θi, φ),

where V (θ is the variance of yWi .

Working residual: Let z and W be the pseudo-response vector and the weight matrix

in the last step of the computational algorithms. The working residual is defined

as the residual after z is fitted to X with weight matrix W .

Checking weight function

In principle, the techniques used to check departures in generalized linear models

can be used in the same way in checking departures in generalized weighted models.

We will briefly discuss those techniques in the next subsection. However the weight

function, which is the only additional component in weighted models, needs additional

treatment. The adequacy of the weight function can be checked either by informal

ways through the residuals or, as mentioned in §2.5.1, by a formal likelihood ratio

test. In the following, we develop an informal technique for checking the adequacy of

the weight function.

Consider the Pearson’s residuals. Suppose that the true weight function, which

is unknown, is w0(y). Let w1(y) be the weight function used in the model. Denote

by ω0(θ, φ) and ω0(θ, φ), respectively, the expectations of w0(Y ) and w2(Y ) under

the original exponential family. Let θi and φ denote the estimates of θi and φ. For

convenience, we suppress the denominator in the Pearson’s residual and consider

yi − µW (θi, φ). If w1(y) is the same as w0(y) then the residual should have roughly

expectation zero. Otherwise, we have

yi − µW1 (θi, φ) = [yi − µW

0 (θi, φ)] + [µW0 (θi, φ) − µW

1 (θi, φ)],

where µW0 and µW

1 are the means of the weighted distributions with weight functions

w0 and w1 respectively. Thus, the expectation of the residual will not be zero, in

stead, will be a positive or negative number µW0 (θi, φ) − µW

1 (θi, φ) (depending on

whether or not µW0 is bigger than µW

1 ). The sign of the difference is determined by

37

Page 45: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

the relationship between the weight functions w0 and w1. It follows from result (iii)

given in §2.1 that if w0(y)/w1(y) is increasing in y the difference will be positive,

otherwise, the difference will be negative. Now, let np denote the number of positive

Pearson’s residuals. If the weight function is adequate, i.e., w1(y) = w0(y), then np

will roughly follow a binomial distribution with index n and probability of success

π = 1/2. The expected np will be n/2. If np is too large (or too small), it indicates

that the weight function w1(y) increase with y too slowly (or too fast), supposing

that the weight function is increasing in y, and suggests that a weight function that

increase with y faster (or slower) should be used. For example, if the weight function

used in the model is of the form w1(y) = yα. If np is too large, it suggests that a larger

α value should be adopted for the weight function. If np is too small, it suggests that

a smaller α value should be adopted for the weight function. To decide whether np

is significantly too large or too small, a one-sided test for testing the null hypothesis

H0 : π = 1/2 against the alternative H1 : π > 1/2 or π < 1/2, could be carried out

assuming np follows the binomial distribution B(n, π).

Checking other aspects of the model

Other aspects of the weighted models such as the link function, the variance function

(essentially the exponential family) and the predictor can be checked in the same way

as in the generalized linear models. McCullagh and Nelder (1989, Chapter 12) has an

extensive discussion on informal graphical methods and formal testing methods for

the checking of generalized linear models. Here, we only briefly discuss the following

residual plots:

(i) the plot of deviance residuals against the estimated predictor for check-

ing whether or not systematic departure from the assumed model is

present;

(ii) the plot of the absolute residuals against fitted values for checking the

adequacy of the assumed variance function;

(iii) the plot of adjusted dependent variable against the estimated predic-

tor for checking the adequacy of the link function;

38

Page 46: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(iv) the partial residual plot for checking the scales of covariates in the

case of generalized weighted linear models.

When these plots are used for checking the weighted models, we should bear in

mind the following points: (1) If some non-null pattern shows up in plot (i), in

addition to possible departures from the assumed link function, variance function

and scales of covariates, a departure from the assumed weight function might also

be the cause of the non-null pattern; (2) When a plot reveals the inadequacy of the

variance function, it is, in the current context, an indication that the exponential

family currently considered might not be appropriate.

39

Page 47: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 3

Generalized Additive Models withWeighted Exponential Families

The generalized linear models have been further extended to a larger class of statistical

models called generalized additive models (GAM). In GAM, the linear predictor of

a model is replaced by an additive function of the covariate variables which is not

in a parametric form. A detailed account of GAM is given by Hastie and Tibshirani

(1990). In this chapter, we extend the GAM to the case when the distribution of the

response variable follows a weighted exponential family. Similar to the generalized

linear model case, we refer to the original GAM as the GAM with ordinary exponential

family and the extended class as the GAM with weighted exponential family.

3.1 Specification of a Generalized Additive Model

with a Weighted Exponential Family

3.1.1 General Assumptions

A generalized additive model differs from a generalized linear model by only one

component, i.e., the predictor function. For a generalized linear model, the predictor

function is linear. However, for a generalized additive model, the predictor function

is an additive function of the covariates Xj’s, not necessarily linear.

The relationship between a GAM with weighted exponential family and the GAM

with the original exponential family is exactly the same as that between a GLIM

40

Page 48: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

with weighted exponential family and the GLIM with the original exponential family.

Therefore, we can obtain a GAM with weighted exponential family in the same way

we arrive at a GLIM with weighted exponential family from a GLIM with ordinary

exponential family.

A GAM with weighted exponential family for the data (Y Wi , xi) : i = 1, . . . , n can

be specified by the following components:

• The weighted exponential family (2.5), i.e., the Y Wi ’s are assumed to be inde-

pendent and follow distributions with probability density function (2.5).

• The additive predictor function, i.e., a function of the covariates given by the

form

ηi = α + f1(x1i) + f2(x2i) + · · · + fp(xpi), (3.1)

where fj, j = 1, . . . , p, are arbitrary functions subject to∫

fj(xj)dxj = 0,

j = 1, . . . , p, and certain other constraints. The restriction∫

fj(xj)dxj = 0 is

required for the model to be identifiable by the additive predictor.

• The weighted response function (2.6) which relates the mean of Y Wi to the

additive predictor.

In fact, a GAM with weighted exponential family differs from a GLIM with weighted

exponential family only by the predictor function. The former can be obtained from

the latter by simply replacing the linear predictor with the additive predictor.

3.1.2 Modeling of the Additive Predictor

Though the functions fj in the additive predictor are not specified by any parametric

forms, they can not be completely arbitrary. They are usually specified by function

spaces with certain properties. That is, the additive predictor is modeled as

41

Page 49: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

η(x) = α + f1(x1) + · · · + fp(xp),

fj ∈ Hj, j = 1, . . . , p, (3.2)

where Hj’s are spaces of univariate functions with certain properties.

There are two major methodologies to specify the function spaces Hj in the lit-

erature. One is through the use of B-splines and the other is through the use of

reproducing kernel Hilbert spaces. These correspond to two major smoothing tech-

niques: regression spline method and smoothing spline method. In the following, we

elaborate on these methods.

Regression Spline Method. The regression spline method is comprehensively in-

vestigated in Stone et al. (1997). In the regression spline method, Hj is specified

by a space spanned by a basis of B-splines. The knots and the number of the basis

B-splines are determined by the data. The details on B-splines can be found in De

Boor (1978). Suppose that the B-spline basis for Hj is determined as Bj1, . . . , Bjmj,j = 1, . . . , p. Then fj can be represented by a linear combination of these basis B-

splines as

fj(xj) =

mj∑k=1

cjkBjk(xj).

Therefore, the additive predictor ηi can be expressed as

ηi = α +

p∑j=1

mj∑k=1

cjkBjk(xij), i = 1, . . . , n.

Thus the additive predictor can be treated as if it is a linear predictor.

Smoothing Spline Method. An extensive account of smoothing spline method can

be found in Wahba (1990), Green and Silverman (1994) and Gu (2002). In this

method, each of the Hj’s is specified as a reproducing kernel Hilbert space of uni-

variate functions. The general theory of reproducing kernel Hilbert spaces and their

applications in statistics can be found in Weinert (1982). In the following, we briefly

discuss some major properties of reproducing kernel Hilbert spaces which are rele-

vant to our development of GAM with weighted exponential families. A reproducing

42

Page 50: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

kernel Hilbert space H is a Hilbert space of functions defined on a set, say, T , such

that any evaluation functional x : f −→ f(x), x ∈ T , f ∈ H, is bounded under the

norm induced by the inner product of the Hilbert space. For a reproducing kernel

Hilbert space, there exists a unique function Q(x, y) defined on T × T that has the

following properties:

(a) for any fixed x ∈ T , Q(x, ·) belongs to H;

(b) for any x ∈ T and f ∈ H, 〈Q(x, ·), f〉 = f(x) where 〈·, ·〉 is the inner product

on the Hilbert space;

(c) Q(x, y) = Q(y, x) where z denotes the complex conjugate of z;

(d) Q(x, y) is positive definite in the sense that∑n

i=1

∑nj=1 aiajQ(xi, xj) ≥ 0 for

any n and any x = (x1, . . . , xn)T , a = (a1, . . . , an)

T . If x is chosen such that

xi = xj, i = j then∑n

i=1

∑nj=1 aiajQ(xi, xj) = 0 if and only if a = 0.

The function Q(x, y) is called the reproducing kernel of H and property (b) is referred

to as the reproducing property.

The so called Sobelev space W(m)2 [0, 1] of functions defined on [0, 1] is an example

of reproducing kernel Hilbert space. The space W(m)2 [0, 1] is defined as

W(m)2 [0, 1] = f |f, f

′, . . . , f (m−1) abs.cont., f (m) ∈ L2

where abs.cont. stands for “absolutely continuous” and L2 is the space of square

integrable functions. The inner product on W(m)2 [0, 1] is given by

〈f, g〉 =m−1∑r=0

∫ 1

0

f (r)(x)dx

∫ 1

0

g(r)(x)dx +

∫ 1

0

f (m)(x)g(m)(x)dx,

and the corresponding reproducing kernel is given by

R(x, y) = 1 +m−1∑r=1

kr(x)kr(y) + km(x)km(y) − k2m(|x− y|),

43

Page 51: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where kr(x) = Br(x)/r!, Br(x) being the Bernoulli polynomials on x ∈ [0, 1] defined

by letting B0(x) = 1, 1r+1

ddx

Br+1(x) = Br(x) and choosing the constant of integration

so that∫ 1

0Br(x)dx = 0 for r ≥ 1. In particular,

k1(x) = x − 1

2,

k2(x) =1

2(x2 − x +

1

6),

k3(x) =1

6(x3 − 3x2

2+

x

2),

k4(x) =1

24(x4 − 2x3 + x2 − 1

30).

For details of this space and other useful reproducing kernel Hilbert spaces, see, e.g.,

Gu (2002).

If in a problem the domain of the functions of concern is a general interval y ∈[a, b] in stead of the standard unit interval [0, 1], a simple linear transformation

y −→ x = (y−a)/b−a) will transfer the problem into the standard problem. Without

loss of generality, we assume the range of the covariates xj’s in the models we consider

is the unit interval hereafter.

To model the components fj in the additive predictor, we can specify Hj as

Hj = W(mj)2 [0, 1]/1,

where 1 is the space spanned by the constant function 1. The Hj so defined is again

a reproducing kernel Hilbert space, and the inner product and the corresponding

reproducing kernel of Hj are given, respectively, by

〈f, g〉j =

mj−1∑r=1

∫ 1

0

f (r)(xj)dxj

∫ 1

0

g(r)(xj)dxj +

∫ 1

0

f (mj )(xj)g(mj)(xj)dxj,

Rj(x, y) =

mj−1∑r=1

kr(x)kr(y) + kmj (x)kmj (y)− k2mj (|x− y|).

The space Hj can be decomposed as

Hj = Hj0 ⊕Hj1,

44

Page 52: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where

Hj0 = Spank1, . . . , kmj−1,

Hj1 = f |f ∈ Hj,

∫ 1

0

f (r)(x)dx = 0, r = 1, . . . , mj − 1.

Both Hj0 and Hj1 are still reproducing kernel Hilbert spaces. In particular, the inner

product and the reproducing kernel of Hj1 are given, respectively, by

〈f, g〉j1 =

∫ 1

0

f (mj)(xj)g(mj)(xj)dxj,

Rj1(x, y) = kmj (x)kmj (y) − k2mj (|x − y|).

Now, we decompose the additive predictor η into

η = η0 + η1,

where

η0 = α +m1−1∑r=1

β1rkr(x1) + · · · +mp−1∑r=1

βprkr(xp),

η1 = g1(x1) + · · · + gp(xp), gj ∈ Gj(= Hj1).

Note that η0 is linear in terms of the unknown parameters. When η1 = 0, the additive

predictor reduces to a linear predictor. By modeling the additive predictor in the way

above, not only we can see explicitly that the additive predictor is an extension of

the linear predictor, but also we have separated the linear part and the other part

in an additive predictor. The separation of the linear part and other part has an

important implication in model building. It provides us with a structure for testing

a linear model against a general non-linear additive model, at least, in principle.

In this thesis, we will concentrate on the reproducing kernel Hilbert space approach

for modeling the additive predictor because of its obvious advantages.

3.2 Estimation of GAM with Weighted Exponen-

tial Families

In this section, we deal with the problem of estimation for GAM with weighted

exponential families. We focus on the method of penalized maximum likelihood es-

45

Page 53: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

timation (PMLE). In Section 3.2.1, we introduce a general backfitting procedure. In

Section 3.2.2, the method of PMLE is described and is shown to be equivalent to

an iterated penalized weighted least square (IPWLS) procedure. In Section 3.2.3,

a backfitting framework for the IPWLS is considered. In Section 3.2.4, the back-

fitting framework is used to derive a simplified formulation for the IPWLS proce-

dure. In Section 3.2.5, algorithms with fixed smoothing parameters are described.

In Section 3.2.6, the issue on the choice of smoothing parameter is discussed. In

Section 3.2.7, algorithms with the choice of smoothing parameters incorporated are

described.

3.2.1 Backfitting Procedure

The backfitting procedure is widely used in the literature of additive models, see

Hastie and Tibshirani (1990). In this section, we briefly discuss the backfitting pro-

cedure.

Let y = (y1, . . . , yn)T be the vector of observed response variable on n subjects.

Let H = f = (f(x1), . . . , f(xn))T |f ∈ H where x1, . . . , xn are the values of the

covariate x on the n subjects. A smoother is defined as an operator, denoted by S,

that maps y to Sy ∈ H . The smoother S, which is determined by the method of

smoothing, does not depend on y except it depends on certain smoothing parameters

whose choice might be affected by y. In the context of additive models, let Sj be

the univariate smoother that maps an response vector to the jth function space

Hj = f j = (fj(xj1), . . . , fj(xjn))T |fj ∈ Hj.

The general backfitting algorithm for fitting the additive function

f(x) = f1(x1) + · · ·+ fp(xp) is an iterative procedure that fits the components alter-

natively as follows:

f 1 = S1(y −∑

j =1 f j),

f 2 = S2(y −∑

j =2 f j),

· · ·f p = Sp(y −

∑j =p f j).

(3.3)

At each step, the components appearing on the right hand side of the equations are

46

Page 54: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

replaced by the most updated values from the previous steps. The advantage of the

backfitting algorithm is that it reduces a multivariate smoothing problem to univariate

smoothing problems. Usually, the multivariate smoothing problem involves a linear

system of order pn× pn but a univariate smoothing problem involves a linear system

of order only n × n.

The following lemma is due to Breiman and Freidman (1985):

Lemma 6 Suppose that the smoothers Sj’s are non-negative definite matrices with

eigenvalues less than 1. Let Bj = Sj(I−Sj)−1 and B =

∑pj=1 Bj . Then the equations

(3.3) have the unique solution given by

f j = Bj(I + B)−1y, j = 1, . . . , p.

3.2.2 Penalized Maximum Likelihood Estimation and Iter-

ated Penalized Weighted Least Square procedure

We consider the penalized maximum likelihood estimation for the GAM with weighted

exponential families in this section. For the sake of convenience, we denote the

known functions, kr(xj)’s, in the component η0 of the additive predictor by hr(x), r =

1, . . . , d, where d =∑p

j=1 mj − p, let

g0(x) = α +

d∑r=1

βrhr(x),

and denote Rj1 by Qj.

Let η = (η1, . . . , ηn) where

ηi = α +d∑

r=1

βrhr(xi) +

p∑j=1

gj(xij).

The log likelihood function of the GAM with weighted exponential family specified

in Section 3.1 is given by

l(η, φ) =n∑

i=1

[yW

i θ(ηi) − bW (θ(ηi), φ)

a(φ)+ cW (yW

i , φ)

],

47

Page 55: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where the function θi = θ(ηi) is determined through the following relation:

∂bW (θi, φ)

∂θi= hW (ηi).

The functions bW , hW and cW are the same as in Chapter 2.

The penalized maximum likelihood method fits the GAM with weighted exponen-

tial family by maximizing the following penalized likelihood:

l(η, φ)− n

2

p∑j=1

λj

∫[g

(mj)j (xj)]

2dxj, (3.4)

subject to gj ∈ Gj, j = 1, . . . , p,

where λj ’s are smoothing parameters to be determined.

The penalized maximum likelihood method arises from the fact that the estimates

of the functions gj, j = 0, 1, . . . , p, cannot be obtained by maximizing the log likeli-

hood function l(η, φ) without any constraint on gj, j = 1, . . . , p, since ηi’s can then

take any values and the likelihood will be maximized by making hW (ηi) = yWi . The

natural constraint on gj is to require that the norm of gj is bounded by a given

constant, i.e., ||g(mj)j ||2 =

∫ 1

0[g

(mj)j (xj)]

2dxj ≤ Dj for some Dj . The conditional max-

imization of the likelihood under this constraint is equivalent to the maximization of

the penalized likelihood.

Let

gj(x) =

n∑k=1

ckjQj(xkj, x) + ρj(x),

where ρj ∈ Gj and is perpendicular to each Qj(xkj, ·). Then, by the properties of the

reproducing kernel Qj, we have

gj(xij) = 〈Qj(xij, ·), gj〉 =n∑

k=1

ckjQj(xkj, xij),

||gj||2 =

∫ 1

0

[g(mj)j (xj)]

2dxj =n∑

k=1

n∑i=1

ckjcijQj(xkj , xij) + ||ρj ||2.

It is then clear that for the maximizer gj, ρj must be taken to be zero.

48

Page 56: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Let

β = (α, β1, . . . , βd)T ,

cj = (c1j, . . . , cnj)T ,

T =

⎛⎝ 1 h1(x1) · · · hd(x1)

· · · · · · · · · · · ·1 h1(xn) · · · hd(xn)

⎞⎠ ,

Aj =

⎛⎝ Qj(x1j, x1j) · · · Qj(x1j, xnj)

· · · · · · · · ·Qj(xnj, x1j) · · · Qj(xnj , xnj)

⎞⎠ .

It can be easily obtained that the maximization of (3.4) is equivalent to the maxi-

mization of

lp(η, φ) = l(η, φ)− n

2

p∑j=1

λjc′jAjcj (3.5)

where

η = Tβ +

p∑j=1

Ajcj.

In the following, we show that the Newton-Rhapson algorithm with Fisher scoring

for the maximization of (3.5) is equivalent to an iterated weighted penalized least

square procedure.

By following the same procedure as in Section 2.3.1, we have

∂l(η, φ)

∂cj=

1

a(φ)

n∑i=1

[Y W

i − ∂bW (θi, φ)

∂θi

]∂θi

∂cj

=1

a(φ)

n∑i=1

wiqj(i)(YWi − µW

i )∂ηi

∂µWi

.

where qTj(i) is the ith row of Aj and

wi =

[∂hW (ηi)

∂ηi

]2 [∂2bW (θi, φ)

∂θ2i

]−1

.

As before, letting W = Diag(w1, . . . , wn) and∂η

∂µW = Diag( ∂η1

∂µW1

, . . . , ∂ηn

∂µWn

), we can

write∂l(η, φ)

∂cj

=1

a(φ)A

′jW

∂η

∂µW(Y w − µw).

49

Page 57: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Similarly, we can derive

∂l(η, φ)

∂β=

1

a(φ)T

′W

∂η

∂µ(Y w − µw).

Therefore, we obtain

∂lp(η, φ)

∂cj=

1

a(φ)A

′jW

∂η

∂µW(Y w −µw) − nλjAjcj,

∂lp(η, φ)

∂β=

1

a(φ)T

′W

∂η

∂µ(Y w − µw).

Furthermore, we have

E

[∂2lp(η, φ)

∂cj∂c′l

]= −Cov

(∂l(η, φ)

∂cj

,∂l(η, φ)

∂cl

)− nλj

∂Ajcj

∂c′l

,

=

− 1

a(φ)A

′jWAl j = l,

− 1a(φ)

A′jWAj − nλjAj j = l.

Similarly, we obtain

E

[∂2lp(η, φ)

∂β∂β′

]= − 1

a(φ)T

′WT,

E

[∂2lp(η, φ)

∂β∂c′j

]= − 1

a(φ)T

′WAj.

The updating step of the Newton-Rhapson algorithm with Fisher scoring, which up-

dates β(0), c(0)j , j = 1, · · · , p to β, cj, l = 1, · · · , p, is then amount to solving the

following equation:

⎛⎜⎜⎝

T′WT T

′WA1 · · · T

′WAp

A′1WT A

′1WA1 + na(φ)λ1A1 · · · A

′1WAp

· · · · · · · · · · · ·A

′pWT A

′pWA1 · · · A

′pWAp + na(φ)λpAp

⎞⎟⎟⎠

⎛⎜⎜⎜⎝

β − β(0)

c1 − c(0)1

· · ·cp − c

(0)p

⎞⎟⎟⎟⎠

=

⎡⎢⎢⎢⎣

T′W

∂η∂µw (Y w −µw)

A′1W

∂η∂µw (Y w −µw) − na(φ)λ1A1c

(0)1

· · ·A

′pW

∂η∂µw (Y w −µw) − na(φ)λpApc

(0)p

⎤⎥⎥⎥⎦ ,

50

Page 58: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where∂η

∂µw , Y w and µw are evaluated at β(0) and c(0)j , j = 1, . . . , p. After absorbing

a(φ) into the λj’s and moving the terms involving β(0) and c(0)j ’s on the left hand side

to the right hand side, we have⎛⎜⎜⎝

T′WT T

′WA1 · · · T

′WAp

A′1WT A

′1WA1 + nλ1A1 · · · A

′1WAp

· · · · · · · · · · · ·A

′pWT A

′pWA1 · · · A

′pWAp + nλpAp

⎞⎟⎟⎠

⎛⎜⎜⎝

βc1

· · ·cp

⎞⎟⎟⎠ =

⎛⎜⎜⎝

T′Wz

A′1Wz· · ·

A′pWz

⎞⎟⎟⎠ ,

(3.6)

where

z = η(0) +∂η

∂µw(Y w −µw).

Now consider the following penalized weighted least square problem:

Minimizing1

n(z − Tβ −

p∑j=1

Ajcj)′W (z − Tβ −

p∑j=1

Ajcj) +

p∑j=1

λjc′jAjcj

with respect to β, cj , j = 1, . . . , p. (3.7)

It is easy to obtain by taking derivatives and setting the derivatives to zero that

the normal equation of the above penalized weighted least square problem is given

by (3.6). Thus we have shown that the Newton-Rhapson algorithm with Fisher scor-

ing for the computation of the penalized maximum likelihood procedure is equivalent

to an iterated penalized weighted least square procedure. At each iteration, the

weight matrix W and the vector z are updated in exactly the same way as in the

IWSL procedure for the GLIM with weighted exponential families.

3.2.3 Relation of PMLE with Backfitting

We have seen in the last section that the penalized maximum likelihood estimation can

be carried out by an IPWLS procedure. At each iteration of the IPWLS procedure,

a linear system of the form given by (3.6) needs to be solved. This system is of order

(pn + d + 1) × (pn + d + 1). We are going to show that this order can be greatly

reduced. To this end, we shall establish the equivalence of the PMLE and a backfitting

system in terms of g0, g1, . . . , gp. The backfitting system is then used to derive explicit

expressions for the vectors β and cj’s which, in turn, are used to reduce the order of

51

Page 59: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

the linear system (3.6). An equivalent backfitting system in terms of f1, . . . , fp has

been obtained by Hastie and Tibshirani (1990, Section 6.5). However, their purpose

is to facilitate a backfitting algorithm, which we are not going to adopt for fitting a

GAM with weighted exponential family.

As a preparation, first let us consider the following simpler problems: Minimizing

1

n(yj −Ajcj)

′W (yj − Ajcj) + λjc

′jAjcj,

with respect to cj. By differentiating with respect to cj and setting the derivatives

to zero, we obtain

AjWAjcj + nλjAjcj = AjWyj.

It can be easily obtained that

W 1/2Ajcj = (W 1/2AjW1/2 + nλjI)−1W 1/2AjWyj

Next, consider the weighted least square problem: Minimizing

1

n(y0 − Tβ)

′W (y0 − Tβ),

with respect to β. The solution is given by

Tβ = T (T′WT )−1T

′Wy0.

Let

Aj = W 1/2AjW1/2,

S0 = W 1/2T (T′WT )−1T

′W 1/2,

Sj = (Aj + nλjI)−1Aj,

g0 = W 1/2Tβ,

gj = W 1/2Ajcj ,

yj = W 1/2yj , j = 0, 1, . . . , p.

The solutions for the above simpler problems can then be expressed as

gj = Sjyj, j = 0, 1, . . . , p.

52

Page 60: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Now suppose that β, cj , j = 1, . . . , p, jointly minimize

(z − Tβ −p∑

j=1

Ajcj)′W (z − Tβ −

p∑j=1

Ajcj) + n

p∑j=1

λjc′jAjcj.

Then β also minimizes

[(z −p∑

j=1

Ajcj) − Tβ]′W [(z −

p∑j=1

Ajcj) − Tβ],

and cj minimizes

[(z − T β −p∑

l =j

Alcl) − Ajcj]′W [(z − T β −

p∑l =j

Alcl) − Ajcj] + nλjc′jAjcj.

Let

y0 = z −p∑

j=1

Ajcj,

yj = z − T β −p∑

l =j

Alcl, j = 1, . . . , p.

Let g0, gj denote W 1/2T β and W 1/2Ajcj and let z = W 1/2z. Then we obtain

g0 = S0(z −p∑

j=1

gj),

g1 = S1(z − g0 −∑j =1

gj), (3.8)

. . .

gp = Sp(z − g0 −∑j =p

gj).

That is, the solution of penalized weighted least square problem satisfies a backfitting

system.

3.2.4 Simplification of IPWLE procedure

In the last section, we deduced that g0, gj, j = 1, . . . , p satisfy the backfitting sys-

tem (3.8). Specifically, we have

gj = Sj [(z − g0) −∑l =j

gl], j = 1, . . . , p. (3.9)

53

Page 61: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

In this section, we use the sub-backfitting system (3.9) to obtain explicit expressions

for the vectors cj’s which will lead to a simplification of the linear system (3.6) for

the penalized weighted least square problem.

It is easy to check that the Sj, j = 1, . . . , p, are all positive definite with eigenvalues

less than 1. Then, by using Lemma 6 of Breiman and Freidman, we have

gj = Bj(I + B)−1(z − g0), j = 1, . . . , p, (3.10)

where Bj = Sj(I −Sj)−1 and B =

∑pj=1 Bj. Recall that Sj = (Aj +nλjI)−1Aj. From

(Aj + nλjI)−1(Aj + nλjI) = (Aj + nλjI)(Aj + nλjI)−1 = I,

we have

(Aj + nλjI)−1Aj = Aj(Aj + nλjI)−1,

I − (Aj + nλjI)−1Aj = nλj(Aj + nλjI)−1.

Therefore,

Bj = Sj(I − Sj)−1

= (Aj + nλjI)−1Aj[nλj(Aj + nλjI)−1]−1

=1

nλjAj.

Thus we can write (3.10) as

gj =1

nλjAj(I + B)−1(z − g0).

That is,

W 1/2Ajcj =1

nλjW 1/2AjW

1/2(I + B)−1(z − g0), j = 1, . . . , p.

Finally, we obtain

cj =1

nλj

W 1/2(I + B)−1(z − g0), j = 1, . . . , p.

54

Page 62: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

The important implication of this result is that the coefficient vectors cj’s differ

only by the scalar factor 1/λj . Therefore, we can write cj = λ−1j c for some vector

c. Let ϑ = (ϑ1, . . . , ϑp) = (λ−11 , . . . , λ−1

p ). Substituting cj of this form into (3.5)

and (3.7), the penalized likelihood and the corresponding penalized weighted sum of

squares become, respectively,

lp(η, φ) = l(η, φ) − n

2c

′Aϑc

and1

n(z − Tβ − Aϑc)

′W (z − Tβ −Aϑc) + c

′Aϑc.

where Aϑ =∑p

j=1 λ−1j Aj and η = Tβ + Aϑc. The linear system for minimizing the

above penalized weighted sum of squares is given by(T

′WT T

′WAϑ

AϑWT AϑWAϑ + nAϑ

)(βc

)=

(T

′Wz

AϑWz

). (3.11)

Thus the linear system (3.6) is reduced to the above system whose order is only

(n + d + 1) × (n + d + 1).

Let

T = W 1/2T,

Aϑ = W 1/2AϑW 1/2,

c = W−1/2c.

Equation (3.11) can then be reduced to(T

′T T

′Aϑ

T Aϑ + nI

)(βc

)=

(T ′zz

). (3.12)

Further, it is easy to see that equation (3.12) is equivalent to

Tβ + (Aϑ + nI)c = z,

T′c = 0.

(3.13)

Equation (3.13) can be solved by standard numerical method. Let

T = FR∗ = (F1, F2)

(R0

)

55

Page 63: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

be the QR-decomposition of T . Then the solution of (3.13) is explicitly given by

c = F2(FT2 AϑF2 + nI)−1F T

2 z,

β = R−1F T1 (z − Aϑc).

The computation of c can be done with a forward substitution followed by a backward

substitution using the Cholesky decomposition F T2 AϑF2 + nI = GT G where G is

upper triangular.

3.2.5 Algorithms with Fixed Smoothing Parameters

In this section, we describe the algorithms for the implementation of the penalized

maximum likelihood estimation with fixed smoothing parameters. In previous sec-

tions, it is implicitly taken that the dispersion parameter φ is fixed. Although, in the

final formulation given in Section 3.2.4, φ does not explicitly appear in the formu-

las, the estimates of β and c still depend on φ, since the weight matrix W and the

pseudo-response vector z both involve φ. This is what is different from the GAM with

ordinary exponential families. Therefore, we need to distinguish the case of known φ

and the case of unknown φ.

Let T and Aϑ be as given in the previous section.

Algorithm 3 (φ is known).

Step 1. Initialization:

(a) Set z = (gW (y1), . . . , gW (yn))

T , W = W (y).

(b) Compute T = W 1/2T and Aϑ = W 1/2AϑW 1/2.

(c) Compute the QR-decomposition T = FR∗ = (F1, F2)

(R0

).

(d) Compute F Tz, F TAϑF , from which F T2 z, A∗

ϑ= F T

2 AϑF2, F T1 z and F T

1 AϑF2

can be extracted.

(e) Compute Cholesky decomposition A∗ϑ + nI = GT G.

56

Page 64: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(f) Solve for u from GT u = F T2 z by forward substitution and for v from Gv = u

by backward substitution.

(g) Return c = F2v and β = R−1(F T1 z − F T

1 AϑF2v).

Step 2. Iteration:

(a) Set W− = W

(b) Compute η = Tβ + AϑW1/2− c and µW = hW (η), from which compute

W = W (µW ) and z = η +∂η

∂µW T (y − µW ).

(c) Repeat (b) — (g) of Step 1 and return β, c.

(d) Check convergence conditions, if the conditions are satisfied, stop; otherwise, go

to (a).

Note that the value of φ is implicitly involved in W and z. For the sake of convenience,

we refer to Step 2 of Algorithm 3 as Algorithm 3′.

Algorithm 4 (φ is unknown).

Step 1. Initialization:

(a) Set φ = φ0. Call Algorithm 3 and return W , β, c.

(b) Compute η = Tβ + AϑW 1/2c. Maximize l(η, φ) with respect to φ and return

the maximizer φ−.

Step 2. Iteration:

(a) Set φ = φ−. Call Algorithm 3′and return W , β, c.

(b) Compute η = Tβ + AϑW 1/2c. Maximize l(η, φ) with respect to φ and return

the maximizer φ−.

(c) Check convergence conditions. If the conditions are satisfied, stop; otherwise, go

to (a).

57

Page 65: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

3.2.6 The Choice of Smoothing Parameters

In the literature of smoothing, a battery of data-driven methods for the choice of

smoothing parameters have been developed. Gu (2002, Chapters 3 and 5) has an

extensive discussion on these methods. In this section, we first give a brief review on

some of these methods and then extend them to the GAM with weighted exponential

families. An estimate of η with smoothing parameter ϑ will be denoted by ηϑ.

One approach for the choice of smoothing parameters is cross validation. The

basic idea of cross validation is as follows. The observations are left out one at a

time, at each time, the remaining observations are used to fit the model and the

observation left out is used to evaluate the prediction error of the fitted model. A

function called the cross validation (CV) score is formed from this procedure. The

smoothing parameters are then chosen to minimize the CV score. The CV score is

usually an estimate of certain loss function.

In the context of penalized weighted least square smoothing, the cross validation

amounts to minimize the following CV score:

V0(ϑ) =1

n

n∑i=1

ωi(yi − η[i]

ϑ(xi))

2,

where η[i]

ϑis the minimizer of

1

n

∑k =i

ωk(yk − η(xk))2 + Jϑ(η).

This CV score is taken as an estimate of the weighted square error loss given below:

L(ϑ) =1

n

n∑i=1

ωi[ηϑ(xi) − η(xi)]2.

In the context of penalized maximum likelihood smoothing with exponential fam-

ilies, the CV score is given as follows:

V0(ϑ) =1

n

n∑i=1

−yiθ(η[i]

ϑ(xi)) + b(θ(ηϑ(xi))). (3.14)

58

Page 66: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

This score is an estimate of the so-called relative Kullback-Leibler distance loss given

below:

RKL(η, ηϑ) =1

n

n∑i=1

−µ(xi)θ(ηϑ(xi)) + b(θ(ηϑ(xi))).

A CV score is usually computationally infeasible. In the context of penalized least

square smoothing, Craven and Wahba (1979) proposed to replace the CV score by a

generalized cross validation (GCV) score developed from the CV score. The method

of Craven and Wahba (1979) has been applied in other contexts to derive GCV scores

from CV scores. The GCV score for the penalized weighted least square smoothing

is given as follows:

V (ϑ) =n−1yT (I − H(ϑ))2y

[n−1tr(I − H(ϑ))]2,

where y = Ω1/2y, H(ϑ) is the smoothing matrix that maps y to Ω1/2η.

In the context of penalized maximum likelihood smoothing with exponential fam-

ilies, when the canonical parameter is directly modeled, i.e., the link function is taken

as the canonical link, the CV score becomes

V0(ϑ) =1

n

n∑i=1

−yiη[i]

ϑ(xi) + b(ηϑ(xi)).

Xiang and Wahba (1996) derived a generalized approximate cross validation score

from this CV score. Let their generalized approximate cross validation score be

denoted by GACV(ϑ).

Gu (2002) also considered two other types of scores for the choice of smoothing

parameters. One is an unbiased estimate of relative loss. The other one is based on a

Bayes model. Here, we are only concerned with the unbiased estimate of relative loss

which, for convenience, will be referred to as the U-score. In the context of penalized

weighted least square smoothing, the U-score is given by

U(ϑ) =1

ny[I − H(ϑ)]2y + 2

σ2

ntrH(ϑ),

where σ2 = Var(√

ωiyi). For the U-score to be applicable, σ2 must be known or a

good estimate free of the smoothing parameters must be available.

59

Page 67: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

In the context of penalized maximum likelihood smoothing with exponential fam-

ilies, since even for a fixed smoothing parameter a Newton-Rhapson procedure is

needed to compute the fitted function, different strategies have been proposed for

the choice of smoothing parameter. Gu (2002, chapter 5) discussed two strategies

which he referred to as direct cross-validation and performance-oriented iteration.

The direct cross-validation can be considered as a double Newton-Rhapson proce-

dure, an outer Newton-Rhapson iteration which updates the smoothing parameters

according to a score function and an inner Newton-Rhapson iteration which computes

the fitted function with fixed smoothing parameters. The smoothing parameters in

the outer Newton-Rhapson iteration are updated at the convergence of each inner

Newton-Rhapson iteration. The score function could be a GCV score for penalized

weighted least square smoothing based on the last step of the inner Newton-Rhapson

iteration (Wahba 1990) or a GACV score for penalized maximum likelihood smooth-

ing (Xiang and Wahba, 1996). The performance-oriented iteration is a quite different

strategy. The roles of the outer and inner iterations are interchanged. The outer

iteration updates the fitted function, but the smoothing parameters are not fixed.

Each updated fitted function has different smoothing parameters. The inner itera-

tion updates the smoothing parameters by using either the GCV score or the U-score

for penalized weighted least square smoothing. The differences of the two strategies

are as follows. The direct cross validation minimizes a certain loss function but the

computation is much more demanding. The performance-oriented iteration does not

minimize any loss function but the amount of computation is much less. Gu (1992)

compared the performance of the two strategies by simulation studies and found that,

for the performance-oriented iteration, use of the U-score is generally better than the

use of the GCV score when σ2 is known, and that, in most of cases, the performance-

oriented iteration approace is comparable with the direct cross validation approach.

See also Gu (2002, Section 5.2).

In the remainder of this section, we give the GCV score, U-score and GACV score

for the GAM with weighted exponential families.

For the GCV score and U-score, the vector y is to be replaced by z = Ω1/2z where

60

Page 68: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

W is the weight matrix and z is the pseudo-response vector defined in the previous

sections. The smoothing matrix H(ϑ) is given by

H(ϑ) = I − nF2(FT2 AϑF2 + nI)−1F T

2 ,

where F2 is the matrix in the QR-decomposition of T = (F1, F2)

(R0

). The details

to arrive at this result can be found in Gu (2002, chapter 3).

We now derive the GACV score for the GAM with weighted exponential families

by following the method of Xiang and Wahba (1996). In Xiang and Wahba (1996) and

Gu (2002), it is the canonical parameter of an exponential family that is modeled by

a smoothing function in a certain reproducing kernel Hilbert space. It is equivalent to

fitting a smooth predictor using the canonical link. If no special structure is imposed

on the smooth predictor, it does not matter whether to use the canonical link or

any other link. By modeling the canonical parameter directly, the formulation of

the model becomes simpler. But, if a special structure such as a linear function

or an additive function is imposed on the predictor, the link function matters since

the practical interpretation of the model differs when the link function differs. For

example, an additive function for the canonical parameter and an additive function

for the mean will have completely different implication in practice, which correspond

to the canonical link and the identical link. Therefore, we consider a general link

function, not necessarily the canonical link, for the GAM with weighted exponential

families. We can write the CV score for the GAM with a weighted exponential family

as follows.

V0(ϑ) =1

n

n∑i=1

−yiθ(η[i]

ϑ(xi)) + bW (θ(ηϑ(xi)))

=1

n

n∑i=1

−yiθ(ηϑ(xi)) + bW (θ(ηϑ(xi))) + yi[θ(ηϑ(xi)) − θ(η[i]

ϑ(xi))]

= l(ϑ) +1

n

n∑i=1

yi

θ(ηϑ(xi)) − θ(η[i]

ϑ(xi))

yi − µ[i]

ϑ(xi)

[yi − µ[i]

ϑ(xi)]

≈ l(ϑ) +1

n

n∑i=1

yi∂θ

∂η(ηϑ(xi))

ηϑ(xi) − η[i]

ϑ(xi)

yi − µ[i]

ϑ(xi)

[yi − µ[i]

ϑ(xi)]

61

Page 69: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

= l(ϑ) +1

n

n∑i=1

yi[yi − µϑ(xi)]∂θ

∂η(ηϑ(xi))Di, (3.15)

where

Di =

[ηϑ(xi) − η

[i]

ϑ(xi)

yi − µ[i]

ϑ(xi)

]/

[1 −

µϑ(xi) − µ[i]

ϑ(xi)

yi − µ[i]

ϑ(xi)

]. (3.16)

Since µϑ(xi) = ∂bW

∂θ(θ(ηϑ(xi))), by Taylor expansion, we have

µϑ(xi) − µ[i]

ϑ(xi) =

∂bW

∂θ(θ(ηϑ(xi))) −

∂bW

∂θ(θ(η

[i]

ϑ(xi)))

≈ ∂2bW

∂θ2(θ(ηϑ(xi)))

∂θ

∂η(ηϑ(xi))[ηϑ(xi) − η

[i]

ϑ(xi)]. (3.17)

Substituting (3.17) into (3.16), we have

Di = Ui/[1 −∂2bW

∂θ2(θ(ηϑ(xi)))

∂θ

∂η(ηϑ(xi))Ui],

where

Ui =ηϑ(xi) − η

[i]

ϑ(xi)

yi − µ[i]

ϑ(xi)

.

In what follows, we derive an approximate expression for Ui which will be computa-

tionally more tractable. Let

Iϑ(η, y) = −n∑

j=1

l(yj, η(xj)) +n

2c

′Aϑc,

where η = Tβ + Aϑc. Let

T = (F1, F2)

(R0

)= F1R

be the QR-decomposition of T , that is, (F1, F2) is orthogonal and R is upper-triangular.

Let

Σϑ = F2(F′2AϑF2)

−1F′2.

We now verify that

c′Aϑc = η

′Σϑη.

62

Page 70: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

It follows from (3.13) that T′c = 0 since T

′c = T

′W 1/2W−1/2c = T

′c. Therefore,

F′1c = 0 and c = F2a for some a. Since (F1, F2) is orthogonal, we have T

′F2 =

R′F

′1F2 = 0. Now we have

η′Σϑη = (β

′T

′+ c

′Aϑ)F2(F

′2AϑF2)

−1F′2(Tβ + Aϑc)

= a′F

′2AϑF2(F

′2AϑF2)

−1F′2AϑF2a

= a′F

′2AϑF2a

= c′Aϑc.

Hence, we have

Iϑ(η, y) = −n∑

j=1

l(yj, η(xj)) +n

′Σϑη

= −n∑

j=1

[yjθ(η(xj)) − bW (θ(η(xj)))] +n

′Σϑη. (3.18)

From (3.18) we can obtain

∂2Iϑ∂η(xi)∂η(xj)

=

∂2bW

∂θ2 (θ(η(xi)))[

∂θ∂η

(η(xi))]2

− [yi − µ(xi)]∂2θi

∂η2i

+ nσϑii, i = j,

nσϑij, i = j,

∂2Iϑ∂yi∂η(xj)

=

− ∂θi

∂ηi, i = j,

0, i = j,

where σϑijis the (i, j)th entry of Σϑ. Note that ∂2bW

∂θ2 (θ(η(xi)))[

∂θ∂η

(η(xi))]2

= ωi, the

ith diagonal element of the weight matrix Ω, since ∂θ∂η

(η(xi)) = ∂hW

∂η(xi)

[∂2bW

∂θ2 (θ(η(xi)))]−1

.

Let

∆ = Diag([y1 − µ(x1)]∂2θ1

∂η21

, . . . , [yn − µ(xn)]∂2θn

∂η2n

).

We can then write∂2Iϑ∂η∂η′ = Ω − ∆ + nΣϑ. (3.19)

By following exactly the same argument as in Xiang and Wahba (1996), we can obtain

Ui ≈ gii∂θi

∂ηi

,

63

Page 71: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where gii is the ith diagonal element of G = [Ω − ∆ + nΣϑ]−1 and the quantities

depending on η are evaluated at ηϑ. Substituting Ui into (3.16) and Di into (3.15),

we obtain

V0(ϑ) ≈ l(ϑ) +1

n

n∑i=1

yi[yi − µϑ(xi)][∂θi

∂ηi]2gii/[1 −

∂2bW

∂θ2i

(∂θi

∂ηi)2gii],

= l(ϑ) +1

n

n∑i=1

yi[yi − µϑ(xi)][∂θi

∂ηi]2gii/[1 − wigii].

Replacing gii and ωigii by (1/n)trG and (1/n)trΩ1/2GΩ1/2 respectively, we finally

arrive at the GACV score given below:

GACV(ϑ) =1

n

n∑i=1

−yiθ(ηϑ(xi)) + bW (θ(ηϑ(xi)))

+trG

ntr(I − Ω1/2GΩ1/2)

n∑i=1

yi[yi − µϑ(xi)][∂θi

∂ηi]2. (3.20)

3.2.7 Algorithms with Choice of Smoothing Parameters In-

corporated

In this section, we deal with the algorithms for fitting GAM with weighted expo-

nential families when the smoothing parameters are not fixed and must be selected.

We first describe the algorithm with performance oriented choice of smoothing pa-

rameters and the algorithm with direct cross validation in the context of GAM with

weighted exponential families. Then we propose a hybrid approach which combines

the performance oriented iterations and direct cross validation together.

For reasons that will become clear later, let the smoothing parameters be re-

parameterized as (λ1, . . . , λp) = (λτ−11 , . . . , λτ−1

p ) where the τj’s are subject to∑pj=1 τ 2

j = 1. Let τ = (τ1, . . . , τp). Then ϑ = λ−1τ , Aϑ = λ−1Aτ and equation

(3.13) becomesTβ + (Aτ + nλI)c∗ = z,

T′c∗ = 0,

(3.21)

where c∗ = λ−1c and Aτ = W 1/2Aτ W 1/2. Note that Aϑc = Aτ c∗ = W 1/2Aϑc and

η = W−1/2(Tβ + Aτ c∗). For the sake of convenience, by an abuse of notation, we

64

Page 72: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

denote c∗ by c in this section. The solution of (3.21) is explicitly given by

c = F2(FT2 Aτ F2 + nλI)−1F T

2 z,

β = R−1F T1 (z − Aτ c).

The matrix H(ϑ) appearing in the GCV score and U-score associated with (3.21) is

given by

H(λ, τ ) = I − nλF2(FT2 Aτ F2 + nλI)−1F T

2 .

For fixed τ , let F T2 Aτ F2 = U∆UT be a decomposition such that U is orthogonal and

∆ is tridiagonal. Then H(λ, τ ) can be expressed as

H(λ, τ ) = I − nλF2U(∆ + nλI)−1UT F T2 .

As mentioned in the last section, when σ2 = a(φ) is known, the performance

oriented algorithm with U-score is generally better than that with GCV score. Since,

in the context of GAM with weighted exponential families, a(φ) is either known or

fixed in the iteration for updating η, the U-score will be used in the algorithms we

are going to describe. The U-score associated with (3.21) is given by

U(λ, τ )

=1

nzT [I − H(λ, τ )]2z +

2a(φ)

ntrH(λ, τ )

= nλ2zT F2(FT2 Aτ F2 + nλI)−2F T

2 z

−2a(φ)λtr(F T2 Aτ F2 + nλI)−1 + 2a(φ) (3.22)

= nλ2xT (∆ + nλI)−2x − 2a(φ)λtr(∆ + nλI)−1 + 2a(φ) (3.23)

= U∗(λ, τ ) + 2a(φ), say,

where x = UT F T2 z. When U(λ, τ ) ( or U∗(λ, τ ) ) is treated as a function of τ with

λ fixed, (3.22) is to be used, and when it is treated as a function of λ with τ fixed,

(3.23) is to be used.

The following algorithm, which is quoted from Gu (2002, page 79-80) with slight

modifications, is the core of the performance oriented algorithm.

65

Page 73: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Algorithm 5 Given T , Aj, j = 1, . . . , p,, φ and starting values τ 0, perform the fol-

lowing steps to minimize U∗(λ, τ ) and return the associated vectors β and c:

Step 1. Initialization:

(a) Compute the QR-decomposition T = FR∗ = (F1, F2)

(R0

).

(b) Compute F T z and F TAjF , from which F T2 z, A∗

j = F T2 AjF2, F T

1 z and F T1 AjF2

can be extracted.

(c) Set ∆τ = 0, τ− = τ 0 and u− = ∞.

Step 2. Iteration:

(a) For τ = τ−+∆τ , scale τ such that τ has length 1 and compute A∗τ =

∑pj=1 τjA

∗j .

(b) Compute A∗τ = UDUT , where U is orthogonal and D is tridiagonal. Compute

x = UT F T2 z.

(c) Minimize U∗(λ, τ ) with respect to λ. Return the minimizer λ0 and u = U∗(λ0, τ ).

If u > u−, set ∆τ = ∆τ/2, go to (a); else proceed.

(d) Evaluate g = (∂/∂τ )U∗(λ0, τ ) and H = (∂2/∂τ∂τT )U∗(λ0, τ ).

(e) Calculate ∆τ = −H∗−1g, where H∗ = H + Diag(e) is positive definite. If H

itself is positive definite “enough”, e is set to zero.

(f) Check convergence conditions. If the conditions fail, set τ− = λ−10 τ , u− = u, go

to (a).

Step 3. Compute return values:

(a) Compute v = U(D + nλ0I)−1x at the convergent λ0 and τ .

(b) Return c = F2v and β = R−1(F T1 z − F T

1 Aτ F2v).

66

Page 74: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

The modifications we have done over Gu’s algorithm are: (i) in Step 2 (a), we

scale τ such that the length of τ is 1 in stead of that the trace of A∗τ is fixed, because

fix of τ is easier to handle than fix of A∗τ , and (ii) in Step 2 (d) — (f), the Newton

iteration is indeed used to update λ−1τ rather than τ itself.

Algorithm 5 can be illustrated in the case p = 2 by Figure 1. The point (ϑ∗1, ϑ

∗2)

stands for the optimum point of (ϑ1, ϑ2). The point (τ10, τ20) on the arc of the unit

circle is the starting point of (τ1, τ2). Before the first updating of (τ1, τ2), a univariate

search on the line λ−1(τ10, τ20) updates (ϑ1, ϑ2) to the point (ϑ10, ϑ20) = λ∗−1(τ10, τ20)

(Step 2 (c)). Then a Newton iteration (Step 2 (d) and (e)) updates (ϑ1, ϑ2) from

(ϑ10, ϑ20) to (ϑ10 + ∆τ1, ϑ20 + ∆τ2) which, when normalized, produces the updates

(τ11, τ21) . The next univariate search on the line λ−1(τ11, τ21) updates(ϑ1, ϑ2) to the

point (ϑ11, ϑ21) = λ∗∗−1(τ11, τ21), and so on.

In Algorithm 5, the step of univariate search on a straight line determined by a

point (τ1, τ2) on the arc of the unit circle seems superfluous. However, it is this step

that greatly speeds up the convergence of the algorithm. This can be explained as

follows. In the smoothing procedure, the roles of λ and τj ’s are different. Roughly

speaking, λ determines the total amount of penalty (or smoothness of the additive

predictor) and τj’s determine how the total amount of penalty is distributed to dif-

ferent components of the additive predictor. It is the value of λ that dominates the

scores to be minimized. Therefore, a univariate search over λ has the potential to

update the smoothing parameters faster towards the optimum point. Besides, the

amount of computation required for the univariate search is negligible compared with

the total amount of computation required in each iteration. As noted by Gu (2002),

each iteration of Step 2 takes altogether 4pn3/3 + O(n2) flops while Step 2 (c) takes

only O(n2) flops.

67

Page 75: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 1. An illustration of Algorothm 5

Performance Oriented Algorithm

We describe in the following the performance oriented algorithm for the fitting of

GAM with weighted exponential family. The score U∗(λ, τ ) whose forms are given

by (3.22) and (3.23) is used in the algorithm.

Algorithm 6 (φ is known).

Given T , Aj, j = 1, . . . , p, and y.

Step 1. Initialization:

68

Page 76: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(a) Set z = (gW (y1), . . . , gW (yn))

T and W = W (y).

(b) Compute T = W 1/2T , Aj = W 1/2AjW1/2, j = 1, . . . , p, and z = W 1/2z. Set

τ 0 = (1/p, . . . , 1/p).

(c) For T , Aj, j = 1, . . . , p, z and τ 0, call Algorithm 5 to minimize U∗(λ, τ ) and

return the associated vectors β, c and the minimizers λ, τ .

Step 2. Iteration:

(a) Set W− = W .

(b) Compute η = Tβ + Aτ W1/2− c and µW = hW (η), from which compute

W = W (µW ) and z = η +∂η

∂µW T (y − µW ).

(c) Compute T = W 1/2T , Aj = W 1/2AjW1/2, j = 1, . . . , p, and z = W 1/2z. Set

τ 0 = τ

(d) For T , Aj, j = 1, . . . , p, and τ 0, call Algorithm 5 to minimize U∗(λ, τ ) and

return the associated vectors β, c and the minimizers λ, τ .

(c) Check convergence conditions. If the conditions are satisfied, return β, c and

λ, τ , otherwise, go to (a).

If φ is unknown, an algorithm of the type of Algorithm 4, which alternates between a

step that updates the vectors β and c and a step that updates φ, can be developed.

Algorithm 7 (φ is unknown)

Step 1. Initialization:

(a) Set φ = φ0. Call Algorithm 6 and return β, c and λ, τ .

(b) Compute η = Tβ + Aτ W1/2− c. Maximize l(η, φ) with respect to φ and return

the maximizer φ−.

Step 2. Iteration:

69

Page 77: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(a) Set φ = φ−. Call Algorithm 6 and return β, c and λ, τ .

(b) Compute η = Tβ + Aτ W1/2− c. Maximize l(η, φ) with respect to φ and return

the maximizer φ−.

(c) Check convergence conditions. If the conditions are satisfied, stop; otherwise, go

to (a).

Direct Cross Validation Algorithm

For the direct cross validation algorithm, it is more convenient to work with the

parameterization ϑ = (ϑ1, . . . , ϑp) = (λ−11 , . . . , λ−1

p ). Let V (ϑ) denote the GACV

score evaluated at ϑ. The direct cross validation algorithm for the fitting of GAM

with weighted exponential family with known φ is described in this paragraph. The

remark at the end of last paragraph applies when φ is unknown.

Algorithm 8 (Direct cross validation)

Step 1. Initialization:

(a) Set ϑ = ϑ0, a specified starting vector.

(b) For fixed ϑ, call Algorithm 3 and return β, c and W .

(c) Compute g = (∂/∂ϑ)V (ϑ) and H = (∂2/∂ϑ∂ϑT )V (ϑ).

(d) Compute ∆ϑ = −H∗−1g, where H∗ = H + Diag(e) is positive definite. If H

itself is positive definite enough, e is set to zero.

(e) Set ϑ− = ϑ.

Step 2. Iteration:

(a) Let ϑ = ϑ− + ∆ϑ, β0 = β, c0 = c.

(b) For fixed ϑ and starting values β0, c0, call Algorithm 3′, return β, c and W .

70

Page 78: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

(c) Compute g = (∂/∂ϑ)V (ϑ) and H = (∂2/∂ϑ∂ϑT )V (ϑ).

(d) Compute ∆ϑ = −H∗−1g, where H∗ = H + Diag(e) is positive definite. If H

itself is positive definite enough, e is set to zero.

(e) Check convergence conditions. If convergence conditions are satisfied, return ϑ,

β and c; otherwise, set ϑ− = ϑ and go to (a).

A Hybrid Approach

In the direct cross validation algorithm, the strategy of searching over λ for fixed τ

can not be implemented with negligible amount of computation, since we do not have

a form similar to (3.23) which is unchanged throughout the iterations in Step 2 of

Algorithm 5 and Algorithm 6. The computation amount will be tremendously heavy

unless we have a starting vector ϑ0 which is very close to the optimum point. The

performance oriented algorithm needs much less amount of computation. However,

since it does not minimize the cross validation (or generalized approximate cross

validation) score, the resultant choice of the smoothing parameters will not perform

as well as the ones chosen by the direct cross validation algorithm.

To take advantage of the computation-saving feature of the performance oriented

algorithm and retain the optimality which can be achieved by the direct cross vali-

dation algorithm, we propose a hybrid approach. The hybrid approach combines the

two algorithms together. First the performance oriented algorithm is implemented to

produce a convergent vector of smoothing parameters. Then this vector is used as

the starting value in the direct cross validation algorithm which is implemented to

yield the final choice of the smoothing parameters. The vector resulted from the per-

formance oriented algorithm can be obtained with much less amount of computation

than the optimal vector which minimizes the GACV score by the direct cross valida-

tion algorithm. On the other hand, though the vector resulted from the performance

oriented algorithm does not necessarily minimize the GACV score, it can be expected

that it will be in a neighborhood of the optimal vector which minimizes the GACV

71

Page 79: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

score. Therefore, when it is used as the starting value in the direct cross validation

algorithm, the latter algorithm can be expected to converge in only a few iterations.

The hybrid algorithm is described as follows.

Algorithm 9 (Hybrid)

Step 1. Initial minimization with performance oriented algorithm:

(a) Given T , Aj, j = 1, . . . , p, and y, call Algorithm 6 and return λ, τ , β and c.

(b) Set ϑ− = λ−1τ and ∆ϑ = 0.

Step 2. Iteration with direct cross validation:

(a) Let ϑ = ϑ− + ∆ϑ, β0 = β, c0 = c.

(b) For fixed ϑ and starting values β0, c0, call Algorithm 3′, return β and c.

(c) Compute g = (∂/∂ϑ)V (ϑ) and H = (∂2/∂ϑ∂ϑT )V (ϑ).

(d) Compute ∆ϑ = −H∗−1g, where H∗ = H + Diag(e) is positive definite. If H

itself is positive definite enough, e is set to zero.

(e) Check convergence conditions. If convergence conditions are satisfied, return ϑ,

β and c; otherwise, set ϑ− = ϑ and go to (a).

72

Page 80: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 4

Special Models with WeightedExponential Families

In Chapters 2 and 3, we developed the general theory for models with weighted

exponential families including generalized linear and additive models. In this chapter,

we consider some important special models. In §4.1, we discuss models for binomial

data. Models for count data are given in §4.2, followed with §4.3 on models for data

having constant coefficient of variation.

4.1 Models for binomial data

4.1.1 Introduction

A random variable Y is said to have a binomial distribution if its probability density

function is given by

P (Y = j) =

(n

j

)πj(1 − π)n−j, j = 0, 1, . . . , n,

where n is a fixed integer and π (0 < π < 1) is a unknown parameter, usually

referred to as probability of success. The binomial distribution arises in either of two

ways. First, in n independent and identical Bernoulli trials, the number of successes

is a binomial random variable. A Bernoulli trial is an experiment that has only two

outcomes, one is called success and the other is called failure. The success occurs with

probability π and the failure occurs with probability 1 − π. Second, if two Poisson

73

Page 81: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

random variables are independent, then the conditional distribution of one of them

given the sum of them is a binomial distribution. In the first case, the binomial

random variable is the sum of n binary variables. However, in the second case, the

binomial random variable arises from count variables. The genesis of the binomial

distribution implies that models with binomial distributions not only apply to binary

data but also apply to certain count data.

The binomial distribution is an exponential family. In the context of generalized

linear or additive models, it is more convenient to deal with the proportion variable

Y = Y /n rather than Y itself. In the form of exponential family, the probability

density function of Y is given by

P (Y = y) = exp

ny ln

1 − π

)+ n ln(1 − π) + ln

(n

ny

)

= expyθ − b(θ)

a(φ)+ c(y, φ),

y = 0, 1/n, 2/n, . . . , 1,

where

θ = ln

1 − π

),

b(θ) = ln(1 + eθ),

a(φ) = 1/n,

c(y, φ) = ln

(n

ny

).

Therefore, the cumulant generating function of Y is given by

κ(t) =1

a(φ)[b(θ + ta(φ))− b(θ)]

= n[ln(1 + eθ+t/n) − ln(1 + eθ)],

= n ln(1 − π + πet/n).

In terms of π, the first four cumulants are given by

κ1 = π,

74

Page 82: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

κ2 = π(1 − π)/n,

κ3 = π(1 − π)(1 − 2π)/n2,

κ4 = π(1 − π)[1 − 6π(1 − π)]/n3.

Weighted binomial distributions arise naturally in the ascertainment of family data

for certain human heredity studies and social surveys. In human heredity studies, a

primary purpose in collecting family data is to make inference on the proportion

of affected children in a family. If the inherited trait of concern is rare, a random

sampling of families is not appropriate, since it will result in a large number of families

that do not possess the trait and yield no information. A quite common method

of ascertainment is to ascertain the affected children first. Once an affected child

is ascertained, his or her family is then ascertained. If each affected child has an

equal chance of being ascertained, then the chance for a family to be ascertained

depends on the number of affected children in the family. The larger the number of

affected children, the larger the chance. Thus, for an ascertained family, the number

of affected children follows a weighted binomial distribution in stead of the original

binomial distribution in reality. An well-known example on a study of albinism was

given and considered by Fisher (1934), Haldane (1938), Rao (1965) and Patil and

Rao (1977). Examples of weighted binomial data in social surveys can be found in

Rao (1965).

When the probability of success of a binomial random variable is related to other

variables and the binomial random variable is ascertained under the mechanism of

weighted distribution, the generalized linear models or generalized additive models

with weighted binomial distributions come naturally into play. Rao (1965) considered

an example for inference on sex ratio and family size. In this example, the numbers

of brothers and sisters in families of 104 boys who were admitted to a post graduate

course at the Indian Statistical Institute in a period of 5 years were recorded. The

numbers were obtained by first sampling these 104 boys and then asking them the

numbers. Weighted binomial distributions are suitable to describe the data in this

example. If we were interested in finding out the effect of factors such as family

75

Page 83: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

income, education level of family head and race, etc. on the sex ratio and family size,

a generalized linear model or additive model with weighted binomial distributions is a

natural choice for the analysis of the data with additional information on those factors.

In this section, we deal in detail with models using weighted binomial distributions.

4.1.2 Weight functions for models with binomial distribu-

tions

In this sub-section, we discuss the weight functions to be used in models with binomial

distributions. We consider the following weight functions:

(i) w1(y) = 1 − (1 − β)y, 0 < β < 1,

(ii) w2(y) = y,

(iii) w3(y) = yα, α > 0.

The above weight functions have been considered by Rao (1965), Neel and Schull

(1966) and Patil and Rao (1977).

In the human heredity study mentioned in the last subsection, if the chance for

each affected child to be ascertained is the same and independent of which family he

or she belongs, the weight function is then given by w1. Here β is the probability with

which an affected child will be ascertained and w1(y) is the probability with which at

least one of the affected children in a family of y affected children will be ascertained.

If β is small, we can expand w1 as a function of β at β = 0 to obtain

w1(y) = βy + O(β2).

It is obvious that, as β approaches zero, the weighted probability density function

becomes

P (Y W = y) =yP (Y = y)

EY,

where EY is the un-weighted expectation of Y . This gives rise to the weight function

w2.

The weight function w3, which is a natural generalization of w2, gives a family of

weighted functions.

76

Page 84: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

If we work with the proportion variable Y in stead of Y , the weight function w1

must be adjusted to w1(y) = 1− (1− β)ny whereas w2 and w3 remain unchanged. In

the remainder of this section, our discussion will be in terms of Y unless otherwise

mentioned.

Let ωi(θ) denote the parameter function corresponding to Wi(y) for i = 1, 2, 3.

These parameter functions are given below:

ω1(θ) =n∑

j=0

(n

j

)[1 − (1 − β)j]πj(1 − π)n−j

= 1 − [1 − βπ]n

= 1 −[1 − βeθ

1 + eθ

]n

= 1 −[1 + (1 − β)eθ

1 + eθ

]n

ω2(θ) =1

n

n∑j=0

(n

j

)jπj(1 − π)n−j

= π =eθ

1 + eθ.

ω3(θ) =1

n∑j=0

(n

j

)jαπj(1 − π)n−j

=1

n∑j=0

(n

j

)jα ejθ

(1 + eθ)n.

The weighted binomial distributions with the weight functions W1, W2 and W3

have their bW (θ) functions given below:

bW1 (θ) = ln(1 + eθ) +

1

nln

[1 −

(1 + (1 − β)eθ

1 + eθ

)n]

=1

nln[(1 + eθ)n − (1 + (1 − β)eθ)n],

bW2 (θ) = ln(1 + eθ) +

1

nln

(eθ

1 + eθ

)

n+

n − 1

nln(1 + eθ),

bW3 (θ) = ln(1 + eθ) +

1

nln

(1

n∑j=0

(n

j

)jα ejθ

(1 + eθ)n

)

77

Page 85: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

=1

n

[ln

(n∑

j=0

(n

j

)jαejθ

)− α lnn

].

The first derivatives of the above bW -functions are derived as follows.

dbW1 (θ)

dθ=

eθ[(1 + eθ)n−1 − (1 + (1 − β)eθ)n−1(1 − β)]

(1 + eθ)n − (1 + (1 − β)eθ)n

=π[1− (1 − β)(1 − βπ)n−1]

1 − (1 − βπ)n= µW

1 (π);

dbW2 (θ)

dθ=

1

n+

n − 1

n

1 + eθ=

1

n+

n − 1

nπ = µW

2 (π);

dbW3 (θ)

dθ=

1

n

ν(α + 1, θ)

ν(α, θ)=

1

n

ν(α + 1, ln[π/(1 − π)])

ν(α, ln[π/(1− π)])= µW

3 (π),

where

ν(α, θ) =n∑

j=0

(n

j

)jαejθ.

For the second derivatives, it is easier to derive in the form

d2bW (θ)

dθ2=

dµW (π)

dθ.

Note that dπdθ

= π(1 − π) in terms of π. It is derived that

dµW1 (π)

dπ=

d exp(lnµW1 (π))

dπ= µW

1 (π)d lnµW

1 (π)

=µW

1 (π)

π

[1 − nβπ(1− βπ)n−1

1 − (1 − βπ)n

]+

(n − 1)(1 − β)βπ(1− βπ)n−2

1 − (1 − βπ)n

=nµW

1 (π)(1 − µW1 (π))

π(1 − π)− (n − 1)

1 − (1 − β)(1 − βπ)n−2

1 − (1 − βπ)n.

dµW2 (π)

dπ=

n − 1

n.

Hence, in terms of π,

d2bW1 (θ)

dθ2= nµW

1 (π)(1− µW1 (π))− (n − 1)π(1 − π)

1 − (1 − β)(1 − βπ)n−2

1 − (1 − βπ)n.

d2bW2 (θ)

dθ2=

n − 1

nπ(1 − π).

78

Page 86: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

For bW3 , it derived directly that

d2bW3 (θ)

dθ2=

1

n

ν(α, θ)ν(α + 2, θ) − (ν(α + 1, θ))2

(ν(α, θ))2

=1

n

ν(α, ln(π/(1 − π))ν(α + 2, ln(π/(1 − π)) − (ν(α + 1, ln(π/(1 − π)))2

(ν(α, ln(π/(1 − π)))2.

4.1.3 Link and response functions for models with binomialdistributions

We now turn to the link and response functions for models with binomial distributions.

In principle, a link function for models with binomial distributions must be a function

that maps the unit interval [0, 1] to the whole real line (−∞, +∞), because the range

of the mean µ = EY is [0, 1] and the linear or additive predictor η could take any real

values. Therefore, we can invert any distribution function with support (−∞, +∞) to

obtain a link function. The distribution function will then be the response function.

The following link functions, which are taken from McCullagh and Nelder (1989)

and have been commonly in use for analyzing binomial data, are examples of link

functions obtained this way:

• Logit link:

g1(µ) = ln

1 − µ

).

• Probit link:

g2(µ) = Φ−1(µ),

where Φ−1 denotes the inverse of Φ, the distribution function of the standard

normal distribution.

• Complementary log-log link:

g3(µ) = ln[− ln(1 − µ)].

The corresponding response functions are the following distribution functions:

h1(η) =eη

1 + eη,

79

Page 87: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

h2(η) = Φ(η),

h3(η) = 1 − e−eη

.

In the following, we derive the response functions of the weighted models with the

above link functions and the weight functions given in the last sub-section. Let hWik (η)

denote the response function of the weighted model with link function gi and weight

function Wj. Corresponding to each link function, the weighted response functions

are given below.

• Logit link:

hW1l (η) =

dbWl (η)

dη, l = 1, 2, 3,

where the derivatives are given in the last subsection, since with the logit link,

θ = η .

• Probit link:

hW21(η) =

Φ(η)[1 − (1 − β)(1− βΦ(η))n−1]

1 − (1 − βΦ(η))n,

hW22(η) =

1

n+

n − 1

nΦ(η),

hW23(η) =

1

n

ν1(ln[Φ(η)/(1 − Φ(η)])

ν0(ln[Φ(η)/(1 − Φ(η)]).

• Complementary log-log link:

hW31(η) =

(1 − e−eη)[1 − (1 − β)(1 − β + βe−eη

)n−1]

1 − (1 − β + βe−eη)n,

hW32(η) = 1 − n − 1

ne−eη

,

hW33(η) =

1

n

ν1(ln(1 − e−eη) + eη)

ν0(ln(1 − e−eη ) + eη).

4.1.4 Estimation of weighted generalized linear models andweighted generalized additive model with binomial data

Consider the data set (Yi, xi) : i = 1, . . . , n where Yi is ascertained with a weight

function from a binomial distribution with index mi and probability of success πi, and

80

Page 88: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

xi = (xi1, . . . , xip) is the vector of the values of p predictor variables associated with

the ith subject. Note that the distribution of Yi is a weighted binomial distribution.

The general algorithms for fitting weighted generalized linear models described in

Chapter 2 are to be applied for the estimation of the models with binomial data. Since

there is no unknown dispersion parameter in the binomial distributions, we apply

Algorithm I given in Chapter 2 for the estimation. In this subsection, we supplement

details in the general algorithm when binomial data is of concern. Algorithm I is a

iterated weighted least square procedure. At each iteration, a pseudo-response vector

Z = (z1, . . . , zn) and a weight matrix W = diag(w1, . . . , wn) are updated from the

previous estimated parameters, then zi is regressed on xi with weight wi and the

estimates of the parameters are updated. In the following, we give the formulae

for updating the pseudo-responses and weights with particular response and weight

functions discussed in this section. Given link function gl(π) and weight function

Wk(y), the pseudo-responses and weights are given respectively by

zi = ηi +

[∂hl(ηi)

∂ηi

∂µWk (πi)

∂πi

]−1

(yi − hWlk (ηi)),

wi =1

πi(1 − πi)

[∂hl(ηi)

∂ηi

]2∂µW

k (πi)

∂πi,

where∂µW

k (πi)

∂πiand hW

lk (ηi) are given at the bottom of page 78 in the previous sub-

sections, and the ∂hl(ηi)∂ηi

are given below:

∂h1(ηi)

∂ηi=

eηi

(1 + eηi)2,

∂h2(ηi)

∂ηi= φ(ηi),

∂h3(ηi)

∂ηi= eηi−eηi

.

The general algorithms for fitting weighted generalized additive models described

in Chapter 3 will be applied for the estimation of the models with binomial data. As

there is no unknown dispersion parameter in the binomial distribution, we apply Al-

gorithm 6 given in Chapter 3 for the estimation. The pseudo-responses and weights

81

Page 89: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

which are used in the algorithm are exactly the same as those in the weighted gener-

alized linear model which are described above. The GACV need modified accordingly

as well.

It is interesting to notice that, when Y W is the length-biased version of a binomial

random variable with index m and probability of success π, then Y W −1 will follow an

ordinary binomial distribution. Indeed, we have that the probability density function

of Y W is given by

P (Y W = j) =

(mj

)πj(1 − π)m−jj

=

(m − 1

j − 1

)πj−1(1 − π)(m−1)−(j−1),

j = 1, . . . , m.

Thus, Y W −1 follows an ordinary binomial distribution with index m−1 and the same

probability of success π. Therefore, the weighted models can be estimated through

the estimation of un-weighted models based on the transformed data Y W − 1.

4.2 Models for count data

4.2.1 Introduction

A random variable Y is said to have a Poisson distribution if its probability density

function is given by

p(y) = P (Y = y) = e−λλy/y!; y = 0, 1, . . .

where λ (> 0) is a unknown parameter. The Poisson distribution is usually used to

model count data. It arises in the Poisson counting process as the distribution of the

number of events occurred in any given interval. It also arises as a limiting distribution

of a sequence of binomial random variables. A sequence of random variables with

binomial distribution B(n, p) converges to a Poisson random variable with parameter

λ if, as n → ∞, np → λ. The density function of the Poisson distribution can be

expressed in exponential family form as

p(y) = expy lnλ − λ − ln y!.

82

Page 90: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

It can be identified that

θ = lnλ, b(θ) = eθ, a(φ) = 1, c(y) = − ln y!.

The cumulant generating function of Y is given by

κ(t) = λ(et − 1).

Hence, the cumulants are a constant

κk = λ, k = 1, 2, . . .

The weighted Poisson distribution arises most naturally in aerial survey of wild

animals. Suppose the group size of the wild animals is of concern. Suppose that

an entire group of animals is observed if any member of the group is observed and

that each individual animal is sighted with the same probability. If the group size

is modeled by a Poisson random variable (indeed a Poisson random variable minus

one), then the observed group size follows a weighted Poisson distribution.

The generalized linear or additive models with weighted Poisson distributions are

dealt with in this section.

4.2.2 Weight and link functions for models with count data

The same weight functions, i.e., w1(y) = 1 − (1 − β)y, w2(y) = y and w3(y) =

yα, considered for models with binomial data can be used for models with Poisson

distributions. The parameter functions ωi(λ), i = 1, 2, 3 corresponding to these weight

functions for the weighted Poisson distribution are given below:

ω1(θ) = 1 − e−βλ

= 1 − e−βeθ

,

ω2(θ) = λ

= eθ,

ω3(θ) =∑

yαe−λλy/y!

=∑

yαe−eθ+θy/y! = m(α, θ),

83

Page 91: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

where m(α, θ) is the αth moment of the Poisson distribution.

The function bW (θ) in the weighted Poisson distribution associated with the weight

functions W1, W2 and W3 are given below:

bW1 (θ) = eθ + ln(1 − eβeθ

),

bW2 (θ) = eθ + θ,

bW3 (θ) = eθ + lnm(α, θ).

The first derivatives of the above functions, which provide the means of the

weighted distributions, are given below:

∂bW1 (θ)

∂θ= eθ(1 +

β

eβeθ − 1)

= λ(1 +β

eβλ − 1) = µW

1 ,

∂bW2 (θ)

∂θ= eθ + 1

= λ + 1 = µW2 ,

∂bW3 (θ)

∂θ=

m(α + 1, θ)

m(α, θ)= µW

3 .

The second derivatives of the bW functions are given below:

∂2bW1 (θ)

∂θ2= eθ

[1 +

β

eβeθ − 1

(1 − eθβeβeθ

eβeθ − 1

)]

= λ

[1 +

β

eβλ − 1

(1 − λβeβλ

eβλ − 1

)],

∂2bW2 (θ)

∂θ2= eθ = λ,

∂2bW3 (θ)

∂θ2=

m(α + 2, θ)

m(α, θ)−

[m(α + 1, θ)

m(α, θ)

]2

.

We now turn to the link and response functions for models with Poisson distri-

butions. Since the range of the mean of a Poisson distribution is [0, +∞) and the

predictor η could take any real values, in principle, any monotone function that maps

[0, +∞) to (−∞, +∞) can serve as a link function for the Poisson model. Here we

84

Page 92: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

only consider the canonical log link which is in common use with Poisson models.

The canonical log link function is given by

η = ln(µ).

The corresponding response function is simply the exponential function

µ = eη.

With the log link, η = θ, the weighted response functions corresponding to the

three weighted functions are given by the first derivative of the bW functions as derived

earlier. That is,

hW1 (η) = eη(1 +

β

eβeη − 1),

hW2 (η) = eη + 1,

hW3 (η) =

m(α + 1, η)

m(α, η).

4.2.3 Estimation of weighted generalized linear models with

count data

Consider the data set (Yi, xi) : i = 1, . . . , n where Yi is ascertained with a weight

function from a Poisson distribution with λi, and xi = (xi1, . . . , xip) is the vector of

the values of p predictor variables associated with the ith subject. The observed Yi

follows a weighted Poisson distribution. Again, for convenience, we have suppressed

the superscript W on the weighted variables.

As in the case of binomial distribution, there is no unknown dispersion parameter

in the Poisson model. Therefore, Algorithm I and Algorithm 6 can be applied to

fit the generalized linear and additive models with Poisson distribution respectively.

Note that η = θ under the log link. Therefore, the formulae for updating the

pseudo-responses and the weights of the Poisson model in the implementation of the

Algorithms are given as follows:

zi = ηi +

[∂2bW

l (ηi)

∂η2i

]−1

[yi − hWl (ηi)]

85

Page 93: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

= ηi +

⎧⎪⎪⎪⎨⎪⎪⎪⎩

e−ηi

[1 + β

eβeηi −1

(1 − eηi βeβeηi

eβeηi −1

)]−1

[yi − eηi(1 + β

eβeηi −1)], l = 1,

e−ηi[yi − eηi − 1], l = 2,[m(α+2,ηi)m(α,ηi)

−(

m(α+1,ηi)m(α,ηi)

)2]−1

[yi − m(α+1,ηi)m(α,ηi)

], l = 3.

wi =∂2bW

l (ηi)

∂η2i

=

⎧⎪⎪⎨⎪⎪⎩

eηi

[1 + β

eβeηi −1

(1 − eηiβeβeηi

eβeηi −1

)], l = 1,

eηi, l = 2,m(α+2,ηi)

m(α,ηi)−

(m(α+1,ηi)

m(α,ηi)

)2

, l = 3.

Let Y W be the length-biased version of a Poisson random variable with parameter

λ . Indeed, we have that the probability density function of Y W is given by

P (Y W = j) =je−λλj/j!

λ= e−λλj−1/(j − 1)!,

j = 1, 2, . . . .

Thus, Y W − 1 follows an ordinary Poisson distribution with the same parameter

λ. Therefore, the weighted models can be estimated through the estimation of un-

weighted models based on the transformed data Y W − 1.

4.3 Models for data with constant coefficient of

variation

4.3.1 Introduction

Data with constant coefficient of variation is well modeled by a Gamma distribution.

A random variable Y is said to have a Gamma distribution if its probability density

function is given by

f(y; µ, ν) =1

Γ(ν)(ν

µ)νyν−1exp(−νy

µ), y ≥ 0,

where µ and ν are unknown positive parameters. In the form of exponential family,

this density function is expressed as

f(y; µ, ν) = expy(−µ−1) − lnµ

ν−1+ [ν ln ν + (ν − 1) ln y − lnΓ(ν)].

86

Page 94: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

It can be identified that

θ = −µ−1, b(θ) = − ln(−θ), a(φ) = ν−1,

c(y, φ) = ν ln ν + (ν − 1) ln y − ln Γ(ν).

The cumulant generating function of Y is given by

κ(t) = −ν ln(1 − µt/ν).

The first four cumulants are

κ1 = µ,

κ2 = µ2/ν,

κ3 = 2µ3/ν2,

κ4 = 6µ4/ν3.

4.3.2 Components of models with weighted Gamma distri-

bution

The reasonable weight functions for weighted Gamma distributions are provided by

the power family:

w(y) = yα, α ≥ 0.

When α = 1 the weighted distribution becomes the length biased distribution, and

when α = 0 the weighted distribution reduces to the ordinary Gamma distribution.

The parameter function corresponding to the power weight function is given below:

ω(θ, ν) =Γ(ν + α)

Γ(ν)

ν

=Γ(ν + α)

Γ(ν)

(− 1

νθ

.

The bW function in the weighted distribution is given by

bW (θ, ν) = − ln(−θ) − ν−1α ln(−νθ) + lnΓ(ν + α)

Γ(ν).

87

Page 95: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

The derivatives of bW (θ, ν) are

∂bW (θ, ν)

∂θ= −1

θ

ν + α

ν=

ν + α

νµ = µW ,

∂2bW (θ, ν)

∂θ2=

1

θ2

ν + α

ν=

ν + α

νµ2.

The following link functions are usually used with Gamma distributions:

• Canonical link:

η1 = g1(µ) = −µ−1.

• Log link:

η2 = g2(µ) = ln(µ).

• identity link:

η3 = g3(µ) = µ.

The corresponding weighted response functions are given below:

hW1 (η1) = − 1

η1

ν + α

ν,

hW2 (η2) = eη2

ν + α

ν,

hW3 (η3) = η3

ν + α

ν.

4.3.3 Estimation of generalized linear additive models withweighted Gamma distributions

The data for generalized linear and additive models with weighted Gamma distri-

butions is of the form: (Yi, xi) : i = 1, . . . , n, where Y Zi is ascertained with a

weight function from a Gamma distribution with mean µi and dispersion parameter

ν, and xi = (xi1, . . . , xip) is the values of p predictor variables associated with the ith

subject.

When the dispersion parameter ν is known, Algorithm I and Algorithm 6 are

used for the estimation of generalized linear and additive models respectively, as in the

case of binomial and count data. However, most of the time the dispersion parameter

88

Page 96: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

is unknown. In this case the general algorithm, Algorithm II and Algorithm 7

are used for the generalized linear and additive models respectively. The following

updating formulae for the pseudo-response and weights are used in the algorithms:

zi = ηi +∂ηi

∂µWi

(yi − µWi )

= ηi +

⎧⎨⎩

ηi

(1 + ν

ν+αηi

), with canonical link,

νν+α

yie−ηi − 1, with log link,

νν+α

yi − ηi, with identity link.

wi =

(∂hW

k (ηi)

∂ηi

)2 (∂2bW (θi)

∂θ2i

)−1

=

ν+αν

1η2

i, k = 1, 3,

ν+αν

, k = 2.

Note that, with the considered weight function W (y) = yα, the weighted Gamma

distribution is still a Gamma distribution but with different parameters. Let Y W be

the weighted version of a Gamma random variable with parameter µ and ν. Indeed,

the probability density function of Y W is given by

P (Y W = y) =1

Γ(ν + α)

µ

)ν+α

yν+α−1e−νµ

y

=1

Γ(ν + α)

(ν + α

µ(ν + α)/ν

)ν+α

yν+α−1e−ν+α

µ(ν+α)/νy.

That is, Y W follows an ordinary Gamma distribution with parameters µW = µ +

(µα)/ν and νW = ν + α. However, the estimation for the weighted generalized

linear or additive models is more complicated than that for unweighted models, since

both µW and νW depends on a common parameter ν which is interwoven with the

estimation procedure.

89

Page 97: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 5

Comparison of Weithed andUnweighted Models by SimulationStudies

In practice, it is common that the bias in sampling caused by various weighted mech-

anisms is unduly neglected. It is then natural to ask what the consequence will be

by ignoring the bias in sampling. To investigate the effect of the bias in sampling

and evaluate the validity of the weighted models, we carried out simulation studies

to compare the weighted models with unweighted models when both are fitted to

data arising from weighted samplings. In this chapter, we report the results of these

simulation studies. The results on weighted generalized linear models are reported in

Section 5.1 and those on weighted generalized additive models are reported in Sec-

tion 5.2. In Section 5.3, we give general discussions based on the findings from the

simulation studies.

5.1 The Effect of Weighted Sampling on General-

ized Linear Models

In this section, we study the effect of weighted sampling on generalized linear mod-

els. The weighed binomial, weighted Poisson and Gamma distributions with various

weight functions are considered. The general methodology we adopted for this study

is as follows. First, the data are generated under the mechanism of a weighted sam-

90

Page 98: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

pling. Then both the weighted model and the corresponding unweighted model are

fitted to the generated data. The discrepancy between the fitted results are studied

to evaluate the effect of weighted sampling.

5.1.1 Studies on generalized linear models with weighted Bi-

nomial distributions

For the study of weighted binomial distributions, we considered a generalized linear

model with the following linear predictor:

η = β0 + β1X1 + β2X2,

where X1 and X2 are covariates distributed uniformly in the interval [1, 4], and the

true values of the coefficients are taken as β0 = 0, β1 = −1 and β2 = 1. The logit link

function was considered. In addition, the following weight function were considered:

1. W (y) = y.

2. W (y) = y2.

Three sample sizes, n = 50, 100, 200, were considered.

The case with weight function W (y) = y

Assuming that yi, i = 1, 2, · · · , n, follow binomial distributions B(mi, πi), and that

the observations yWi , i = 1, 2, · · · , n follow length biased binomial distributions, i.e.,

with weight function w(y) = y. As mentioned in the previous chapter that yWi − 1

follows an ordinary binomial distribution B(mi − 1, πi). Hence, yWi , i = 1, 2, · · · , n,

can be easily generated when mi and πi are given.

The indices mi’s are generated from a Poisson distribution with mean 6. The

simulated number which is either less than 2 or greater than 10 are ignored. The

reason we ignore the case mi = 1 is that, when mi = 1, the observation in length

biased sample contains no information about the parameter π because the distribution

of Y Wi is P (yW

i = 1) = 1 for any π.

91

Page 99: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Table 5.1: Comparison of weighted and unweighted generalized linear models withBinomial distributions and weight function w(y) = y

Weighted model Un-weighted modelTrue Average Empirical Average Empirical

std. std.n Para. value MLE of MLE MLE of MLE50 β0 0 0.005 0.573 0.456 0.799

β1 −1 −1.041 0.106 −0.785 0.096β2 1 1.029 0.077 0.771 0.100

100 β0 0 0.010 0.392 0.469 0.573β1 −1 −1.004 0.040 −0.801 0.066β2 1 1.005 0.035 0.783 0.071

200 β0 0 0.002 0.125 0.413 0.306β1 −1 −1.003 0.016 −0.781 0.060β2 1 1.001 0.017 0.783 0.062

For each of the sample sizes, the simulation was carried out 200 times. The

average estimated values of β0, β1 and β2 over the 200 simulations and their empirical

standard deviations when fitted with the weighted generalized linear model and the

corresponding unweighted model are given in Table 5.1.

We can see from Table 5.1 that the estimates are stabilized when n reaches 200,

judging by the small differences between the values when n = 100 and n = 200, and

that the estimates with the weighted model are essentially unbiased but those with

the unweighted model have significant biases.

The case with weight function W (y) = y2

When the weight function is given by W (y) = y2, the density function of the

weighted distribution is

Pr(yW = y) =y2(

my

)πy(1 − π)m−y

mπ(1 + (m − 1)π)

=(m− 1)π

1 + (m − 1)πB(y − 2, m − 2, π) +

1

1 + (m − 1)πB(y − 1, m − 1, π).

92

Page 100: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Table 5.2: Comparison of weighted and unweighted generalized linear models withBinomial distributions and weight function W (y) = y2

Weighted model Un-weighted modelTrue Average Empirical Average Empirical

std. std.n Para. value MLE of MLE MLE of MLE50 β0 0 −0.017 0.604 0.611 0.785

β1 −1 −1.016 0.202 −0.754 0.287β2 1 1.023 0.177 0.778 0.258

100 β0 0 0.017 0.418 0.644 0.733β1 −1 −1.012 0.144 −0.768 0.254β2 1 1.004 0.135 0.772 0.249

200 β0 0 0.005 0.339 0.672 0.727β1 −1 −1.002 0.100 −0.762 0.250β2 1 1.000 0.095 0.756 0.254

Therefore, the weighted distributions is a mixture of the two distributions B(y −2, m− 2, π)and B(y − 1, m− 1, π). Hence, yW

i , i = 1, 2, · · · , n can be easily generated

when mi and πi are given. Again, mis are generated from a Poisson distribution with

mean 6, but the simulated number which is either less than 2 or greater than 10 were

ignored from the same consideration as in the case W (y) = y. The results over 200

replicates of simulations are presented in Table 5.2.

The same features as in Table 5.1 can be observed from Table 5.2. While the

estimates with the weighted model remain essentially unbiased the biases of the es-

timates with the unweighted model become even larger. The other point to notice

is that, under both models, the variations of the estimates become larger compared

with the case W (y) = y. However, while the inflation of variation in the estimates

with the weighted model is slight and decreases when n becomes larger, those in the

estimates with the unweighted model is large and remains unchanged even when n is

large.

93

Page 101: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

5.1.2 Studies on generalized linear models with weighted Pois-

son distributions

For the study of weighted Poisson distributions, we considered the weight function

w(y) = 1 − (1 − β)y with three different β values, i.e., β = 0.2, 0.5, 0.9. The log link

function was considered.

When Yi follows a Poisson distribution with parameter λi, the weighted distribu-

tion of Y Wi with the above weight function is as follows.

P (yWi = y) =

(1 − (1 − β)ye−λλj

(1 − e−βλ)j!

The linear predictor takes the form

log(λ) = η = β0 + β1X1 + β2X2,

where X1 and X2 are uniformly distributed in the interval [2, 4], and the true values

of the coefficients are taken as β0 = 1, β1 = −1, and β2 = 1. The other settings of

the simulation are the same as before, that is, the simulation size is 200 and three

sample sizes, 50, 100, 200, are considered. The simulated results are summarized in

Table 5.3.

It can be observed again from Table 5.3 that the estimates with weighted models

are unbiased and that the estimates with unweighted models suffer significant biases.

There exists an obvious trend of the biases in Table 5.3 when the parameter β in

the weight function changes. As β changes towards 1, the biases become smaller.

This trend is expected since as β goes to 1, the weighted distribution tends to the

unweighted distribution. As noted previously, as β goes to 0, the weighted distribution

converges to the length-biased distribution with order 1. The simulation results here

suggest that the biases incurred by fitting a length-biased sample with the unweighted

model is serious.

94

Page 102: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 5.3. Comparison of weighted and unweighted generalized linear models with

Poisson distributions and weight function W (y) = 1 − (1 − β)y

weighted model un-weighted modelTrue Average Empirical Average Empirical

std. std.β n Para. value MLE of MLE MLE of MLE0.2 50 β0 1 1.086 1.024 1.351 0.741

β1 -1 -1.020 0.057 -0.806 0.069β2 1 0.991 0.039 0.784 0.081

100 β0 1 0.988 0.338 1.279 0.269β1 -1 -0.999 0.017 -0.775 0.066β2 1 1.003 0.016 0.781 0.063

200 β0 1 0.996 0.149 1.280 0.183β1 -1 -0.996 0.010 -0.762 0.065β2 1 0.996 0.008 0.768 0.062

0.5 50 β0 1 0.941 0.616 1.178 0.441β1 -1 -1.010 0.040 -0.777 0.078β2 1 1.024 0.044 0.792 0.078

100 β0 1 0.982 0.352 1.259 0.294β1 -1 -0.994 0.020 -0.804 0.057β2 1 0.999 0.017 0.793 0.057

200 β0 1 1.008 0.193 1.237 0.193β1 -1 -1.001 0.011 -0.785 0.055β2 1 0.997 0.009 0.786 0.055

0.9 50 β0 1 0.926 0.562 0.985 0.344β1 -1 -0.982 0.039 -0.853 0.055β2 1 1.005 0.023 0.901 0.031

100 β0 1 0.969 0.419 1.088 0.265β1 -1 -0.996 0.023 -0.814 0.054β2 1 1.004 0.018 0.843 0.041

200 β0 1 1.004 0.155 1.160 0.130β1 -1 -1.009 0.011 -0.836 0.037β2 1 1.005 0.010 0.835 0.037

95

Page 103: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Table 5.4. Comparison of weighted and unweighted generalized linear models with

Gamma distribution, log link and weight function W (y) = yα

weighted model un-weighted modelTrue Average Empirical Average Empirical

std. std.α n Para. value MLE of MLE MLE of MLE0.5 50 β0 0 0.024 0.268 0.169 0.316

β1 1 0.919 0.592 0.919 0.592β2 2 2.044 1.864 2.044 1.863ν 3 3.337 0.726 3.701 0.761

100 β0 0 0.014 0.211 0.162 0.211β1 1 1.004 0.375 1.004 0.375β2 2 1.924 1.209 0.924 1.209ν 3 3.209 0.503 3.665 0.582

200 β0 0 0.001 0.154 0.153 0.217β1 1 0.988 0.292 0.988 0.292β2 2 2.040 0.941 2.040 0.941ν 3 3.074 0.339 3.541 0.415

1 50 β0 0 -0.027 0.328 0.245 0.403β1 1 1.088 0.575 1.088 0.575β2 2 2.019 1.925 2.019 1.925ν 3 3.391 0.903 4.249 0.945

100 β0 0 0.009 0.180 0.287 0.337β1 1 1.014 0.397 1.014 0.397β2 2 1.929 1.118 1.929 1.118ν 3 3.210 0.605 4.121 0.646

200 β0 0 0.006 0.160 0.290 0.328β1 1 1.005 0.296 1.005 0.296β2 2 1.944 0.835 1.944 0.835ν 3 3.085 0.417 4.066 0.479

2 50 β0 0 0.027 0.264 0.503 0.551β1 1 1.022 0.486 1.022 0.486β2 2 1.966 1.427 1.966 1.427ν 3 3.575 1.218 5.348 1.249

100 β0 0 -0.018 0.189 0.484 0.511β1 1 1.033 0.327 1.033 0.327β2 2 2.107 1.047 2.107 1.047ν 3 3.201 0.729 5.161 0.749

200 β0 0 0.011 0.138 0.510 0.523β1 1 0.983 0.230 0.983 0.230β2 2 2.036 0.754 2.036 0.754ν 3 3.149 0.489 5.100 0.559

96

Page 104: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Table 5.5. Comparison of weighted and unweighted generalized linear models with

Gamma distribution, canonical link and weight function W (y) = yα

weighted model un-weighted modelTrue Average Empirical Average Empirical

std. std.α n Para. value MLE of MLE MLE of MLE0.5 50 β0 1 1.026 0.454 0.887 0.407

β1 1 1.034 0.880 0.894 0.770β2 2 1.901 2.840 1.635 2.457ν 3 3.352 0.827 3.715 0.883

100 β0 1 0.994 0.261 0.855 0.265β1 1 1.061 0.609 0.914 0.528β2 2 1.930 1.622 1.664 1.434ν 3 3.175 0.553 3.623 0.624

200 β0 1 1.020 0.194 0.875 0.206β1 1 0.960 0.431 0.824 0.407β2 2 1.988 1.229 1.709 1.099ν 3 3.075 0.378 3.551 0.406

1 50 β0 1 1.058 0.420 0.806 0.367β1 1 0.931 0.788 0.707 0.654β2 2 1.615 2.468 1.242 2.149ν 3 3.407 0.869 4.260 0.931

100 β0 1 0.998 0.277 0.752 0.326β1 1 1.039 0.542 0.779 0.460β2 2 1.992 1.849 1.492 1.471ν 3 3.144 0.573 4.054 0.610

200 β0 1 0.962 0.172 0.729 0.298β1 1 1.058 0.426 0.803 0.378β2 2 2.100 1.109 1.590 0.924ν 3 3.177 0.410 4.149 0.465

2 50 β0 1 0.980 0.337 0.610 0.434β1 1 0.921 0.741 0.571 0.615β2 2 2.104 2.062 1.312 1.461ν 3 3.664 1.147 5.382 1.138

100 β0 1 1.003 0.296 0.610 0.425β1 1 0.959 0.534 0.584 0.527β2 2 1.914 1.632 1.173 1.293ν 3 3.264 0.720 5.109 0.791

200 β0 1 1.020 0.176 0.615 0.397β1 1 0.943 0.394 0.567 0.489β2 2 1.934 1.066 1.167 1.045ν 3 3.119 0.516 5.061 0.589

97

Page 105: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

5.1.3 Studies on generalized linear models with weighted Gamma

distributions

For the study of weighted Gamma distributions, we consider the weight function

W (y) = yα with α = 0.5, 1 and 2. Note that, when W (y) = yα, the weighted Gamma

distribution is again a Gamma distribution with parameter (ν+αν

µ, ν+α) where (µ, ν)

are the parameters of the original Gamma distribution. Hence, the weighted sample

can be easily simulated from Gamma distributions when α, µ and ν are given. Two

link functions, log link and canonical link, are considered in the simulation.

For the Gamma models, we also consider the linear predictor of the form:

η = β0 + β1X1 + β2X2,

where X1 and X2 are uniformly distributed on the rectangle [0.05, 0.5] × [0.05, 0.2],

and the true values of the coefficients are taken as (β0, β1, β2) = (0, 1, 2) when the link

function is taken as the log link, and (β0, β1, β2) = (1, 1, 2) when the link function

is taken as the canonical link. The dispersion parameter ν is taken to be ν = 3 in

both cases. The results with log link and those with canonical link are summarized

in Table 5.4 and Table 5.5 respectively.

It is interesting to notice that, when the link function in the generalized linear

model is the log link, the estimates of β1 and β2 are exactly the same from both the

weighted and unweighted models. This is not accidental. Note that the mean of the

weighted Gamma distribution is given by µW = ν+αν

µ, where µ is the mean of the

original Gamma distribution. With the log link, the linear predictor η is related to

the mean of the original Gamma distribution by

ln(µ) = η = β0 + β1X1 + β2X2.

It is related to the mean of the weighted Gamma distribution by

ln(µW ) = ln(ν + α

ν) + η = β0 + β1X1 + β2X2.

That is, when treated as general generalized linear models, the weighted model and

the unweighted model have the same link function, and their linear predictors differ

only by the intercept term.

98

Page 106: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

While the estimates of β1, β2 are the same under both the weighted and unweighted

models, the estimates of β0 and ν differ under the two models. For the estimation of

the latter two parameters, the same feature we have observed for binomial and Poisson

models shows up. That is, the estimates under the weighted model are asymptotically

unbiased but those under the un-weighted model are significantly biased. As can be

expected, when α gets larger, the biases become more prominent, since the weighted

model becomes more distinguished from the un-weighted models.

Unlike in the case of the log link, the bias in the sampling affects the estimation

of all the parameters when the link is the canonical one. As the general feature,

the unbiasedness of the weighted model and the biasedness of the un-weighted model

present themselves in Table 5.5 again.

5.2 The Effect of Weighted Sampling on General-

ized Additive Models

In this section, we investigate the effect of bias in sampling on generalized additive

models by the same methodology used for studying generalized linear models. That

is, we generate data under the mechanism of weighted sampling but fit the data with

both weighted and unweighted models and then evaluate the discrepancies between

the two models. For the sake of convenience, we only consider length biased sam-

plings. The results with general weight functions will be similar. We consider two

distribution families for the study of generalized additive models: the family of bino-

mial distributions and the family of Gamma distributions. The generalized additive

models with length biased binomial distributions are treated in Section 5.2.1 and

those with length biased Gamma distributions are treated in Section 5.2.2.

5.2.1 Studies on generalized additive models with length bi-

ased Binomial distribution

For the length biased binomial distributions, we consider the weight function W (y) =

y, that is we consider the length biased distributions with order α = 1. The data

99

Page 107: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

are similarly simulated as in Section 5.1.1. The mi’s are generated from a Poisson

distribution with mean 6 and the simulated numbers which are less than 2 or greater

than 10 are discarded. For given mi and πi, the length biased observation Y Wi is

generated as y′i + 1 where Y ′

i follows B(mi − 1, π) distribution.

The additive prediction function is taken as follows.

η = 0.3(106(x111 )(1 − x1)

6 + 104(x31(1 − x1)

10)) − 2 + βx2,

where x1 is a continuous variable with range [0, 1] and x2 is a dummy variable taking

values 0 or 1, and the true value of β is taken as β = 0.5. The logit link is considered

in the generalized additive models.

Two sample sizes, n = 100 and 200, are considered in the simulation study. For

each sample size, half of the samples are generated for x2 = 0 and the other half

are generated for x2 = 1. For fixed x2 value, the values of x1 are generated from a

uniform distribution over interval [0, 1].

The simulation is repeated 200 times. To see the effects of the bias in sampling,

the average fitted mean curves over the 200 repetitions together with the empiri-

cal pointwise 2.5% and 97.5% quantiles under weighted and unweighted models are

plotted. Figures 2(a) and 2(b) plot the curves corresponding to x2 = 0 and x2 = 1

respectively when the sample size is 100. Figures 3(a) and 3(b) plot the curves cor-

responding to x2 = 0 and x2 = 1 respectively when the sample size is 200. In the

figures, the solid curves are the true mean curves derived from the additive prediction

function through the logit link. The dashed curves correspond to the weighted model

and the dotted curves correspond to the unweighted model.

It is obvious from these figures that the fitted mean curves with the unweighted

model are seriously biased. The bias is intrinsic. It does not reduce as n changes

from 100 to 200. The true curve even falls below the 2.5% quantile curve fitted with

the unweighted model at some points. However, the fitted curves with the weighted

model are essentially unbiased. When n changes from 100 to 200, the band flanked

by the two quantile curves shrinks to the true mean curve.

100

Page 108: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 2(a). Comparison of fitted mean curves together with the empirical 2.5% and

97.5% quantile curves under the unweighted and weighted model with length biased

binomial distribution when n = 100 and x2 = 0

The similar feature manifests itself in the estimation of β. When n = 100, the

estimated β using weighted model is 0.53 and that using un-weighted model is 0.43.

When n = 200, the estimated β using weighted model is 0.50 and that using un-

weighted model is 0.38. With the weighted model, there is a small bias 0.03 when

101

Page 109: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 2(b). Comparison of fitted mean curves together with the empirical 2.5%

and 97.5% quantile curves under the unweighted and weighted model with length

biased binomial distribution when n = 100 and x2 = 1

n = 100 and this bias reduces 0.0002 when n = 200. But the biases with the

unweighted model, 0.07 and 0.12, are large, which is expected since the bias with the

unweighted model is intrinsic.

102

Page 110: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 3(a). Comparison of fitted mean curves together with the empirical 2.5% and

97.5% quantile curves under the unweighted and weighted model with length biased

binomial distribution when n = 200 and x2 = 0

103

Page 111: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 3(b). Comparison of fitted mean curves together with the empirical 2.5%

and 97.5% quantile curves under the unweighted and weighted model with length

biased binomial distribution when n = 200 and x2 = 1

104

Page 112: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

5.2.2 Studies on generalized additive models with length bi-

ased Gamma Distribution

For the weighted gamma distributions, we consider the weight function

W (y) = y0.5.

The length biased observations are generated from the Gamma distribution with

parameters (ν+0.5ν

µ, ν + 0.5) for given µ and ν as in the case of generalized linear

models.

The following additive prediction function is taken,

η = ln((106(x111 )(1 − x1)

6 + 104(x31(1 − x1)

10)) + 0.1) + βx2,

where x1 is a continuous variable with range [0, 1] and x2 is a dummy variable taking

value of 0 and 1, and β is taken as 0.5, ν is taken as 3. Only the log link is considered

in this simulation study. The other settings of the simulation are the same as in the

case of length biased binomial distributions considered in the last sub-section.

The same types of curves as in the length biased binomial case are plotted. Figures

4(a) and 4(b) plot the curves corresponding to x2 = 0 and x2 = 1, respectively,

obtained from the generalized additive model with Gamma distributions described

above when the sample size is 100. Figures 5(a) and 5(b) plot those when the sample

size is 200.

By the same reason as in the generalized linear model case, when the log link is

used, the estimates of the parameter β are the same with weighted and unweighted

models. But the estimated mean curves and the dispersion parameter are different

with the weighted and unweighted models. The average of the estimates of β are 0.54

and 0.49 when n = 100 and n = 200 respectively. The average of the estimates of ν

with weighted model changes from 3.57 to 3.29, approaching the true value 3, when

n changes from 100 to 200. However, the averages with unweighted model, which are

2.62 and 2.56 when n = 100 and n = 200 respectively, remain far away from the true

value.

105

Page 113: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 4(a). Comparison of fitted mean curves together with the empirical 2.5% and

97.5% quantile curves under the unweighted and weighted model with length biased

gamma distribution with order α = 0.5 when n = 100 and x2 = 0

The same features of the fitted averaged mean curves and quantile curves as have

been observed in the case of length biased binomial distribution case show up again

in Figures 4(a), 4(b) and 5(a), 5(b). This again demonstrates adverse effects of the

sampling bias on generalized additive models when the sampling bias is ignored.

106

Page 114: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 4(b). Comparison of fitted mean curves together with the empirical 2.5%

and 97.5% quantile curves under the unweighted and weighted model with length

biased gamma distribution with order α = 0.5 when n = 100 and x2 = 1

107

Page 115: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 5(a). Comparison of fitted mean curves together with the empirical 2.5% and

97.5% quantile curves under the unweighted and weighted model with length biased

gamma distribution with order α = 0.5 when n = 200 and x2 = 0

108

Page 116: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Figure 5(b). Comparison of fitted mean curves together with the empirical 2.5%

and 97.5% quantile curves under the unweighted and weighted model with length

biased gamma distribution with order α = 0.5 when n = 200 and x2 = 1

109

Page 117: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 6

Concluding Remarks and OpenProblems

6.1 Conclusion

In this thesis, we have established a general theory on the generalized linear and

additive models with weighted exponential families. We discussed how to formulate

such models, developed a host of algorithms for the estimation of the models, touched

on some inferential issues and conducted simulation studies to illustrate the practical

importance of the weighted models.

Unlike the case of exponential family, the mean of a weighted exponential family

depends on dispersion parameter, so the estimate of the linear predictor or additive

predictor depends on the dispersion parameter. We distinguished two cases in our

development of computational algorithms: the dispersion parameter is known and

the dispersion parameter is unknown. If the dispersion parameter is known, our al-

gorithm for the computation of the model is essentially the same with that for an

ordinary generalized linear or additive model. For the weighted generalized linear

models the computation of the maximum likelihood estimates is reduced to an iter-

ated weighted least square procedure. For the weighted generalized additive models,

the computation is reduced to an iterated penalized least square procedure. If the

dispersion parameter is unknown, we developed double iterated procedures for the

computation of the models.

110

Page 118: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

For the weighted generalized linear models, we derived the asymptotic distribution

of the maximum likelihood estimates of the coefficients in the linear predictor. It

is multivariate normal but with different asymptotic variance-covariance matrices

depending on the dispersion parameter is known and unknown.

We demonstrated by simulation studies the necessity of the weighted models when

there is sampling bias. Our simulation results show that ignoring the sampling bias

causes serious biases in the estimation. These biases, which are intrinsic, do not

diminish as the sample size gets large. Inferences made from fitting an unweighted

model to data with weighted distribution could be very misleading. Therefore, it is

necessary and important to consider a weighted model rather than an unweighted

model when there is sampling bias.

6.2 Further Investigation

We focused in this thesis on the weighted distributions with a closed form parameter

function w(θ, φ) = E(Y |θ, φ). If the parameter function does not have a closed form,

the computation of the model will be much more intensive. More efficient algorithms

need to be developed to cater for this situation.

Another practical issue which needs to address further is the selection of the

weight function. The weight function is not always completely determined. In certain

situations, the weight function can only be determined up to an unknown parameter.

When this is the case, the first strategy might be to treat the unknown parameter in

the weight function as an additional parameter in the distribution family and estimate

it together with other parameters in the estimation procedure. This strategy needs

a more powerful computing facility to implement. Another possible strategy is to

sequentially try a few values of the unknown parameter coupled with the diagnostic

techniques briefly discussed in §2.5. There are still many other situations, even the

form of the weight function can not be determined from the scheme of ascertainment.

This makes the determination of the weight function even more difficult. A strategy

to deal with this situation is to try a few families of weight functions as listed in

111

Page 119: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Chapter 1 and select an appropriate one by certain criteria such as certain goodness-

of-fit measures. For example, the weight function can be selected by minimizing either

AIC or BIC score. The details of these strategies need to be further investigated.

Diagnostic techniques for the weighted models need to be further developed, es-

pecially, for checking the adequacy of the weight function. The effect of incorrectly

specifying a weight function also needs to be evaluated.

Although we touched on the asymptotic theory of the weighted generalized lin-

ear models, the asymptotic properties of the weighted generalized additive model

still need to be investigated. The case when the dispersion parameter is know is

relatively easy to handle. For the ordinary generalized additive models, Gu (2002)

showed that, under certain conditions, the penalized maximum likelihood estimate of

the additive predictor is consistent. Moreover, Stone (1986) demonstrated that the

additive components can be estimated with optimal rates of convergence. Since the

weighted generalized additive models with known dispersion parameter is essentially

the same as the ordinary generalized additive models, the result of Gu and Stone

continue to hold for the weighted models. However, when the dispersion parameter

is unknown, the picture is not clear. Further research is needed to deal with the

asymptotic properties of the weighted generalized additive models in this case.

112

Page 120: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Bibliography

Bayarri, M.J. and DeGroot, M.H. (1987). Information in selection models. In prob-

ability and Bayesian Statistics (ed. R. Viertl), 39-51.

Bayarri, M.J. and DeGroot, M.H. (1989). Comparison of Experiments with Weighted

Distributions. Statistical Data Analysis and Inference (ed Yaholah Dodge),

North Holland Publishing Co., 185-197.

Bhattacharyya, B.B, Franklin, L.A and Richardson, G.D. (1988). A comparison of

nonparametric unweighted and length-biased density of fibers. Comm. Statist.

A 17, 3629-44.

Breiman, L. and Friedman, J.H. (1985). Estimating optimal transformations for

multiple regression and correlation. J. Amer. Statist. Assoc. 80, 580-598 (with

discussion).

Chong Gu (1992). Cross-validating non-Gaussian data. J. Comput. Graph. Statist.

1, 179-179.

Chong Gu (2002). Smoothing Spline ANOVA Models, Springer-Verlag, New York.

Cook, R.C. and Martin, F.B. (1974). A model for quadrat sampling with visibility

bias. Journal of the American Statistical Association, 69, 345-349.

Cox, D.R. (1962). Renewal Theory. Methuen’s Monograph. Barnes and Noble, Inc.,

New York.

Cox, D.R. (1969). Some sampling problems in technology. In New Developments in

Survey Sampling, (ed. N.L. Jahnson and H. Smith), Wiley, New York. 506-27.

Craven, P. and Wahba, G. (1979). Smoothing noisy data with spline functions:

Estimating the correct degree of smoothing by the method of generalized cross-

validation. Number. Math. 31, 377-403.

de Boor, C. (1978). A Practical Guide to Splines, Springer-Verlag, New York.

113

Page 121: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Drummer, T.D. and McDonald, L.L (1987). Size-bias in line transect sampling,

Biometrics, 13, 13-21.

Fisher, R.A. (1934). The effects of methods of ascertainment upon the estimation

of frequencies. Annals of Eugenics, 6, 13-25.

Godambe, B.P. and Rajarhi, M.b. (1989). Optimal estimation for weighted distri-

butions: semi-parametric model. Statistical Data Analysis and Inference (ed.

Yadolah Dodge), North Holland Publishing Co. 199-208.

Green, P.J. and Silverman, B.W. (1994). Nonparametric Regression and Generalized

Linear Models. Chapman and Hall, London.

Haldane, J.B.S. (1938). The estimation of the frequency of recessive conditions in

man. Annals of Eugenics (London), 7, 255-262.

Hanis, C.L. and Chakraborty, R. (1984). Nonrandom sampling in human genetics:

Familial correlations. J. Math. Appl. Med. and Biol., 1, 193-213.

Hastie, T. and Tibshirani, R. (1990). Generalized Additive Model, Chapman and

Hall, London.

Johns, M.C. (1991). Kernel density estimation for length biased data. Biometrics,

78, 3, 511-519.

McCullagh, P. and Nelder, J.A. (1989). Generalized Linear Model, second edition,

Chapman and Hall.

Mahfoud, M. and Patil, G.P. (1981). Size biased sampling, weighted distributions,

and Bayesian estimation. Statistics in Theory and Practice: Essays in Honor

of Bertil Matern, (ed. B. Ranneby), Swedish University Agricultural Science,

Umea, Sweden, 173-187.

Mahfoud, M. and Patil, G.P. (1982). On weighted distributions. Statistics in Theory

and Practice: Essays in Honor of C.R.Rao, (eds. G. Kallianpur et al.), North-

Holland, Amsterdam, 479-492.

114

Page 122: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Patil, G.P. and Rao, C.R. (1977). Weighted distributions: A survey of ascertain-

ment: what population does a sample represent? A celebration of Statistic,

the ISI Centenary Volume. (eds. Anthony C. Aitkinson and S.E. Fienberg),

Springer-Verlag, 543-569.

Patil, G.P. and Rao, C.R. (1978). Weighted distributions and size-biased sampling

with applications to wildlife populations and human families. Biometrics 34,

179-89.

Patil, G.P. and Ord, J.K. (1975). On size-biased sampling and related form-invariant

weighted distributions. Sankhya, 38, 48-61.

Patil, G.P., Rao, C.R. and Ratnaparkhi, M.V. (1986). On discrete weighted dis-

tributions and their use in model choice for observed data. Comm. Statist.

-Theor. Meth., 15(3), 907-918,

Patil, G.P., Rao, C.R. and Zelen, M. (1988). Weighted distributions. In Encyclo-

pedia of Statistical Sciences, Vol. 9, (eds. Kotz, S. and Johnson, N.L.), John

Wiley, New York, 565-571.

Patil, G.P. and Taillie, C. (1989). Probing encountered data, meta analysis and

weighted distribution methods. Statistical Data Analysis and Inference (ed.

Yaholah Dodge), North Holland Publishing Co.,317-45.

Patil, G.P. (2002). Weighted distributions. Encyclopedia of Environmetrics, Vol. 4,

(eds. Abdel H. El-Shaarawi and Walter W. Piegorsch), John Wiley and sons,

Chichester, 2369-2377.

Rao, C. R. (1965). On discrete distributions arising out of methods of ascertainment.

In Classical and Contagious Discrete Distributions, (ed. G.P. Patil), Pergamon

Press and Statistical Publishing Society, Calcutta, 320-332.

Rao, C.R. (1985). Weighted distribution arising out of methods of ascertainment:

What population does a sample represent?. In A Celebration of Statistics, (eds.

Atkinson, A.C. and Fienberg, S.E.), Springer-Verlag, New York, 543-569.

115

Page 123: GENERALIZED LINEAR AND ADDITIVE MODELS WITH WEIGHTED ... · generalized linear and additive models with weighted distribution shen liang (m.sc., national university of singapore )

Sen, P.K. (1987). What do the arithmetic, geometric and harmonic means tell us in

length-biased sampling. Statistical and Probability Letters, 5, 95-98.

Stone, C.J. (1986). The dimensionality reduction principle for generalized additive

models. Ann. Statist. 14, 590-606.

Stone, C.J., Hansen, M.H., Kooperberg, C., and Truong, Y.K. (1997). Polynomial

splines and their tensor products in extended linear modeling. Ann. Statist.

25, 1371-1470 (with discussions).

Wahba, G. (1990). Spline Models for Observational Data, Volume 59 of CBMS-NSF

Regional Conference Series in Applied Mathematics. Philadelphia: SIAM.

Weinet, H.L. (1982). Reproducing Kernel Hillbert Spaces: Applications in Statistical

Signal Processing. Hutchinson Ross, New York.

Xiang, D. and Wahba, G. (1996). A generalized approximate cross validation for

smoothing splines with non-Gaussian data. Statist. Sin. 6, 675-792.

Zelen, M. (1969). On the theory of screening for chronic disease. Biometrika, 56,

601-614.

Zelen, M. (1971). problems in the early detection of disease and the finding of faults.

Bull. Int. Statist. Inst., Pros. 38, Session I, 649-661.

Zelen, M. (1974). Problems in Cell Kinetics and Early Detection of Disease. Relia-

bility and Biometry, SIAM, Philadelphia, 701-726.

116