13
Research Article Received 8 March 2011, Accepted 26 July 2011 Published online 14 October 2011 in Wiley Online Library (wileyonlinelibrary.com) DOI: 10.1002/sim.4366 Simple estimation of hidden correlation in repeated measures Thuan Nguyen a * and Jiming Jiang b In medical and social studies, it is often desirable to assess the correlation between characteristics of interest that are not directly observable. In such cases, repeated measures are often available, but the correlation between the repeated measures is not the same as that between the true characteristics that are confounded with the measurement errors. The latter is called the hidden correlation. Previously, the problem has been treated by assuming prior knowledge about the measurement errors or by using relatively complex statistical models, such as the mixed-effects models, with no closed-form expression for the estimated hidden correlation. We propose a simple estimator of the hidden correlation that is very much like the Pearson correlation coefficient, with a closed-form expression, under assumptions much weaker than the mixed-effects model. Simulation results show that the proposed simple estimator performs similarly as the restricted maximum likelihood (REML) estimator in mixed models but is computationally much more efficient than REML. We also made simulation comparison with the Pearson correlation. We considered a real data example. Copyright © 2011 John Wiley & Sons, Ltd. Keywords: correlation coefficient; hypothesis testing; repeated measures 1. Introduction In medical and social studies, it is often desirable to assess the correlation between characteristics of interest in the presence of repeated measures. For example, the repeated measures may be obtained under different circumstances, such as (i) by different observers (e.g., doctors, nurses, technicians); (ii) at dif- ferent occasions (e.g., times, visits); or (iii) from different individuals sharing the same characteristics. If the characteristics of interest are directly observed, a well-known simple measure of the correlation between them is the Pearson correlation, also known as the correlation coefficient (e.g., [1]), computed by r D P n i D1 .x i N x/.y i N y/ q P n i D1 .x i N x/ 2 q P n i D1 .y i N y/ 2 ; (1) where .x i ;y i / is the pair of observations from subject i , i D 1;:::;n, N x D n 1 P n i D1 x i , and N y D n 1 P n i D1 y i . Here, r is considered an estimator of the population correlation coefficient between x and y . However, in many cases, the variables of interest are not directly observable; instead, repeated, ‘indirect’ measures of these variables are available. It should be noted that the term ‘repeated mea- sures’ should be understood in a broader sense as any observation that involves errors so that the true characteristic of interest is not observed. We use a real data example for illustration. Example. Here, we consider a daily study conducted by Dr. Ryan Olson and his colleagues at Oregon Health and Science University to measure home care worker’s exposure levels to demanding tasks and also to understand how work exposure might relate to daily physical symptoms, wellness, and lifestyle behaviors. Twenty-three home care workers participated in the study. After undertaking a baseline demo- graphic and psychosocial survey, the participants completed 10–14 days of daily self-monitoring by a Department of Public Health and Preventive Medicine, Oregon Health and Science University, Portland, OR, U.S.A. b Department of Statistics, University of California, Davis, Davis, CA, U.S.A. *Correspondence to: Thuan Nguyen, Department of Public Health and Preventive Medicine, Oregon Health and Science University, Portland, OR, U.S.A. E-mail: [email protected] Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415 3403

Simple estimation of hidden correlation in repeated measures

Embed Size (px)

Citation preview

Page 1: Simple estimation of hidden correlation in repeated measures

Research Article

Received 8 March 2011, Accepted 26 July 2011 Published online 14 October 2011 in Wiley Online Library

(wileyonlinelibrary.com) DOI: 10.1002/sim.4366

Simple estimation of hidden correlationin repeated measuresThuan Nguyena*† and Jiming Jiangb

In medical and social studies, it is often desirable to assess the correlation between characteristics of interest thatare not directly observable. In such cases, repeated measures are often available, but the correlation betweenthe repeated measures is not the same as that between the true characteristics that are confounded with themeasurement errors. The latter is called the hidden correlation. Previously, the problem has been treated byassuming prior knowledge about the measurement errors or by using relatively complex statistical models, suchas the mixed-effects models, with no closed-form expression for the estimated hidden correlation. We proposea simple estimator of the hidden correlation that is very much like the Pearson correlation coefficient, with aclosed-form expression, under assumptions much weaker than the mixed-effects model. Simulation results showthat the proposed simple estimator performs similarly as the restricted maximum likelihood (REML) estimatorin mixed models but is computationally much more efficient than REML. We also made simulation comparisonwith the Pearson correlation. We considered a real data example. Copyright © 2011 John Wiley & Sons, Ltd.

Keywords: correlation coefficient; hypothesis testing; repeated measures

1. Introduction

In medical and social studies, it is often desirable to assess the correlation between characteristics ofinterest in the presence of repeated measures. For example, the repeated measures may be obtained underdifferent circumstances, such as (i) by different observers (e.g., doctors, nurses, technicians); (ii) at dif-ferent occasions (e.g., times, visits); or (iii) from different individuals sharing the same characteristics.If the characteristics of interest are directly observed, a well-known simple measure of the correlationbetween them is the Pearson correlation, also known as the correlation coefficient (e.g., [1]), computed by

r D

PniD1.xi � Nx/.yi � Ny/qPn

iD1.xi � Nx/2

qPniD1.yi � Ny/

2

; (1)

where .xi ; yi / is the pair of observations from subject i , i D 1; : : : ; n, Nx D n�1PniD1 xi , and

Ny D n�1PniD1 yi . Here, r is considered an estimator of the population correlation coefficient between

x and y. However, in many cases, the variables of interest are not directly observable; instead, repeated,‘indirect’ measures of these variables are available. It should be noted that the term ‘repeated mea-sures’ should be understood in a broader sense as any observation that involves errors so that the truecharacteristic of interest is not observed. We use a real data example for illustration.

Example. Here, we consider a daily study conducted by Dr. Ryan Olson and his colleagues at OregonHealth and Science University to measure home care worker’s exposure levels to demanding tasks andalso to understand how work exposure might relate to daily physical symptoms, wellness, and lifestylebehaviors. Twenty-three home care workers participated in the study. After undertaking a baseline demo-graphic and psychosocial survey, the participants completed 10–14 days of daily self-monitoring by

aDepartment of Public Health and Preventive Medicine, Oregon Health and Science University, Portland, OR, U.S.A.bDepartment of Statistics, University of California, Davis, Davis, CA, U.S.A.*Correspondence to: Thuan Nguyen, Department of Public Health and Preventive Medicine, Oregon Health and ScienceUniversity, Portland, OR, U.S.A.

†E-mail: [email protected]

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3403

Page 2: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

using the safety task assessment tool developed by Dr. Olson’s research group. It was noted that the par-ticipants might meet different clients on those days. The work task survey focuses on measuring time andfrequency exposure to demanding tasks and was completed by the workers during their natural breaksat work. The end-of-day survey focuses on recording physical, behavioral, and psychosocial symptomsthat could be affected by the work task exposure and was completed in the evening between dinner andbedtime. In short, the daily data include the 24 variables recording the counts and times spent on varioustasks. The correlations between these variables were of interest, for example, that between daily occu-pational fatigue and daily negative affect. However, these variables were recorded on the basis of theworkers’ estimates, or memories, say, during the natural breaks. Such estimates might not be accurate.Besides, there is variation due to the occasion (i.e., day) that exits even if the workers report their dailyactivities accurately. In other words, the true characteristics of interest (e.g., daily occupational fatiguestress and daily negative affect) were measured with combinations of different sources of errors.

Nevertheless, with the repeated measures, we may be able to recover the true correlation between thecharacteristics of interest. Let .xi ; yi / denote the pair of unobserved characteristics of interest for subjecti , and let .xis; yis/; s D 1; : : : ; ni be the repeated measures of .xi ; yi /. Imagine that xis is a combinationof the true value xi plus a measurement error (which combines different sources of variations); that is,xis D xi C �is , where �is is the measurement error for xi . Similarly, yis D yi C �is , and �is is the mea-surement error for yi . For example, in the aforementioned example, xi is the mean daily occupationalfatigue, and yi is the mean daily negative affect for worker i . Our interest is �h D cor.xi ; yi /, the pop-ulation correlation coefficient for the pair of characteristics across all the workers (here, the subscript hrefers to ‘hidden correlation’; see in the succeeding discussion). However, this correlation is confoundedwith the measurement errors. For example, the latter often add noise to the ‘signals’ (i.e., xi and yi ),making one think that the signals are weakly correlated. In other words, the correlation between the sig-nals is hidden, at least to some extent, by the noise. In fact, even if the measurement errors �is and �is areuncorrelated, the correlation between xi and yi is not the same as that between the repeated measures,xis and yis , but greater in absolute value than the latter, because of the variations of the measurementerrors (see Sections 2 and 4 for more details). It should be pointed out that the repeated measures do notcreate the hidden correlation problem; on the contrary, they provide a way to solve the problem. In fact,the hidden correlation can never be recovered without the repeated measures.

Also note that the problem considered here is different from the one regarding the ICC (e.g., [2]).Fisher [3] originally proposed the latter to assess consistency, or conformity, of measurements made bymultiple observers measuring the same quantity [4]. The following example may help to highlight thedifference. Several medical laboratories examined each blood sample drawn from a patient with autismto obtain measures of total cholesterol and 7-dehydrocholesterol. If one wants to know the consistencyof the medical laboratories on their cholesterol measurements (over the population of all the patients),one has an ICC problem. On the other hand, if one wants to know the population correlation (across allthe patients) between cholesterol and 7-dehydrocholesterol, one has a hidden correlation problem.

Some studies involve more than two variables (e.g., the aforementioned example on home care work-ers). If one is interested in the correlations between these variables, by considering one pair of variableseach time, these correlations can be estimated, provided that the aforementioned problem regarding thehidden correlation between xi and yi can be solved.

The hidden correlation problem has been studied in the literature. In most cases, it has been assumedthat the measurement errors, �is and �is , are uncorrelated. For example, ‘case 3’ in [5, pp. 217–218]assumed that measurement errors are independent and normally distributed. However, the author did notprovide a satisfactory solution to the problem, noting the ‘difficulty’ due to the non-observability of thevariables of interest. Following the same assumptions, Horn and Guiard [6] considered the problem asthat of estimating the variance components, namely the variances of x; y; �; �, and �h. They used theanalysis of variance method to estimate the variances and then a formula simplified by the assumptionc�� D 0 (see (3)) to estimate �h. Rosner and Willett [7] used the simplified connection (4), with c�� D 0,and assumed, in addition, that cxx; cyy ; c�� ; c�� are obtained from ‘reproducibility studies’ (see also[8,9] for regression error-in-variable settings). Unlike the aforementioned study, Schaalje and Butts [10]noted that it is sometimes unreasonable to assume that the measurement errors are uncorrelated, and theeffects of ignoring the correlation. They proposed a method without assuming that c�� D 0, but theyrequired that estimates of c�� , c�� , and c�� are ‘obtained elsewhere’; in other words, knowledge aboutthe variances and correlations of the measurement errors is available (see also Thoresen and Laake [11]who considered the regression error-in-variable setting, with c�� ¤ 0, but the focus was estimation ofthe regression coefficients).

3404

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 3: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

A ‘modern’ approach to estimation of the hidden correlation is to use a mixed-effects model (e.g.,[12]). For the most part, under a mixed-effects model, xi and yi are treated as random effects whose cor-relation �h is part of the so-called variance components. Thus, by estimating the variance componentsusing, say, the restricted maximum likelihood (REML; e.g., [13]) method, one obtains an estimate of �h

(see Section 4 for more details). A procedure based on the mixed-effects model approach is availablein PROC MIXED in SAS [14]. However, to use this procedure to estimate the hidden correlation, oneneeds to know the mixed-effects models very well to understand which particular model should be used.This is often not the case to non-statistician users, not to mention the methodology background behind it.As many medical or social researchers are more familiar with the Pearson correlation, it might be moreappealing to these researchers if an estimate of the hidden correlation can be computed in a way similarto the Pearson correlation, in particular, with a closed-form expression. Furthermore, the mixed-effectsmodel approach is based on a highly parametric model involving the normality assumption, which maynot be satisfied in practice. For example, in many medical or social studies, the repeated measures arediscrete (e.g., counts, integer-valued levels, as in the aforementioned example on home care workers).Such data are obviously not normal, so the normality-based mixed-effects model is questionable. Also,as noted by Schaalje and Butts [10], it is impractical to assume that c�� D 0 in many cases. For example,if the repeated measures are carried out by different observers (e.g., doctors, nurses), there may be acorrelation between the measurement errors, �is and �is , as they are due to the same technician. There-fore, we are not going to assume that c�� D 0, or any knowledge about the variance/covariance of themeasurement errors, as in the previous works (with the exception of the mixed-effects model approach),and still obtain a simple estimator of the hidden correlation as well as its standard error. These are someof the important features of our work.

We propose a simple model in the next section and derive a closed-form estimator of the hidden cor-relation �h in a way very much like (1) under this model. The derivation shows, in particular, that blindlyusing the Pearson correlation and ignoring the difference between the true (hidden) correlation and thecorrelation between the observations can lead to incorrect, or inaccurate, estimate of the true correla-tion. In Section 3, we derive large sample tests for the hidden correlation, which are often of interestin medical/social studies. In Section 4, we carry out simulation studies to evaluate the performance ofthe proposed estimator in terms of both estimation and testing. The simulation studies also include com-parisons of our method with the Pearson correlation, and REML based on a linear mixed model. Thesimulation results support the theoretical finding that blindly using the Pearson correlation can be seri-ously misleading. Furthermore, the simulation results show that our method performs similarly to theREML method in terms of estimation and testing performance, especially when the sample size is rel-atively large, but (our method) is computationally much more efficient than REML. In Section 5, werevisit the aforementioned example on the home care worker study. We highlight the main results inSection 6 and offer discussions on the potential use of bootstrap and other alternative methods. We defertechnical details in Appendix A.

2. A statistical model and derivation of estimator

Suppose that xis; yis; i D 1; : : : ; mI s D 1; : : : ; ni are observed, where ni > 1, such that

xis D xi C �is; yis D yi C �is; (2)

where i represents the subject so that .xis; yis/ is the sth repeated pair of measures from subject i and niis the number of repeated measures from subject i , which is allowed to depend on i . Furthermore, xi ; yirepresent the characteristics of interest that are not observable, but do not depend on s, and �is; �is arethe random fluctuations, or errors, due to the repeated measures. The following assumptions are made.

A1. The vectors .xi ; yi /; i D 1; : : : ; m are independent and identically distributed (i.i.d.).A2. The vectors .�is; �is/; i D 1; : : : ; mI s D 1; : : : ; ni are i.i.d. with mean 0.A3. For each i , .xi ; yi /, and .�is; �is/; s D 1; : : : ; ni are uncorrelated.

Assumption A1 is the basis for the population correlation coefficient to be well defined. AssumptionA2 is reasonable if there is no bias against any subject or any repeated measure within the subject. Forexample, if the pairs .�is; �is/ are viewed as random samples from the population of pairs of measure-ment errors for all the medical workers, then assumption A2 is expected to be satisfied. Assumption A3may be interpreted as ‘blindness’ of the measurement errors irrespective of the subject.

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3405

Page 4: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

Our interest is �h D cor.xi ; yi / D cxy=pcxxcyy , where cxy D cov.xi ; yi /, cxx D var.xi /, and

cyy D var.yi /. Note that the correlation �h, the covariance cxy , and the variances cxx and cyy do notdepend on i , because the pairs .xi ; yi / are i.i.d. (assumption A1); in other words, these are populationparameters. Then, we have

cov.xis; yis/D cov.xi ; yi /C cov.�is; �is/

D cxy C c��;(3)

by A1–A3. Equation (3) shows that the covariance between the repeated measures is not the same asthe one between the two variables of interest but includes another source of covariance (the covari-ance between the measurement errors). By similar arguments, one has var.xis/ D cxx C c�� andvar.yis/D cyy C c�� . Thus, we have

cor.xis; yis/Dcxy C c��

pcxx C c��

pcyy C c��

: (4)

The right side of (4), denoted by �, is what one is actually estimating if the Pearson correlation, r , isblindly used. Note that � is most likely not equal to �h, the true correlation that we are interested in.For example, suppose that all the variances are equal to 1, cxy D 0:9, and c�� D 0:1. Then, we have�h D 0:9, while � D .0:9C 0:1/=2 D 0:5. Another special case is when there is no correlation between� and �; that is, c�� D 0. In this case, it is easy to see that j�j is strictly less than j�hj. Thus, in the lattercase, the (blind) Pearson correlation is expected to underestimate in absolute value the true correlation.

On the other hand, by (2), we have

Nxi � D xi C N�i �; Nyi � D yi C N�i �;

where Nxi � D n�1iPnisD1 xis and so on, which imply that cov. Nxi �; Nyi �/D cxy C cov. N�i �; N�i �/ by A1 and A3.

Note that cov. N�i �; N�i �/D n�2iPnis;tD1 cov.�is; �it /D n�2i

PnisD1 cov.�is; �is/D n�1i c�� by A2. Thus, we

have cov. Nxi �; Nyi �/D cxy C c��=ni , or

c�� D nifcov. Nxi �; Nyi �/� cxyg: (5)

By bringing (5) to (3), we obtain

cxy Dni

ni � 1cov. Nxi �; Nyi �/�

cov.xis; yis/

ni � 1: (6)

Because (6) holds for any s D 1; : : : ; ni , by taking the average over s, we have

cxy Dni

ni � 1cov. Nxi �; Nyi �/�

1

ni .ni � 1/

niXsD1

cov.xis; yis/

Dni

ni � 1E. Nxi � Nyi �/�

ni

ni � 1E. Nx1�/E. Ny1�/�

1

ni � 1cov.x11; y11/;

(7)

because A1–A3 imply that E. Nxi �/D E. Nx1�/, E. Nyi �/D E. Ny1�/, and cov.xis; yis/D cov.x11; y11/. Finally,because (7) holds for every i , by taking the average over i , we get

cxy D1

m

mXiD1

ni

ni � 1E. Nxi � Nyi �/�

1

m

mXiD1

ni

ni � 1

!E. Nx1�/E. Ny1�/

1

m

mXiD1

1

ni � 1

!cov.x11; y11/:

(8)

The first term on the right side of (8) can be expressed as

E

1

m

mXiD1

ni

ni � 1Nxi � Nyi �

!;

3406

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 5: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

which has the following consistent estimator (by removing the expectation sign):

s1 D1

m

mXiD1

ni

ni � 1Nxi � Nyi �:

On the other hand, consistent estimators of E. Nx1�/, E. Ny1�/, and cov.x11; y11/ are s2 D m�1PmiD1 Nxi �,

s3 Dm�1PmiD1 Nyi �, and

s4 D1

m

mXiD1

1

ni

niXsD1

xisyis � s2s3;

respectively. Thus, a consistent estimator of cxy is obtained by

Ocxy D s1 �

1

m

mXiD1

ni

ni � 1

!s2s3 �

1

m

mXiD1

1

ni � 1

!s4

D s1 � s2s3 �

1

m

mXiD1

1

ni � 1

! 1

m

mXiD1

1

ni

niXsD1

xisyis

!:

(9)

Similar to (9), Ocxx and Ocyy , the estimators of cxx and cyy , respectively, are obtained (with all the lettery replaced by x for cxx , and the other way around for cyy ; see Appendix A). Therefore, a consistentestimator of �h is obtained as

rh D Ocxy=qOcxx Ocyy : (10)

3. p-value in testing hypothesis

In some cases, the interest is whether the correlation between the characteristics of interest is ‘signifi-cant’. This is related to testing the hypothesis H0 W �h D 0 versus Ha W �h ¤ 0 (or �h > 0; �h < 0).First, note that �h D cor.xi ; yi /D 0 is equivalent to cxy D cov.xi ; yi /D 0. Thus, the null hypothesis isequivalent to

H0 W cxy D 0: (11)

We consider the following t -statistic

tc DOcxy

s:e:. Ocxy/: (12)

Under regularity conditions, (12) has an asymptotic standard normal distribution. Unlike the Pearsoncorrelation whose test statistic, r=

p.n� 2/�1.1� r2/, has a student-t distribution with (n � 2) DOF

under the null hypothesis that � D 0, if the data are bivariate normal (e.g., [15]), the normal or´-distribution is used to determine the critical value for (12). This is more attractive because in the pres-ence of repeated measures and, possibly, non-normal data, it is difficult to determine the effective DOFas for testing the Pearson correlation. Note that the derivation of the (exact) t -distribution is based on theassumption of independent normal observations. The current case is much more complicated with cor-relation among the repeated measures (even under the null hypothesis). Thus, the (exact) t -distributionresult no longer holds, even if the data is normal. Furthermore, we consider situations where the datamay not be normal, so an exact t -distribution under the null hypothesis cannot be expected. As notedearlier, our case may be viewed as a special case of the linear mixed models. In general, it is difficult toderive an exact t , �2, or F distribution for a testing problem in linear mixed models with unbalanceddata (and hence determine the corresponding DOF), even if the data is normal (e.g., [13, Section 2.1];see Section 6 for further discussion on the p-value).

Thus, the p-value for testing (11) can be obtained, provided that we can determine the denominatorof (12). Define u1 D s2; u2 D s3; u3 D s1, and

u4 D1

m

mXiD1

1

ni

niXtD1

xisyis :

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3407

Page 6: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

Then, we have Ocxy D u3 � u1u2 � c1u4 D g.u1; u2; u3; u4/, where

c1 D1

m

mXiD1

1

ni � 1:

Using the delta method (e.g., [16]), we obtain

var�Ocxy��@g

@u0Var.u/

@g

@u; (13)

where uD .u1; u2; u3; u4/0. Furthermore, we have

@g

@uD .�u2;�u1; 1;�c1/

0 (14)

(and @g=@u0 is the transpose of @g=@u). An estimate of Var.u/, OV , is derived from Appendix A. Bycombining these results, we obtain

s:e:. Ocxy/Dq

the right side of .13/ with Var.u/ replaced by OV : (15)

More generally, one may be interested in testing whether the correlation is at or below a given thresh-old; that is, H0 W �h D �h0 versus Ha W �h ¤ �h0 (�h > �h0; �h < �h0/, where �h0 is a given number. Theaforementioned approach, unfortunately, does not apply to this case (note that the equivalence of �h D 0to cxy D 0 simplifies the problem, but there is no such simplification when �h ¤ 0). A large samplet -test for the general hypothesis is based on

tr Drh � �h0

s:e:.rh/; (16)

where s:e:.rh/ D Ov and the expression of Ov2 is given in Appendix A. To get some idea about the accu-racy of this s.e. estimator, or equivalently that of Ov2 as the variance estimator, some small simulationsare run under the scenario described in the first paragraph of the next section. Note that Ov2 is supposedto satisfy d D mfE. Ov2/ � var.rh/g ! 0 as m ! 1. On the basis of K D 1000 simulation runs, wehave d D �0:202 for m D 50, and d D 0:057 for m D 100. The asymptotic distribution of tr underthe null hypothesis is, again, the standard normal distribution. The approximate power of the test canbe obtained by the standard derivation. For example, for testing H0 W �h D �h0 against Ha W �h > �h0,let �h be the true hidden correlation, which is greater than �h0. Then, at the ˛-level of significance, thepower of test (16) at the alternative �h is approximately 1�ˆ.´˛ � .�h � �h0/=s:e:.rh//, where ˆ.�/ isthe cumulative standard normal distribution function and ´˛ is the corresponding ˛ critical value. Theassociated 100.1� ˛/% confidence interval for �h is

Œrh � ´˛=2s:e:.rh/; rhC ´˛s:e:.rh/�: (17)

More simulation results are presented in the next section.

4. Simulation studies

The simulations are carried out under the following setting: E.xi /D E.yi /D 0, var.xi /D var.yi /D 1,cor.xi ; yi / D �h, E.�is/ D E.�is/ D 0, var.�is/ D var.�is/ D 1, and cor.�is; �is/ D ı. Unless specif-ically indicated, data are generated under the normal distribution. First, consider estimation of �h withı D 0:1. We consider three different true values of �h: 0.3, 0.6, and 0.9. In each case, three differentsample sizes are considered: m D 50, m D 100, and m D 200. In all cases, ni D 5; 1 6 i 6 m. InTable I, we report the Monte Carlo (MC) mean and standard error (s.e.) of the proposed estimator, rh,on the basis of K D 1000 simulation runs. Note that each mean reported in Table I is the average ofK MC simulation results (as an approximation to the corresponding true expectation of the estimator).Thus, the MC s.e. (of the MC mean) is the sample standard deviation of the K results (MC s.d.) dividedbypK. So, in particular, from the MC s.e., one can easily obtain the MC s.d. of the estimator. Overall,

the estimator appears to perform very well and show sign of consistency as m increases.The next simulation shows how blindly using the Pearson correlation can be misleading. Because all

the variances are equal to one in our simulations, according to (4), the Pearson correlation coefficient is

3408

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 7: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

Table I. Monte Carlo mean and standard error (s.e.) of estimator.

�h D 0:3 �h D 0:6 �h D 0:9

m Mean s.e. Mean s.e. Mean s.e.

50 0.301 0.005 0.600 0.004 0.907 0.002100 0.296 0.004 0.598 0.003 0.903 0.001200 0.298 0.003 0.600 0.002 0.900 0.001

expected to converge (e.g., in probability) to .cxy C c��/=2 D �. The second row of Table II shows thevalues of � when the true value of �h (D cxy in this case) is 0:6 and when the true value of ı (D c�� inthis case) is �0:4, �0:2, 0, 0:2, 0:4, and 0:6, respectively. It is seen that � is not equal to �h even if ı D 0,that is, when the measurement errors �is and �is are uncorrelated. In fact, the only case that � D �h iswhen the correlation between � and � is the same as that between x and y, that is, 0:6. Next, we carryout a simulation study with m D 100, and again, ni D 5; 1 6 i 6 m. Here, the rows E.rh/ and E.r/correspond to the MC means (MC s.e. in the next row) of rh and r , respectively, where r is the (blind)Pearson correlation coefficient. Again, the results reported in (the last four rows of) Table II are basedon K D 1000 simulation runs. It is seen that the proposed estimator, rh, of the true hidden correlationis doing (much) better for all except the last case, where � D 0:6, compared with the blind Pearsoncorrelation coefficient, r (note that the smaller MC s.e. of r does not really matter when the estimatoris seriously biased). In the last case, of course, r is doing better, but this depends on some luck (that ıhappens to be 0.6 so that � is equal to �h); one cannot count on this in practice.

As noted (Section 1), the hidden correlation may be estimated by assuming a linear mixed-effectsmodel. Write ˛i D .xi ��x; yi ��y/0, where �x D E.xi / and �y D E.yi /. Then, (2) can be expressedas Yis D � C ˛i C �is , i D 1; : : : ; m; s D 1; : : : ; ni , where Yis D .xis; yis/

0, � D .�x; �y/0, and

�is D .�is; �is/0. The latter may be viewed as a linear mixed model (e.g., [13]) with the ˛i ’s being the

random effects. Furthermore, assume that ˛1; : : : ; ˛m are independent and bivariate normal with meanszero, variances 21 ;

22 , and correlation coefficient �h and that the �is’s are independent and bivariate

normal with means zero, variances 21 ; 22 , and correlation coefficient ı. Then, 21 , 22 , �h, 21 , 22 , and ı

are unknown variance components. In the current case, because ni D 5; 1 6 i 6 m, the model may befurther expressed as Yi DXi�CZi˛i C �i , i D 1; : : : ; m, where Yi D .Yis/16s65, Xi DZi D 15˝ I2(here,˝ denotes the Kronecker product (e.g., [13, Appendix B.1]), 15 D .1; 1; 1; 1; 1/0, and I2 is the 2�2identity matrix), and �i D .�is/16s65; or, in the standard matrix expression, Y D X�CZ˛C �, whereY D .Yi /16i6m, X D .Xi /16i6m, Z D diag.Zi ; 1 6 i 6m/, ˛ D .˛i /16i6m, and � D .�i /16i6m. Evenin this simple case, the variance–covariance structure is not simple with six unknown variance compo-nents involved (of which only �h is of interest). The implementation of the procedure in SAS PROCMIXED is not straightforward and requires understanding of the aforementioned linear mixed model,somehow in a more complicated way as described in [14]. Nevertheless, we are able to compare ourmethod with that of Hamlett et al. [14]. Two cases are considered. In the first case, the characteristics ofinterest and measurement errors are both normally distributed. This is the ideal situation for REML. Inthe second case, the characteristics of interest both have a normal mixture distribution that consists 90%of N.4; 0:1/ and 10% of N.�4; 0:1/. This is the same distribution considered by Claeskens and Hart[17], with mean 3:2 and variance 5:86. The distribution is then standardized to satisfy the requirementsset up at the beginning of the section. In addition, the measurement errors have a centralized exponentialdistribution, that is, the distribution ofX�1, whereX has the exponential distribution with a mean equal

Table II. Comparison with Pearson correlation: true �h D 0:6; E D Monte Carlo mean; s.e. D Monte Carlostandard error.

ı �0:4 �0:2 0 0.2 0.4 0.6� 0.1 0.2 0.3 0.4 0.5 0.6

E.rh/ 0.605 0.603 0.601 0.601 0.597 0.595s:e:.rh/ 0.003 0.003 0.003 0.003 0.003 0.002E.r/ 0.096 0.197 0.298 0.400 0.498 0.599s:e:.r/ 0.002 0.002 0.002 0.002 0.002 0.001

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3409

Page 8: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

to 1. Two different sample sizes are considered: m D 50 and m D 100. In Table III and IV, we reportresults of same-data comparison based on K D 100 simulations. The results are remarkably close. Infact, there is hardly any consistent pattern in favor of any method in terms of the estimation performance.On the other hand, there is a significant difference in terms of the computational efficiency. For example,for the normal case with mD 50; �h D 0:3, the SAS PROC MIXED procedure took roughly 35 times ofthe time it took for our procedure to run the same simulations when the jobs were performed on a DellOPTIPLEX 755 desktop computer (3.0 GHz Intel E8400 CPU; 3.25 Gb RAM).

Next, we present some simulation results regarding testing hypotheses for the hidden correlation. First,we consider H0 W �h D 0 versus Ha W �h > 0. We consider both a relatively large sample size, m D 100,and a relatively small one, m D 20. In Table V, we report the results based on K D 1000 simulationruns, where ˛ is the level of significance, the column �h D 0 corresponds to the size of the test, and thecolumns �h D 0:3 and �h D 0:6 corresponds to the powers at the given alternatives. The value of ı isequal to 0:1, as in Table I. It appears that the t -test performs quite well in terms of the size, even with therelatively small sample size (m D 20). On the other hand, the powers are much lower for m D 20 thanfor m D 100, as expected. Next, we consider testing H0 W �h D 0:5 versus Ha W �h > 0:5. We consider,in addition to m D 20 and m D 100, a larger sample size m D 200. In Table VI, we report the results

Table III. Comparison with restricted maximum likelihood (REML)—normal case: MC mean (s.e.), % bias(s.e.).

rh REML

m True �h Mean % bias s.e. % s.e. Mean % bias s.e. % s.e.

50 0.3 0.306 2.1 0.017 5.5 0.306 2.1 0.015 5.150 0.6 0.600 0.0 0.013 2.2 0.596 �0:6 0.013 2.250 0.9 0.900 0.0 0.007 0.7 0.896 �0:5 0.006 0.7100 0.3 0.311 3.6 0.009 3.2 0.311 3.7 0.009 3.1100 0.6 0.618 3.0 0.007 1.2 0.615 2.5 0.008 1.3100 0.9 0.898 �0:2 0.004 0.5 0.897 �0:3 0.004 0.5

Table IV. Comparison with restricted maximum likelihood (REML)—non-normal case: MC mean (s.e.), %bias (s.e.).

rh REML

m True �h Mean % bias s.e. % s.e. Mean % bias s.e. % s.e.

50 0.3 0.319 6.2 0.019 6.2 0.320 6.7 0.018 6.050 0.6 0.599 �0:2 0.017 2.8 0.591 �1:5 0.016 2.750 0.9 0.904 0.4 0.009 1.1 0.899 �0:1 0.009 1.0100 0.3 0.296 �1:4 0.011 3.7 0.297 �1:0 0.011 3.7100 0.6 0.622 3.6 0.009 1.5 0.621 3.5 0.009 1.5100 0.9 0.900 0.0 0.005 0.6 0.898 �0:2 0.005 0.6

Table V. Size and power of t-test (H0 W �h D 0).

mD 20 mD 100

˛ �h D 0 �h D 0:3 �h D 0:6 �h D 0 �h D 0:3 �h D 0:6

0.10 0.118 0.491 0.873 0.093 0.878 1.0000.05 0.057 0.321 0.761 0.050 0.799 1.000

Table VI. Size and power of t-test (H0 W �h D 0:5).

˛ D 0:10 ˛ D 0:05

m �h D 0:5 �h D 0:7 �h D 0:9 �h D 0:5 �h D 0:7 �h D 0:9

20 0.207 0.589 0.942 0.161 0.497 0.918100 0.144 0.898 1.000 0.097 0.838 1.000200 0.118 0.986 1.000 0.058 0.970 1.000

3410

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 9: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

based on K D 1000 simulation runs. It appears that, with the smaller sample size, the test somewhatover rejects under the null hypothesis, especially for m D 20, but the accuracy picks up quickly as thesample size increases. The power follows a reasonable trend as in the previous case.

Finally, we present some simulation results that compare our method with REML in terms of the test-ing performance. We consider testing the hypothesis H0 W �h D 0 versus Ha W �h ¤ 0. We compare ourtest (12) with the REML-based likelihood-ratio test (LRT). The LRT is well-known in the statistical liter-ature. For testing a simple null hypothesis versus a simple alternative, the fundamental Neyman–Pearsonlemma (e.g., [18]) states that the LRT is the most powerful among all tests of the same significancelevel (for a discussion of the LRT in linear mixed models, see [13, Section 2.1]). We consider the caseof normal data. The set-up is considered most favorable for the LRT. Three different sample sizes areconsidered: mD 20, 50, and 100, ranging from small to relatively large. We consider the same null andalternatives as in Table V. The numbers of simulation runs for m D 20; 50; 100 are K D 500; 200; 100,respectively (more simulation runs are needed for smaller sample size in order to obtain stable results).The results are reported in Table VII. It is seen that, overall, the LRT performs somewhat better but thedifference is diminishing as the sample size increases. On the other hand, our method is, again, muchmore favorable in terms of the computational efficiency.

5. Home care worker example revisited

Now let us return to the home care worker study discussed in Section 1. We apply our method to estimatethe hidden correlation between each pair of variables. In Table VIII, we report partial results of the esti-mated correlation coefficients, rh, with the corresponding p-values in parentheses. As noted, the Pearsoncorrelation coefficient is inappropriate for these hidden correlation problems. Nevertheless, it might beinterested in knowing what one is getting if the Pearson method is blindly used. The results show thatthe Pearson correlation coefficient, r , is, in general, lower in absolute value than our estimate, rh, whichis consistent with our expectation (see the end of the discussion following (4)). Table IX provides a sum-mary of these comparisons, where reported are the numbers of estimated correlation coefficients whoseabsolute values are greater than or equal to the cutoff, for both methods. It is clear that the rh estimates

Table VII. Comparison with (restricted maximum likelihood) likelihood-ratio test (LRT).

�h D 0 �h D 0:3 �h D 0:6

m ˛ rh LRT rh LRT rh LRT

20 0.10 0.142 0.106 0.334 0.314 0.748 0.79820 0.05 0.066 0.046 0.232 0.232 0.592 0.69250 0.10 0.105 0.070 0.555 0.585 0.960 0.97550 0.05 0.055 0.050 0.405 0.425 0.930 0.955100 0.10 0.110 0.140 0.830 0.840 1.000 1.000100 0.05 0.080 0.070 0.770 0.770 1.000 1.000

Table VIII. Estimated correlations and p-values: partial results of the estimated correlation coefficient, rh,with the corresponding p-value in parentheses.

Variable 3 4 5 6 7

3 1.00 (NA) 0.62 (0.11) 0.63 (0.05) 0.51 (0.01) 0.94 (0.25)4 1.00 (NA) 0.84 (0.00) 0.52 (0.01) 0.69 (0.02)5 1.00 (NA) 0.45 (0.00) 0.71 (0.01)6 1.00 (NA) 0.41 (0.00)7 1.00 (NA)

Table IX. Number of absolute values of estimates > cutoff.

Cutoff 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

rh 220 176 141 97 62 43 28 17 8r 169 95 63 37 16 7 4 1 0Significant rh 47 45 43 40 30 21 12 9 4

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3411

Page 10: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

are, in general, higher than the r estimates in absolute values. Recall that r may be seriously biased(Table II). As examples of specific comparisons, the r corresponding to the first row of Table VIII are1.00 (NA), 0.33 (0.00), 0.20 (0.00), 0.38 (0.00), and 0.47 (0.00), with the p-values in parentheses. Itshould be noted that the p-value for the Pearson correlation may be misleading in this case, because itis computed on the basis of the assumption that all the repeated measures are independent. However,because the observations are clustered, there may be within-cluster correlations. The implication is thatthe effective sample size (used to determine the standard errors) is not the same as the total sample sizebut is usually smaller. Therefore, in the bottom row of Table IX, we only report the numbers of significantrh results (at the 5% level) that are greater than or equal to the cutoff in absolute values.

Significant strongly correlated (rh > 0:8) variables include heavy object time and pain lag; heavyobject time and non-lagged pain; driving for client and house clean time; bathing/toileting time anddressing time; house clean time and task minute sum; heavy object time and task minute sum; dailyoccupational fatigue and daily negative affect; and so on. These can be easily interpreted and agree quitewell with the common sense.

6. Discussion

We have developed a simple, Pearson-type estimator of the correlation coefficient between unobservablecharacteristics of interest, the hidden correlation. Our method is developed without making stringentassumptions about the variance/covariance of the measurement errors, such that c�� D 0, or the normalityassumption about the data. This makes the method more applicable to practical situations. Furthermore,we have derived large sample standard errors associated with the hidden correlation estimator, whichcan be used for hypothesis testing and confidence intervals. Simulation results suggest that the proposedestimator works well in both estimation and testing performances, especially when the sample size is rel-atively large. Furthermore, we show by a simulation study that the proposed simple estimator performssimilarly to the estimator obtained via the more sophisticated mixed-effects model approach but that ourmethod is much more efficient in terms of computational speed. A potential limitation of our method isthat its testing procedure relies on large sample approximations, which we discuss in further details inthe succeeding text.

The p-value calculation (Section 3) is based on the asymptotic normal distribution. However, the nor-mal approximation may not be accurate when the sample size is small. For example, it is seen in Table VIthat the simulated size is much higher than the nominal level when the sample size is small (m D 20).In such case, permutation or bootstrap methods are often used in other contexts of studies. Permutationtests rely on exchangeability under the null hypothesis. In the current case, however, the null hypothesisdoes not seem to lead to useful exchangeability. For example, it is not clear at all if any exchangeabilitycan be implied under the null hypothesis H0 W �h D 0:3. In fact, even H0 W �h D 0 does not seem helpful,unless it is known in advance that c�� D 0 (see the succeeding text for details).

It also seems difficult to bootstrap the p-value in the current situation. To see this, first, note that thisis not a conventional bootstrap based on the observed data—the null hypothesis has to be involved inthe resampling. The simplest case would be to test H0 W �h D 0. If the xi ’s and yi ’s are directly observ-able, then a non-parametric bootstrap can be carried out by drawing samples x�1 ; : : : ; x

�m and y�1 ; : : : ; y

�m

independently and then pairing x�i with y�i ; i D 1; : : : ; m. Note that the null hypothesis does play arole in this resampling procedure (otherwise, it is not equivalent to resampling .x�i ; y

�i /; i D 1; : : : ; m

from f.xi ; yi /; i D 1; : : : ; mg). However, this strategy does not work if the xi ’s and yi ’s are not directlyobservable. Furthermore, a similar approach does not work with the xi ; yi replaced by xis; yis . This isbecause, even if xi ; yi are independent (under the null hypothesis), there is no independence betweenxis and yis; in fact, cov.xis; yis/D c�� under the null hypothesis, and the latter may not be zero.

Another difficulty of the non-parametric bootstrap is because the ni ’s are not necessarily equal. Notethat if the ni ’s are the same, say, ni D k; 1 6 i 6 m, then the vectors .xis; yis; s D 1; : : : ; k/; i D1; : : : ; m are i.i.d. However, assuming that the ni ’s are equal may not be practical. Even in the specialcase with all the ni ’s equal, it is still unclear how the null hypothesis can come into play in a practicalway.

On the other hand, it is often easier to incorporate the null hypothesis with the model-based or para-metric bootstrap. Once again, consider the simple case of testing H0 W �h D 0. If normality is assumed,one could estimate all the unknown parameters except �h, which is set to 0 under H0, and then use theestimated parameters and �h D 0 (as the true parameters) to generate the data. However, this approachis problematic in our case. First, because a parametric model is not assumed, the parametric-bootstrap

3412

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 11: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

approach does not work. Second, even if one is willing to make stronger assumptions, such as normal-ity, the accuracy of the parameter estimation is, again, dependent on the sample size. Recall that thebootstrap idea is brought about due to the concern of small sample size that results in inaccurate normalapproximation. The very same concern arises with the parameter estimation, which results in inaccu-rate bootstrapping. As written by Efron [19], bootstrap provides ‘approximate frequency statements, notapproximate likelihood statements. Fundamental inference problems remain, no matter how well thebootstrap works.’

In a way, the nature of difficulty in bootstrapping encountered here is similar to that under a mixed-effects model. Note that the latter also involves unobservable random quantities (the random effects) thatcannot be directly bootstrapped (see, e.g., [20, 21] for some recent developments). The strategies in thelatter studies will be explored in our future study. It would also be interesting to explore some other moreadvanced bootstrap ideas, such as the block bootstrap [22].

APPENDIX A.

A.1 Complete analytic expressions of Ocxy ; Ocxx; Ocyy

Ocxy D1

m

mXiD1

ni

ni � 1Nxi � Nyi � �

1

m

mXiD1

Nxi �

! 1

m

mXiD1

Nyi �

!

1

m

mXiD1

1

ni � 1

! 1

m

mXiD1

1

ni

niXsD1

xisyis

!I

Ocxx D1

m

mXiD1

ni

ni � 1. Nxi �/

2 �

1

m

mXiD1

Nxi �

!2

1

m

mXiD1

1

ni � 1

! 1

m

mXiD1

1

ni

niXsD1

x2is

!I

Ocyy D1

m

mXiD1

ni

ni � 1. Nyi �/

2 �

1

m

mXiD1

Nyi �

!2

1

m

mXiD1

1

ni � 1

! 1

m

mXiD1

1

ni

niXsD1

y2is

!:

A.2 Derivation of OV

Define

c2 D1

m

mXiD1

�ni

ni � 1

�2;

Ov11 D

mXiD1

Nx2i � �m � u21;

Ov22 D

mXiD1

Ny2i � �m � u22

[ Nx2i � D . Nxi �/2 and so on.]. Now, let

c2 D1

m

mXiD1

�ni

ni � 1

�2; u5 D

1

m

mXiD1

Nxi � Nyi �:

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3413

Page 12: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

We continue with the definitions of Ov’s:

Ov33 D

mXiD1

�ni

ni � 1

�2Nx2i � Ny

2i � �m � c2 � u

25;

Ov44 D

mXiD1

1

ni

niXsD1

xisyis

!2�m � u24I

Ov12 D

mXiD1

Nxi � Nyi � �m � u1 � u2;

Ov13 D

mXiD1

ni

ni � 1Nx2i � Nyi � �m.c1C 1/ � u1 � u5:

Next, define ´is D xisyis , N i � D n�1iPnisD1 ´is , and u6 Dm�1

PmiD1 N i �. We define

Ov14 D

mXiD1

Nxi � N i � �m � u1 � u6;

Ov23 D

mXiD1

ni

ni � 1Nxi � Ny

2i � �m.c1C 1/ � u2 � u5;

Ov24 D

mXiD1

Nyi � N i � �m � u2 � u6;

Ov34 D

mXiD1

ni

ni � 1Nxi � Nyi � N i � �m.c1C 1/ � u5 � u6:

Extend the definition of Ov according to symmetry; that is, Ovkl D Ovlk; k ¤ l . Then, Var.u/ in (13) isestimated by m�2 OV Dm�2. Ovkl/16k;l64.

A.3 Derivation of s:e:.rh/

The derivation here is ‘semirigorous’. For example, we shall assume that EfoP.m�1/g D o.m�1/,

although some regularity conditions are needed for this to hold. We refer the details to a technical report[23]. First, it is easy to show, by the Taylor expansion and by the fact that Ocxy � cxy D OP.m

�1=2/ andso on, we have

rh � �h DOcxy � cxypcxxcyy

�cxy

2pcxxcyy

�Ocxx � cxx

cxxCOcyy � cyy

cyy

�C oP.m

�1=2/:

Furthermore, it can be shown that Ocxy � cxy D m�1PmiD1 �i;xy C oP.m

�1=2/, Ocxx � cxx Dm�1

PmiD1 �i;xx C oP.m

�1=2/, and Ocyy � cyy Dm�1PmiD1 �i;yy C oP.m

�1=2/, where

�i;xy Dni

ni � 1Nxi � Nyi � � E.y11/ Nxi � � E.x11/ Nyi � � E.x11/E.y11/�

c1

ni

niXsD1

xisyis

� cxy �

�1

ni � 1� c1

�E.x11y11/;

�i;xx Dni

ni � 1Nx2i � � 2E.x11/ Nxi � � fE.x11/g

2 �c1

ni

niXsD1

x2is

� cxx �

�1

ni � 1� c1

�E.x211/;

3414

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

Page 13: Simple estimation of hidden correlation in repeated measures

T. NGUYEN AND J. JIANG

and �i;yy is �i;xx with x replaced by y. Note that all the �’s have mean zero. Thus, combining the results,we have rh � �h Dm

�1PmiD1 �i C oP.m

�1=2/, where

�i D1

pcxxcyy

��i;xy �

cxy

2

��i;xx

cxxC�i;yy

cyy

��:

Note that �i ; i D 1; : : : ; m are independent, with mean zero. Thus, we have var.rh/Dm�2E.

PmiD1 �

2i /C

o.m�1/. A first-order approximation to var.rh/, in the sense that the difference is o.m�1/ (note thatvar.rh/DO.m

�1/) (e.g., [24, Section 4.8]), is given by Ov2 Dm�2PmiD1O�2i , where O�i is �i with

cxy , cxx , cyy , E.x11/, E.y11/, E.x11y11/, E.x211/, E.y211/ replaced by Ocxy , Ocxx , Ocyy , m�1PmiD1 Nxi �,

m�1PmiD1 Nyi �, m�1

PmiD1 n

�1i

PnisD1 xisyis , m�1

PmiD1 n

�1i

PnisD1 x

2is , m�1

PmiD1 n

�1i

PnisD1 y

2is ,

respectively. Therefore, we have the expression s:e:.rh/D Ov Dm�1

qPmiD1O�2i .

Acknowledgements

The authors’ research is partially supported by the NIH grant R01-GM085205A1. In addition, Jiming Jiang’sresearch is partially supported by the NSF grant DMS-0809127. The authors are grateful to Drs. Ryan Olson andBrad Wipfli of the Center for Research on Occupational and Environmental Toxicology at the Oregon Healthand Science University for providing the data from their research as well as information and helpful discussions.The authors are grateful to an associate editor and to two referees for their thoughtful comments that led to theimprovement of the manuscript.

References1. Rodgers JL, Nicewander WA. Thirteen ways to look at the correlation coefficient. The American Statistician 1988;

42:59–66.2. Bland JM, Altman DG. Statistics notes: measurement error and correlation coefficient. British Medical Journal 1996;

313:41.3. Fisher RA. Statistical Methods for Research Workers, 12th. Oliver & Boyd: Edinburgh, 1954.4. Koch GG. Intraclass correlation coefficient. In Encyclopedia of Statistical Sciences, Vol. 4, Kotz S, Johnson NL (eds).

Wiley: New York, 1982; 213–217.5. Graybill FA. An Introduction to Linear Statistical Models, Vol I. McGraw-Hill: New York, 1961.6. Horn M, Guiard V. Correlation and linear regression if the random variables are subject to errors or fluctuations. Biomed

Journal 1986; 28:683–696.7. Rosner B, Willett WC. Interval estimates for correlation coefficients corrected for within-person variation: implications

for study design and hypothesis testing. American Journal of Epidemiology 1988; 127:377–386.8. Rifkin RD. Effects of correlated and uncorrelated measurement error on linear regression and correlation in medical

method comparison studies. Medical Statistics 1995; 14:789–798.9. Archer KJ, Dumur CI, Taylor GS, Chaplin MD, Guiseppi-Elie A, Buck GA, Grant G, Ferreira-Gonzalez A, Gurrett

CT. A disattenuated correlation estimate when variables are measured with error: illustration estimating cross-platformcorrelations. Medical Statistics 2008; 27:1026–1039.

10. Schaalje GB, Butts RA. Some effects of ignoring correlated measurement errors in straight line regression and prediction.Biometrics 1993; 49:1262–1267.

11. Thoresen M, Laake P. On the simple linear regression model with correlated measurement errors. Journal of StatisticalPlanning and Inference 2007; 137:68–78.

12. Anuradha R. Estimating correlation coefficient between two variables with repeated observations using mixed effectsmodel. Biomed Journal 2006; 48:286–301.

13. Jiang J. Linear and Generalized Linear Mixed Models and Their Applications. Springer: New York, 2007.14. Hamlett A, Ryan L, Wolfinger R. SUGI Proceedings: statistics, data analysis and data mining 2003; 198–29:1–7.15. Rahman NA. A Course in Theoretical Statistics. Charles Griffin & Company Ltd.: London, 1968.16. Lehmann EL. Elements of Large-sample Theory. Springer: New York, 1999.17. Claeskens G, Hart J. Goodness-of-fit tests in mixed models (with discussion). TEST 2009; 18:213–270.18. Lehmann EL. Testing Statistical Hypothesis, 2nd edn. Springer: New York, 1986.19. Efron B. Bootstrap method: another look at the jackknife. Annals of Statistics 1979; 7:1–26.20. Hall P, Maiti T. Nonparametric estimation of mean-squared prediction error in nested-error regression models. Annals of

Statistics 2006; 34:1733–1750.21. Chatterjee S, Lahiri P, Li H. Parametric bootstrap approximation to the distribution of EBLUP, and related prediction

intervals in linear mixed models. Annals of Statistics 2008; 36:1221–1245.22. Künsch HR. The jackknife and the bootstrap for general stationary observations. Annals of Statistics 1989; 17:1217–1241.23. Nguyen T, Jiang J. First-order approximation to the variance of the hidden correlation estimator. Tech. Report, Dept.

Statist., Univ. Calif., Davis, CA, 2011.24. Jiang J. Large Sample Techniques for Statistics. Springer: New York, 2010.

Copyright © 2011 John Wiley & Sons, Ltd. Statist. Med. 2011, 30 3403–3415

3415