10
The Need for Research That Is Designed to Support Decisions in the Delivery of Mental Health Services Frederick L Newman Manuel J. Tejeda Florida International University and University of Miami University of Miami M. E. P. Seligman (1995) argued that traditional ap- proaches to mental health services research fail to provide useful information to consumers and practitioners, par- ticularly in an environment increasingly dominated by managed care. The authors recommend 4 guidelines for designing a research program so that the results can sup- port the decisions of the major stakeholders (clients-fam- ilies, practitioners, service managers, and policymakers): (a) Research must be targeted and programmatic and en- compass a strategy of complementary efficacy, effective- ness, and cost-effectiveness studies; (b) study design and measure selection must be sensitive to describe who, and in what context, is best served by which intervention; (c) the design must inform stakeholders as to the type and amount of effort that is required to achieve a behavioral criteria," and (d) the strategy should inform researchers how information should be formatted to best support the decisions of the key stakeholders. S eligman (1995) offered a provocative review of the limitations of our research methods in supplying stakeholders in the delivery of mental health ser- vices with information to support their decisions. Al- though many of the issues raised by Seligman are prob- lematic, central in his argument are the burden and failure that the scientific community bears in supplying fruitful information to the consumers of our research, primarily the stakeholders of the mental health delivery system. We present and discuss four guidelines that should be con- sidered in designing a program of research that generates information to support the decisions of the key stakeholders. Stakeholders Before presenting the four guidelines, we need to identify the key stakeholders and the types of decisions that a program of research could support. To do so, we will use the context of managed care, the most prevalent mech- anism of service delivery of this decade• For the purposes of this discussion we consider four classes of stakeholders: (a) the consumer (including the client, the client's family, and the party paying for the service), (b) the practitioner, (c) the supervisors and service manager(s), and (d) the policymaker who set the standards for care and reim- bursement. The primary set of stakeholders are direct consumers of the service (the client or patient) and the consumer's family or employer (who are directly or in- directly paying for the costs of the services). Consumers ask themselves, "Will I be able to function adequately. . . What is required of me and can I afford the services?" The practitioner(s) providing the services, the clinical su- pervisors, and managers of the service want to understand, "What interventions will have optimal impact on con- sumers of a specific psychosocial profile in the bounds of the resources that have been made available to us?" Those persons who set policies for quality and reimbursement standards want to understand, "What outcomes, pro- cesses, and cost-constraint standards are feasible for each class of consumers?" An additional stakeholder group that would influence publicly funded services are those agen- cies and agents across the community who are indirectly impacted by delivery of such services (e.g., health, legal, or social service providers). They want to understand, "What is our role or responsibility to or for persons with a mental health problem?" Currently, most insurance carriers and managed care organizations are following two lines of evidence to direct their policies: cost containment and consumer satisfac- tion. Few organizations have used behavioral criteria be- yond gross diagnostic categories to determine service eli- gibility or the level of service that is reimbursed. The amount and type of service are often based on a rule of dosage (e.g., 12, 18, or 20 sessions of psychotherapy; 10, 21, or 30 days of inpatient service; or both). Consider applying the logic of these cost containment rules in on- cological medicine: "Mr. Smith, 90% of the tumor is now benign. Only 10% of the tumor remains malignant. Unfortunately you have used up the 20 sessions of radiation treatment allowed under your managed care plan for this year. Please come back next year when your insurance eligibility has been renewed." (Newman & Carpenter, in press) Frederick L Newman, Health Services Administration, Florida Inter- national University and Center for Family Studies, Psychiatry and Be- havioral Sciences Department, University of Miami; Manuel J. Tejeda, Center for Family Studies, Department of Psychiatry and Behavioral Sciences, University of Miami. Five colleagues contributed greatly to this article: Jeff Berman, Gary Bond, Will Shadish, Gregory Teague, and Brian Yates. Correspondence concerning this article should be addressed to Frederick L Newman, HSA-SPM, Florida International University, North, 3000 NE 145th Street, North Miami, FL 33181. Electronic mail may be sent via Internet to [email protected]. 1040 October 1996 American Psychologist Copyright 1996by the American Psychological Association, Inc. 00034)66X/96/$2.00 Vol. 51. No. 10. 1040-1049

The Need for Research That Is Designed to Support

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

The Need for Research That Is Designed to Support Decisions in the Delivery of Mental Health Services

Frederick L Newman Manuel J. Tejeda

Florida International University and University of Miami University of Miami

M. E. P. Seligman (1995) argued that traditional ap- proaches to mental health services research fail to provide useful information to consumers and practitioners, par- ticularly in an environment increasingly dominated by managed care. The authors recommend 4 guidelines for designing a research program so that the results can sup- port the decisions of the major stakeholders (clients-fam- ilies, practitioners, service managers, and policymakers): (a) Research must be targeted and programmatic and en- compass a strategy of complementary efficacy, effective- ness, and cost-effectiveness studies; (b) study design and measure selection must be sensitive to describe who, and in what context, is best served by which intervention; (c) the design must inform stakeholders as to the type and amount of effort that is required to achieve a behavioral criteria," and (d) the strategy should inform researchers how information should be formatted to best support the decisions of the key stakeholders.

S eligman (1995) offered a provocative review of the limitations of our research methods in supplying stakeholders in the delivery of mental health ser-

vices with information to support their decisions. Al- though many of the issues raised by Seligman are prob- lematic, central in his argument are the burden and failure that the scientific community bears in supplying fruitful information to the consumers of our research, primarily the stakeholders of the mental health delivery system. We present and discuss four guidelines that should be con- sidered in designing a program of research that generates information to s u p p o r t the decisions of the key stakeholders.

Stakeholders Before presenting the four guidelines, we need to identify the key stakeholders and the types of decisions that a program of research could support. To do so, we will use the context of managed care, the most prevalent mech- anism of service delivery of this decade• For the purposes of this discussion we consider four classes of stakeholders: (a) the consumer (including the client, the client's family, and the party paying for the service), (b) the practitioner, (c) the supervisors and service manager(s), and (d) the policymaker who set the standards for care and reim- bursement. The primary set of stakeholders are direct consumers of the service (the client or patient) and the

consumer's family or employer (who are directly or in- directly paying for the costs of the services). Consumers ask themselves, "Will I be able to function adequately. • . . What is required of me and can I afford the services?" The practitioner(s) providing the services, the clinical su- pervisors, and managers of the service want to understand, "What interventions will have optimal impact on con- sumers of a specific psychosocial profile in the bounds of the resources that have been made available to us?" Those persons who set policies for quality and reimbursement standards want to understand, "What outcomes, pro- cesses, and cost-constraint standards are feasible for each class of consumers?" An additional stakeholder group that would influence publicly funded services are those agen- cies and agents across the community who are indirectly impacted by delivery of such services (e.g., health, legal, or social service providers). They want to understand, "What is our role or responsibility to or for persons with a mental health problem?"

Currently, most insurance carriers and managed care organizations are following two lines of evidence to direct their policies: cost containment and consumer satisfac- tion. Few organizations have used behavioral criteria be- yond gross diagnostic categories to determine service eli- gibility or the level of service that is reimbursed. The amount and type of service are often based on a rule of dosage (e.g., 12, 18, or 20 sessions of psychotherapy; 10, 21, or 30 days of inpatient service; or both). Consider applying the logic of these cost containment rules in on- cological medicine:

"Mr. Smith, 90% of the tumor is now benign. Only 10% of the tumor remains malignant. Unfortunately you have used up the 20 sessions of radiation treatment allowed under your managed care plan for this year. Please come back next year when your insurance eligibility has been renewed." (Newman & Carpenter, in press)

Frederick L Newman, Health Services Administration, Florida Inter- national University and Center for Family Studies, Psychiatry and Be- havioral Sciences Department, University of Miami; Manuel J. Tejeda, Center for Family Studies, Department of Psychiatry and Behavioral Sciences, University of Miami.

Five colleagues contributed greatly to this article: Jeff Berman, Gary Bond, Will Shadish, Gregory Teague, and Brian Yates.

Correspondence concerning this article should be addressed to Frederick L Newman, HSA-SPM, Florida International University, North, 3000 NE 145th Street, North Miami, FL 33181. Electronic mail may be sent via Internet to [email protected].

1040 October 1 9 9 6 • American Psychologist Copyright 1996 by the American Psychological Association, Inc. 00034)66X/96/$2.00

Vol. 51. No. 10. 1040-1049

The low correlations between indexes of consumer sat- isfaction and actual impact on functioning or quality of life clearly indicate that this approach, albeit currently popular, is limited in its usefulness for guiding consumers in the selection of a service provider.

Whether they are privately or publicly operated, managed care organizations are seen as holding only profit or cost constraints as their guiding principle. However, there is strong evidence that these organizations cannot survive unless they consider both the effectiveness (out- comes) and the costs of providing the therapeutic inter- ventions available under their managed care plan. For long-term survival in a complex industry such as health care, managed care plans must consider the outcomes and the costs to three groups: (a) the payer (including the employer, or the government, and the copaying consumer or family), (b) the service provider, and (c) those in the communi ty affected by the client-patient 's behavior,

Our research strategies are often delinquent in ad- dressing these issues. The consumers of our findings in- form us of this failure. Thus, we argue that our research designs must begin to incorporate the needs of our con- sumers, specifically to support the decision making of all parties involved in the delivery of mental health care.

An additional group of stakeholders is the scientific communi ty itself, who advance and test basic theories about behavior and the psyche. This group cannot be neglected here- -we are that group! Our colleagues are our consumers, and we must urge ourselves to operate under the same ideological umbrella voiced by Bronowski (1959), who recommended that science must be guided by the simultaneous exploration and examination of hu- man values. An underlying value is that our science should provide information useful to each of the four sets of decision makers. In this context, we offer the following set of guidelines for designing a program of research that is intended to support the decisions of these stakeholders.

Efficacy, Effectiveness, and Cost-Effectiveness Research must be targeted, programmatic, and encom- pass a strategy of complementary efficacy, effectiveness, and cost-effectiveness studies. This first guideline rec- ommends that a broad but targeted program of research be envisioned to develop and test the efficacy and effec- tiveness of a therapeutic intervention. Randomized con- trolled trials are typically needed in the early stages of development, not only to refine a technique that may produce stronger effects but also to permit testing for the intervention's safety under controlled conditions.

It is interesting that one requirement of a well-de- signed efficacy study is to institute monitoring methods to assure treatment fideli ty--who does what with which, when, where, and to whom are obsessively documented. These data are precisely the basic ingredients required for many forms of cost-effectiveness research (see Knapp, 1995; Yates, 1996). Once a reasonably strong and safe intervention has been demonstrated by the efficacy stud- ies, a program of effectiveness and cost-effectiveness stud-

ies can help us understand who is best served by the ther- apeutic intervention and in what environmental context. To be sure, both efficacy and effectiveness studies can serve the same master provided there is a clear under- standing of the overall mission of a research program and how complementary strategies can benefit that mission.

Although the meta-analysis of meta-analyses by Lipsey and Wilson (1993) discovered that randomized control trials and quasi-experimental (effectiveness) stud- ies yield similar effect sizes, Shadish and his colleagues (Heinsman & Shadish, 1996; Shadish & Ragsdale, in press) found that randomized trial studies yielded stronger effect sizes. These differences were explained by the fact that effectiveness studies typically have less control over facets that increase error variance (the denominator of an effect-size estimator).

Seligman (1995), among others, correctly noted that randomized control trials and well-controlled effectiveness studies have advantages that the practicing clinician does not. Specifically, the practicing clinician does not have full control over who walks in the door and has less control over spurious events between sessions. Nevertheless, im- portant issues at both ends of the control spectrum can be addressed from the cont inuum of research designs-- from efficacy studies, through the nine-plus quasi-exper- imental designs, to the retrospective surveys represented by the Consumer Reports (CR," 1994) survey. Let us com- pare the differences in the interpretations that come from the CR survey with that of the Shadish and Ragsdale (in press) meta-analysis as a case study that argues for the need for complementary efficacy and effectiveness re- search. In the CR survey, marital and family counselors were not found to be as effective as psychiatrists, psy- chologists, or social workers. Yet the Shadish and Ragsdale (in press) meta-analysis on 64 efficacy studies (with ran- dom assignment) and 36 effectiveness studies (using non- equivalent group designs) found what are often charac- terized as large effect sizes for family psychotherapy. Note that marital therapy was not included in their meta-anal- ysis. They also found that the effect sizes were significantly larger for the efficacy studies (with random assignment) than for the quasi-experimental (nonequivalent group) studies.

More research is needed to partition the effects of the type of therapy from the therapist's discipline to better understand the discrepancy between the clear success of the 100 marital and family therapy studies in the Shadish and Ragsdale (in press) meta-analysis and the findings in the CR (1994) survey. The CR survey results appear to provide the first major evidence that training and licensure requirements do make a difference, whereas other single studies or meta-analyses have not shown the same results (Berman & Norton, 1985). Moreover, Berman (personal communication, February 16, 1996) recommended that "any observed differences between disciplines in the CR survey are likely to be differences in the type, problem severity, and prognosis of clients who choose different types of helpers."

October 1996 • American Psychologist 1041

Should we interpret these results to mean that when marital or family therapy is used in the real world, it is not effective? Or are there alternative explanations, as suggested by Berman (personal communication, February 16, 1996). Shadish and Ragsdale's (in press) results suggest that preselection and facets influencing both internal and external validity strongly affect the predicted outcome. Furthermore, it should be obvious that marital and family therapy is one area that calls for a series of effectiveness studies to help us understand what real-life circumstances could negatively influence the potential positive effects of a therapy with demonstrated efficacy.

Although the CR (1995) article raised several issues that will guide follow-up research, its bold conclusions about family and marital therapy appear to be too broad, and the reporting of the data could have cautioned the reader as to the potential for this misinterpretation. There are notable examples in the scientific literature in which a series of well-designed randomized controlled trials questioned basic assumptions and were key to identifying more appropriate intervention strategies. If popular beliefs were allowed to prevail, these bolder intervention strat- egies might not have had a chance to show their potential. To provide a good example of this, we asked Bond (per- sonal communication, February 19, 1996) to identify random assignment studies that were, and still are, sem- inal to understanding how to effectively deliver mental health services to persons with a severe and persistent mental illness in the community. His recommendations provided a strong argument for randomized control trials in an area in which there has been little belief that com- munity treatment is either efficacious or cost-effective. We quote directly from Bond's electronic mail to us:

May, Tuma, and Dixon (1981), examining inpatients with schizophrenia, compared five treatment conditions: psycho- therapy only, psychotherapy and phenothiazine, electroconvul- sive shock therapy, milieu, and psychotherapy plus medication. The psychotherapy-only condition had the worst outcomes overall, similar to outcomes for those receiving milieu treatment only. This study pretty much dispelled the idea that psycho- therapy as practiced was the best treatment for schizophrenia. Karon and VandenBos (1981) questioned whether the "right" form of psychotherapy was provided and did find that psycho- therapy tailored to the population was shown as cost-effective.

Hogarty, Goldberg, Schooler, and Ulrich (1974) shifted the focus of research to that of combined medication and an array of psychosocial interventions rather than just psychotherapy. Hogarty et al. performed a randomized controlled trial of psy- chotropic medication and major role therapy for people with a diagnosis of schizophrenia. They found that psychotropic med- ication had a major effect in reducing relapse, but that psycho- social interventions also had an incremental effect above and beyond what the medication contributed. From the late 1970s through the early 1990s, a series of randomized clinical trials employing assertive community treatment which integrated medication management with different forms of psychosocial interventions were performed by Stein and Test (1980) and by Bond and his colleagues (see reviews in Bond, 1991, 1992; Bond, McGrew, & Fekete, 1995). These teams of researchers repeatedly demonstrated the efficacy of these models to increase community tenure and decrease number of days of hospitalization.

Until Weisbrod's (1983) economic analysis of the Stein and Test (1980) study, most people believed that the model of assertive community treatment programs had hidden costs that actually made them more expensive than episodic treatment approaches. Weisbrod had a fairly extensive accounting frame- work applied to a randomized controlled trial. He showed that a model program had about the same cost as controls, and with better client outcomes. A clinical trial by Drake, McHugo, Becker, Anthony, and Clark (1996) dispelled the myth that peo- ple with severe mental disorder need a period of career explo- ration before they are ready to start employment. (Bond, per- sonal communication, February 19, 1996)

Taken together, Bond (personal communication, February 19, 1996) argued that a series of randomized control trials over the last three decades has focused our attention on the possibility of cost-effectively serving our most persistently impaired citizens in the community with a possibility of community employment. But one may still ask whether this is true of all subsets of people with a history of persistent and severe mental illness. Are there permutations on the basic model that might serve some subgroups better? Here the issue concerns identifying necessary ingredients in assertive community treatment that may have their most potent impact on different types of consumers. What ingredients are necessary for all groups?

McGrew, Bond, Dietzen, and Salyers (1994) per- formed a meta-analysis across 18 sites on assertive com- munity treatment and identified those ingredients that were associated with the greatest number of days reduc- tion in hospitalizations. Does that retrospective analysis set the stage for a new set of prospective studies in which the ingredients are systematically varied, or should we simply use the results of the retrospective analysis to identify the most effect ingredients? Would those sup- porting the view of outcome represented in the CR (1995) article recommend that we simply accept the retrospective analysis performed by McGrew et al.? Probably not, once someone points out that some of the ingredients (e.g., a 1:10 staff-to-consumer ratio and that both a medical doc- tor and a nurse should be on the team) are expensive.

Finally, treatment-adherence data collected in most efficacy studies contain the basis for cost-effectiveness analysis. In its most basic form, treatment adherence is nothing more than the managerial accounting of the be- haviors of individuals involved in the delivery of service. Moreover, accounting for the behaviors of individuals is not only sound science but effective management practice (Mintzberg, 1973). It requires little more effort in the management of a scientific endeavor to account for the activities of clinical staff on a scientific project. Assessing staff time and whereabouts provides the fundamental pieces for a cost-effectiveness analysis. As argued over a decade ago (Newman & Howard, 1986), our traditional procedures of collecting data on assuring treatment in- tegrity and assessing therapeutic effort in our research studies are sufficient to perform some of the most basic cost-effectiveness studies. Moreover, the procedures de- scribed by Knapp (1995) and his colleagues by either a

1042 October 1996 • American Psychologist

structured interview or a self-report are sufficient to col- lect the information needed for more sophisticated cost- effectiveness studies. Stated quite simply, the technology exists for us to collect the data to conduct cost-effective- ness analysis even with our most basic efficacy studies and certainly with our broader naturalistic studies.

The story here is one in which a combination of randomized control trials, effectiveness, and retrospective studies have come together under a program of research. Given the evolution of this research, it is probably time to perform a CR (1994)-type survey with the consumers of assertive communi ty treatment services. As described later, several such efforts are now in progress. But there is an intermediate issue to be discussed first.

W h o Is Best Served and in W h a t Context? The study design and measures selected must be sensitive to produce results that inform us about who, and in what context, is best served by which intervention. This second guideline recommends that this can be done in a manner that is ultimately useful to the four classes of stakeholders only if the program of research systematically considers what attribute and contextual moderator variables will influence the effectiveness of the intervention. There are a number of moderator variables that should be helpful to both the consuming-paying public and the professional communities in the choice of therapeutic mode or cli- nician. At a min imum, the study design should include the analysis and report of the interaction (covariation) of the initial level of symptom distress, functional status, or self-management ability with treatment effects.

Although the CR (1995) article did at tempt to de- scribe some of the interactions, the questionnaire (CR, 1994) design and the analyses used did not permit a thorough exposition of who is best served by which in- tervention to be addressed. Those involved in the design and analysis of the survey are not to be faulted. Creat- ing a study design and selecting measures that are suffi- ciently reliable and sensitive to detect such interactions are profoundly difficult; with a great deal of debate in the literature as to what can and should be done. What follows is an analysis of the problems with some recommendations.

An important issue embedded here is whether one should separate out those variables that would be im- portant to consider when designing how a therapeutic intervention is delivered so that it has its greatest impact. An approach often used by some researchers is to statis- tically adjust for differences observed among participants after treatment for those differences that were found prior to administering the treatment. Covariance adjustments or residual gain scores are two of these approaches (Cron- bach, 1992; Cronbach & Furby, 1970). The intent of this adjustment is to estimate treatment effects independent of differences in initial status. A variation on this argu- ment is that one can more cleanly test a theory that pre- dicts between group differences if such pesky initial vari- ations are taken out of the equation (the error term). Some

of the analyses in the CR (1995) article used such ad- justments, and some methodologists would question this approach.

An argument against using adjustments for initial differences in a test of between-group differences is one of basic relevance (Newman, 1994; Willett & Sayer, 1994). I f there exists a common source ofcovariation in the day- to-day delivery of services, then that source of the co- variance should be studied rather than subtracted out of the analysis. Specifically, a program of research on the effectiveness of a particular intervention should include the major sources of variation that one might reasonably expect in the delivery of the services in the field. If be- tween-group differences are found to be significantly above an acceptable noise level represented in current practice, then these differences are very much worth knowing.

There are several problems with this view when it is taken to the extreme. Often, an innovative approach requires careful development in its early stages. Early de- velopment and test work probably should be discovery based (Greenberg & Newman, 1996), and early discovery- oriented exploration should be followed by research de- signs that are most sensitive to the intervention's potential benefits. Any known technique that can control error variance should be part of the designs at these early de- velopmental stages. However, it is critical that the program of research lead to testing the robustness of the interven- tion under those conditions that are common in the field of application.

The specific recommendation is that once the re- search program achieves a level in which effectiveness studies are indicated, then the sources of covariation that are due to key moderator variables, including that of the initial status, must be studied and not statistically con- trolled (removed) in the analysis and reports. Including these interactions in the study design will move us toward the goal of identifying for whom a therapeutic intervention is best suited and away from the do-do bird interpretation that all the therapies are winners.

There are two traditional problems that arise when studying interactions between treatment and moderator variables. One is inherent in our methods of measurement (i.e., reliability), and the other is whether our methods of analysis have sufficient power to detect effects. Shoham- Salomon (1991; Shoham-Salomon & Hannah, 1991) and Beutler (I 991) have documented the poor record of iso- lating such interactions in psychotherapy research. In those instances in which interactions have been found, the researchers had a strong theoretical reason for selecting the moderator variable as well as the dependent measures. In those cases in which generic measures of outcome were used, significant results typically were not obtained. One inference from their reviews is that the outcome measures need to be more carefully selected when studying an in- teraction effect. A more important inference is that theory should not be underestimated.

Simulation studies by McClelland and Judd (1993) and by Jaccard and Wan (1995) have provided ample

October 1996 • American Psychologist 1043

demonstration that detecting interactions and moderator effects typically have low power even when difficulties that are due to measurement reliability are minimal. When measurement reliability is a problem, the issue of power worsens. Aiken and West (1993) provided an excellent pr imer on how to deal with testing and interpreting in- teractions through regression techniques.

Recent findings have been more encouraging. Jac- card and Wan (1995) found that structural equation ap- proaches are more powerful in detecting such interactions. Even more exciting are the findings by Willett and Sayer (1994) that structural equation modeling can also be used to test differences in patterns of change among different subgroups within and between treatment groups. Willett and Sayer ended their article with the strong recommen- dation that "we can measure ' change ' - - and we should" (p. 379).

Several authors (Bryk & Raudenbush, 1988; Lyons & Howard, 199 I; Willett & Sayer, 1994) described pro- cedures that can identify when a source of error variance may be masking the interaction of a moderator and a treatment variable that should be investigated in follow- up research. However, as indicated earlier, one alternative explanation for large error variance remains measurement reliability-unreliability (Lyons & Howard, 1991). An- other is that neither the main effect of t reatment nor the interaction effects exist or are sufficiently large to matter.

Another problem in the description of key interac- tion effects is our field's questionable allegiance to statis- tical significance (Cohen, 1994; Schmidt, 1996b). The politics o f p values are ubiquitous in the industry of sci- ence. Our typically underpowered studies are likely to not be published in refereed journals or any journal, for that matter. Although this issue is recognized by editors, we have not found a satisfactory substitute to guide the accepting or rejecting of research reports in our scientific archival journals.

One positive sign is that Division 5 (Evaluation, Measurement, and Statistics) of the American Psycho- logical Association has formed a work group to study the issue with other scientific associations, with an objective to formulate guidelines for editorial review that are less reliant on statistical significance (Schmidt, 1996a). An- other solution may lie in using effect sizes. Although effect sizes, power, and p values are interlinked, small samples can produce low power and nonsignificance while yielding a modest effect size. Overdependence on statistical sig- nificance in small samples while ignoring effect sizes may lead us to ignore relationships that may be i m p o r t a n t - - the Type II error. Science has yet to deal effectively with this issue: Unlike Type I errors in which the probability can be computed exactly, the precise probability of Type II errors is usually, and uncomfortably, unknown.

The bot tom line here is that a program of research should plan to describe the interaction effects of a treat- ment with the principal moderator variables that clini- cians face in their daily practice. Although the literature has noted that there are measurement, design, and anal- ysis techniques that have the potential to do this, an unmet

challenge is whether retrospective surveys, such as the CR (1994) survey, can be designed to be sensitive to these interaction effects. We need to improve on our ability to design prospective studies that are sufficiently sensitive to study such interactions, so that a parallel technology will emerge for retrospective surveys.

Behavioral Criteria Must Drive Research Design The study design must inform us as to the type and amount of effort required to achieve behavioral criteria rather than continuing to perform studies that fix the dosage of the intervention. This third guideline recom- mends that the design of our studies must focus more on what efforts and resources are necessary and sufficient to attain specific levels of outcome. Specifically, we must seek to redesign our research questions, hypotheses, and methods so that we can estimate the type and amount of effort needed to attain a given range of behavioral out- comes. This requires that we move away from studies that fix dosage (i.e., number of sessions) and move toward designs that seek to describe the type and amount of effort likely to attain specific levels (ranges) of behavioral outcome.

There is a need to investigate the relationships be- tween efforts (costs) for an intervention to attain a desired behavioral outcome criteria rather than continuing to perform studies that fix the dosage of the intervention. One way of summarizing the type and amount of efforts required by an intervention strategy is by estimating its costs. Costs should be seen as an estimate of the efforts or resources consumed by the one or more parties in- volved directly or indirectly with the therapeutic inter- vention. What resources are expended by the client, the service provider-therapist, the family-support system, or the community-society? There is a homily that describes the cost-effectiveness issue for the service provider-ther- apist: Every decision made about a client's clinical or functional status is related to resources (and their costs) needed to help the person. Likewise, every decision made about the allocation of resources (and their costs) is related to the potential impact a service can have on a client's clinical status-functioning. We hope that our science will provide adequate cost-effectiveness data to support the decisions about allocating the resources needed to achieve desired outcomes.

The technology to estimate costs of resources and relate these to outcomes has been in our literature for quite some time (Carter & Newman, 1975/1980; Newman & Howard, 1986; Yates & Newman, 1980a, 1980b) and has been updated (Clark et al., 1994; Knapp, 1995; Yates, 1996). However, few research groups have considered re- sources used (through costs or dosage) as a systematic feature of their research program. Four groups do stand out as potential role models. The research groups at the University of Wisconsin--Madison and the Indiana Uni- ve r s i t y -Pu rdue University at Indianapolis have focused on assertive communi ty treatment (Bond, 1991; Weis- brod, 1983) for persons with a persistent and severe men-

1044 October 1996 • American Psychologist

tal illness. The research group at Dar tmouth -New Hampshire Psychiatric Research Center (Clark et al., 1994; Drake et al., 1996) has focused on persons with a dual diagnosis of a severe and persistent mental illness and substance abuse. Howard's (Howard, Moras, Brill, Martinovich, & Lutz, 1996) group at Northwestern Uni- versity has focused on persons seeking psychotherapy, and these have primarily been persons with an affective dis- order or a situational crisis.

Another technology under development is that of linking level of care to level of need (Carter & Newman, 1975/1980; Left, Graves, Natkins, & Bryan, 1985; Left, Mulkern, Lieberman, & Raab, 1994; Newman, Griffin, Black, & Page, 1989; Uehara, Smukler, & Newman, 1994). This technology, as currently understood, assumes that a necessary first step is to perform a holistic assess- ment profile of a consumer population that partitions the population into subgroups that have homogeneous needs for service resources (cost-homogeneous groupings). The profile is described as holistic because it considers physical needs along with psychological, social-interpersonal, and community-functioning needs. Those who work with programs that serve persons with a severe and persistent mental illness have had to face the need for this holistic view in designing and evaluating mental health services since the early 1970s when deinstitutionalization began (Carter & Newman, 1975/1980; Leffet al., 1985). A more specific description of a cost-homogeneous group is one that describes persons with a similar physical and psy- chosocial profile as requiring similar service needs to achieve a given set of behavioral criteria in a given t ime frame. It should be noted that, for some cost-homoge- neous groups, the behavioral criteria may be to maintain functioning in a given range. For others, the criteria might be to improve functioning.

An idealized example describing how behavioral criteria can be used to evaluate the relative cost-effec- tiveness of two or more groups would be useful at this point. This example uses technology that is currently available (with one limitation that we discuss later). The intended message in reviewing this example is that by using existing technology, one can describe the impact of two or more interventions in terms of its cost-effectiveness in achieving observable criteria. The example illustrated in Figure 1 offers a graphical analysis in the form of a progress-outcome report for three hypothetical groups of individuals. Persons in these three groups have been identified as having similar characteristics in their cost- homogeneous grouping. In this example, all three groups have similar initial levels of functioning and have service plans for a 26-week period (182 days), so that the expected costs of administering the treatment plan are the same. However, two groups differ with regard to their behavioral goals. Group 1 is expected to improve in their overall functioning to a point in which they can be placed in another cost-homogeneous grouping requiring a less in- tensive-expensive service plan. Group 2 has a goal of maintaining their functioning in the bounds described as

acceptable to community functioning. Group 3 represents outliers from Group 2 and is discussed later.

The vertical axis of Figure 1 represents a person's overall ability to function in the community. It is under- stood that this global measure must be supported by a multidimensional view of the persons in each group (Newman & Ciado, 1994). The horizontal axis represents the cumulative costs of providing services from the be- ginning of this episode of treatment. The two horizontal dotted lines inside the box represent the behavioral cri- teria set a priori (i.e., the lower and upper bounds of func- tioning for which this mental health service is designed to serve). I f a consumer is observed to behave at a level below the lower dotted line, then she or he should be referred to a more intensive service. Likewise, if a con- sumer is observed to behave at a level above the upper dotted line, then either services should be discontinued or a referral to a less intensive-expensive service should be considered. Finally, the dashed vertical line inside the box represents the cumulative costs of the planned ser- vices for the 26-week period. In this hypothetical case, all three groups have treatment plans with the same ex- pected costs.

T h e circled numbers in Figure 1 track the average progress of consumers in each of the groups over 13 suc- cessive 2-week intervals (a total of six months). The ver- tical placement of the circled number represents the group's average level of functioning at that time, and the horizontal placement represents the group's average costs of services up to that point. The sequence of the 13 circled numbers represents the average progress of persons in each of the groups in 2-week intervals over the 26 weeks of care. For Group 1, with a six-month objective of im- provement above that of the upper dotted line, the ob- jective is met. For Group 2, with the six-month objective of maintaining functioning within the range represented by the area between two dotted lines, the objective is also met. Group 3 is an example of clients who exceeded the bounds of the intended service-treatment plan after four months, two months shy of the objective. These consum- ers were able to maintain their communi ty functioning but only with the commitment of additional resources. A continuous quality improvement program should focus a posthoc analysis on these consumers to determine if they belong to another cost-homogeneous subgroup or whether the services were provided as intended and whether modification of current practices are needed.

A major difficulty in attempting to enact these rec- ommendations is that of obtaining a believable database to determine what the appropriate progress-outcome criteria or what the expected social or treatment costs should be. Some have used expert panels to set the first draft of such criteria (e.g., Newman et al., 1989; Uehara et al., 1994). Howard et al. (1996) have used baseline data from prior studies, along with measures taken on a non- patient population. However, both strategies have prob- lems. The research literature is largely based on studies in which the dosage was fixed within a study, thereby constricting inferences about what type and amount of

October 1996 • American Psychologist 1045

Figure 1 Hypothetical Example of a Graphical Analysis of Changes in Functioning and Costs Relative to Managed Care Behavioral Criteria for Three Cost-Homogeneous Groups of Consumers

Note. Group 1 hod the planned objective of improvement in functioning so that they could move to o less intensive service plan within six months. Group 2 was expected to maintain functioning over the same six months but to use about the same amount of resources. Group 3 was an outlier group who required additional resources to maintain adequat e levels of community functioning. They ore highlighted here for a follow-up study by the program's quality improvement team.

effort can achieve specific behavioral criteria. Howard's early work on a dosage and phase model of psychotherapy was statistically confounded by combining the results of controlled studies in which number of sessions was fixed. Some of the naturalistic data came from persons who had no limit on sessions, and others did have limits on the number of sessions, which was imposed by third-party payers. Having worked closely with Howard over the last decade, it would appear the dosage data as reported are probably trustworthy. However, we know of no studies in which behavioral criteria were set a priori, and the type and amount of efforts were seen as the dependent variables in estimating the success or failure to achieve these criteria. Yet the logic of managed care and the logic of practice guidelines under discussion by the American Psychological Association and the American Psychiatric Association would require that such study results be available to set criteria for using a particular intervention strategy or to set reimbursement standards. This void must be filled by data from well-designed efficacy and

cost-effectiveness studies that can provide empirical sup- port for setting behavioral outcome criteria for managed care programs. Without such data, there is faint hope of changing the current practice of setting dosage guidelines independent of behavioral criteria.

How would one go about establishing such cri- teria? The design of a project currently underway for the Indiana Division of Mental Health represents our ~ understanding of what the literature recommends in the constraints of the real world of Indiana. The project is designed to provide an empirical base for a team of actuaries to set these criteria approximately 2.5 years

1 Richard DeLiberty, deputy director of mental health, is the overall project director and directs the consumer telephone survey project. Kay Hodges of Eastern Michigan University directs the development of the assessment instrument for children and adolescents. John McGrew of Indiana University--Purdue University of Indianapolis directs the field research team activities. Manuel J. Tejeda provides statistical consultation. Frederick L. Newman leads the development of the adult assessment instrument and directs the technical aspects of the project.

1046 October 1996 • American Psychologist

after the necessary field research is completed. The re- search and implementa t ion plan calls for a two-tiered approach with two phases of data collection in each tier. One tier involves telephone surveys with a ran- domly selected cadre of consumers currently receiving services. The i tems collected in the telephone survey resemble those asked in the CR (1994) survey plus i tems derived from the second tier of the research plan. The second tier involves collecting mult idimensional assessment and service data f rom the service providers. For both tiers, the first phase involves developing in- s t ruments and testing the psychometr ic qualities of these instruments. In the first phase of the second tier, in which assessment and service data are collected from the service providers, the focus is on test and devel- opment of methods for training and auditing the mul- t idimensional assessment instruments . Initial pilot ac- tivities are designed to assure psychometr ic quality and to revise the ins t rument or procedures accordingly. The pilots include well-controlled quasi-experimental studies done in a selected set of communi ty mental health centers (CMHCs) before field testing the pro- cedures across all of Indiana 's 30-plus CMHCs. The adult ins t rument ' s development involved input from all stakeholder groups (including consumers). The se- lection of ins t rument i tems and the molding of its s t ructure followed from several sessions of values clar- ification with the stakeholders about the objectives of Indiana ' s managed care plan and the use of the assess- ment instrument . The ins t rument ' s development pro- cess used these values as a guide while adapting i tems f rom the research literature. It should be no surprise that the ins t rument covers s ym p t om distress; personal, interpersonal, and vocational functioning; substance abuse; and sel f -management skills in dealing with ex- isting symptoms or problems. The unique feature here was the stakeholder panel ' s insistence that the language of symp tom or p rob lem intensity be described jointly in te rms of the symptoms and the person's ability to manage his or her functioning given that pat tern of symptoms. Record audits and training (and retraining) are considered critical to this process and will be con- tinued after the system is installed. In the second phase, assessment and service data are tracked to identify cost- homogeneous groupings, as described earlier. Initial criteria will be based on data for the 3rd through the 15th month after Phase 2 is initiated. The criteria will be reviewed semiannual ly and revised either annually or biannually thereafter.

Colleagues have described similar projects now in progress in other states, so one could expect that we will be seeing better data to develop these criteria in the future. It would be wonderful to rewrite this section in half a decade, citing a rich line of randomized control trials and quasi-experimental and cost-effectiveness studies that used behavioral criteria to define successful outcome and to generate the data used to guide the policies of managed care.

Research on Communicating With All Stakeholders How must information be formatted to best support the decisions of the key stakeholders (clients-families, clini- cians, service managers, and policymakers)? This fourth guideline serves to remind the scientific communi ty that research cannot exist in a vacuum: We have a tendency to talk to ourselves. Scholars must learn to communicate their results effectively to all consumers of their research and not simply to one another. We must keep in view each of the key stakeholder groups when exploring and designing how we will communicate our findings. Here, we are indebted to scholars such as Seligman (1995) and Howard and his colleagues (1996) for furthering diagnosis and treatment as scientists, while exploring new ways to test and communicate the findings of the research to those concerned with the delivery of services.

There is a need to perform studies on the use of effectiveness and evaluation data in supporting managed care decisions by the stakeholders. As described earlier, there are at least four sets of stakeholders when consid- ering the managed care environment in its broadest sense: (a) the client or family in their choice of modality or clinician, (b) the clinical staff in supporting their selection of therapeutic strategy, (c) the management staff in the allocation of resources that meet the needs and can po- tentially achieve the behavioral outcome criteria, and (d) those who set funding or accreditation policy about the linkage between level of need and level of resource avail- ability in a managed care environment. Whether it is the family, the clinician, the employer, the insurance carrier, or a government entity, the central issue is how to best manage the available resources to meet the need and achieve a desired outcome. Study designs must assess re- sources used with the traditional measures of initial status, and progress and outcome should support client-family, clinical, supervision, management, and policy decisions in a managed care environment.

There has been a start in the development of con- sumer-oriented report cards on the access, availability, outcome, and satisfaction with mental health services. Mulkern, Left, Green, and Newman (1995) reviewed the published and unpublished materials that were available in the spring and winter of 1995. The technology of de- veloping performance indicators on accessibility, avail- ability, and satisfaction does exist. Kamis-Gould (1978) and Sorensen and his colleagues (Sorensen, Zelman, Hanbery, & Kucic, 1978) developed a battery of perfor- mance indicators to describe access, availability, and fi- nancial stability viability of mental health services. Davies and Ware ( 1991) developed a measure for the consumer satisfaction for the Group Health Association of America that has reasonable psychometric properties.

Outcome measures from the viewpoint of the con- sumer are not readily available. Yes, there are many out- come measures that are described as consumer self-re- ports (e.g., the SF-36 [Ware, Snow, Kosinski, & Gandek, 1993] or the SCL-90 [Derogatis, 1975]), but these have

October 1996 • American Psychologist 1047

been der ived by menta l heal th professionals to es t imate impac t f rom the v iewpoin t o f a men ta l heal th profes- sional. Measures der ived f rom consumers ' views are still being developed and tested for thei r psychomet r i c qual i ty (Mulkern et al., 1995).

The Center for Menta l Hea l th Services ( C M H S ) cre- ated a task force of consumer representat ives and p rogram evaluators to develop such an i n s t rumen t under the Men- tal Heal th Systems Improvemen t Program (MHSIP) . The task force reviewed all o f the in s t rumen t s available and gathered in fo rma t ion f rom a var ie ty o f consumer focus groups (some using fairly sophis t ica ted mul t id imens iona l scaling procedures) . The task force draf ted an ins t rument that is now available. A n u m b e r of states, inc luding In- d iana , are a t t emp t ing to develop and test the use o f their own menta l heal th services r epor t card, using the C M H S mode l as a guide.

The concept o f packag ing research and evaluat ion f indings for the p r inc ipa l sets of s takeholders is not cur- rent ly a focus o f our science or seen as the concern o f the cl inical services researcher. Yet it is centra l to the success o f our field having impac t on managed care guidelines. It may be t ime for us to review the research pe r fo rmed by those in decision sciences, management , and market ing to bet ter unde r s t and how in fo rma t ion can be packaged so tha t the i n fo rma t ion is useful to consumers and thei r families, p rac t ic ing c l in ic ians and their admin is t ra to rs , and po l i cymakers inf luencing the rules o f managed care.

Conclusion The C R (1995) ar t ic le and Se l igman (1995) are to be c red i ted for their provocat ions . O u r in ten t ions have been to share some o f the l imi ta t ions we and others have ob- served in the state o f the science. O u r work as scholars has m a d e great in roads in t rea tment . However, the changing and tu rbu len t env i ronmen t o f heal th care now d e m a n d s a b roade r vision.

We mus t cont ras t the b roade r vision with the de- m a n d for specificity that is an in tegral pa r t o f the social sciences. Indeed, our studies and careers reflect the nar row vision requi red in our research. The s tandard recom- menda t ion in our texts (e.g., Meyers & Well, 1991) is tha t we should seek to m a x i m i z e sys temat ic var iance, min i - mize e r ror var iance, and cont ro l ex t raneous variables. As we strive to create designs o f greater in ternal validity, the consumers of our research become increasingly less central .

O u r studies, a l though scientifically thorough, typi- cally are not devised to assess the real -world impac t o f our in tervent ions . O u r science does jus t i fy cer ta in forms of t r ea tment , bu t we t rad i t iona l ly fail to apprec ia te the manager ia l , legal, and social ramif ica t ions resul t ing f rom endors ing one in te rven t ion over another. Al though we canno t be expected to assess every repercuss ion o f our research, we can begin to inc lude in our research the c o m p o n e n t s tha t begin to help us assess some o f these repercussions.

Too often, menta l heal th care decis ions are made because of fiscal constraints . In par t , we are to b lame.

On the one hand, the costs o f menta l heal th services are t es tament to the increas ing complexi t ies o f providing ser- vices. On the other hand, the technologies have existed for some t ime for us to incorpora te therapeut ic effort, level of care, and cost in fo rmat ion in our research (New- man & Howard, 1986). We have arr ived in t imes in which technologies no t only exist bu t also in which appl ica t ions are requi red to s imul taneous ly answer the quest ions of science, clients, providers, managers , and pol icymakers . This ar t ic le has followed that theme: Technology now exists at such a level tha t we can adap t our techniques to the in fo rmat ion d e m a n d s of the stakeholders in the (managed care) env i ronmen t o f menta l heal th service de- livery. I f we re turn to one quest ion we proposed in the beginning of the article, the c l i en t -consumer is still asking, "Wi l l I be able to funct ion a d e q u a t e l y . . . . W h a t is re- qu i red o f me, and can I afford the services?" Our studies most move to answer, "Yes, and this is how."

REFERENCES

Aiken, L. S., & West, S. G. (1993). Multiple regression: Testing and interpreting interactions. Newbury Park, CA: Sage.

Berman, J. S., & Norton, N. C. (1985). Does professional training make a therapist more effective? Psychological Bulletin, 98, 401-407.

Beutler, L. E. (1991). Have all won and must all have prizes? Revisiting Luborsky et al.'s verdict. Journal qf Consulting and Clinical Psychology, 59, 226-233.

Bond, G. R. (1991). Variations in an assertive outreach model. New Directions in Mental Health Services, 52, 65-80.

Bond, G. R. (1992). Vocational rehabilitation. In R. Liberman (Ed.), Handbook of psychiatric rehabilitation (pp. 244-275). New York: MacMillan.

Bond, G. R., McGrew, J. H., & Fekete, D. (1995). Assertive outreach for frequent users of psychiatric hospitals: A meta-analysis. Journal of Mental Health Administration, 22. 4-16.

Bronowski, J. (1959). &:ience and human values. New York: Harper & Row.

Bryk, A. S., & Raudenbush, S. W. (1988). Hetrogeniety of variance in experimental studies: A challenge to conventional interpretations. Psychological Bulletin, 104, 396-404.

Carter, D. E., & Newman, F. L. (1980). A client oriented system of mental health service delivery and program management: A workbook andguide (DHEW Publication No. ADM 76-307, FN Series No. 4). Washington, DC; U.S. Government Printing Office. (Original work published 1975)

Clark, R. E., Teague, G. B., Ricketts, S. K., Bush, P. W., Keller, A. M., Zubkoff, M., & Drake, R. E. (1994). Measuring resources used in economic evaluations: Determining the social costs of mental illness. Journal of Mental Health Administration. 21, 32-41.

Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49, 997-1003.

Consumer Reports. (1994). Annual questionnaire. Consumer Reports. (1995, November). Mental health: Does therapy help?

734-739. Cronbach, L. J. (1992). Four P.~w'hological Bulletin articles in perspective.

Psychological Bulletin, 112, 389-392. Cronbach, L. J., & Furby, L. (1970). How should we measure "change"--

or should we? Psychological Bulletin, 74, 68-80. Davies, A. R., & Ware, J. E., Jr. (1991). GHAA's ConsumerSatisJaction

Survey (2nd ed.). Washington, DC: Group Health Association of America.

Derogatis, L. R. (1975). The SCL-90. Baltimore: Clinical Psychometric Research.

Drake, R. E., McHugo, G. J., Becker, D. R., Anthony, W. A., & Clark, R. E. (1996). The New Hampshire study of supported employment for people with severe mental illness: I. Vocational outcomes. Journal of Consulting and Clinical Psychology 64, 391-399.

1048 Oc tober 1996 • Amer i can Psychologist

Greenberg, L. S., & Newman, E L. (1996). An approach to psychotherapy process research: Introduction to the special series. Journal of Con- suiting and Clinical Psychology, 64, 435-438.

Heinsman, D. T., & Shadish, W. R. (1996). Assignment methods in experimentation: When do nonrandomized experiments approximate the answers from randomized experiments? Psychological Methods, 1, 154-169.

Hogarty, G. W., Goldberg, S. C.. Schooler, N. R., & Ulrich, R. F. (1974). Drug and sociotherapy in the aftercare of schizophrenic patients: I1. Two year relapse rates. Archives of General Psychiatry, 31. 603-606.

Howard, K. I., Moras, K., Brill, E L., Martinovich, Z., & Lutz, W. (1996). Evaluation of psychotherapy: Efficacy, effectiveness, patient progress: American Psychologist, 51, 1059-1064.

Jaccard, J., & Wan, C. K. (1995). Measurement error in the analysis of interaction effects between continuous predictors using multiple regression: Multiple indicator and structural equation approaches. Psychological Bulletin, 117, 348-357.

Kamis-Gould, E. (1978). The New Jersey performance management system: A state system using simple measures. Evaluation and Program Planning, 10, 249-255.

Karon, B. P., & VandenBos, G. R. ( 1981 ). Psychotherapy of schizophrenia: The treatment of choice. New York: Jason Aronson.

Knapp, M. (Ed.). (1995). The economic evaluation of mental health care. London: Arena, Aldershot.

Left, H. S., Graves, S. C., Natkins, J., & Bryan, J. (1985). A system for allocating mental health resources. Administration in Mental IIealth, 13, 43-68.

Left, H. S., Mulkern, V., Lieberman, M., & Raab, B. (1994). The effects of capitation on service access, adequacy, and appropriateness. Ad- ministration and Policy in Mental Health, 21, 141-160.

Lipsey, M., & Wilson, D. (1993). The efficacy of psychological, educa- tional, and behavioral treatment: Confirmation and recta-analyses. American Psychologist, 48, 1181-1209.

Lyons, J. S., & Howard, K. I. (1991). Main effects analysis in clinical research: Statistical guidelines for disaggregating treatment groups. Journal of Consulting and Clinical Psychology, 59. 745-748.

May, P. R., Tuma, A. H., & Dixon, W. J. (1981). Schizophrenia--A follow-up study of the results of 5 forms of treatment. Archives of General Psychiatry, 38, 776-784.

McClelland, G. H., & Judd, C. M. (1993). Statistical difficulties of de- tecting interactions and moderator effects. Psychological Bulletin, 114, 376-389.

McGrew, J. H., Bond, G. R., Dietzen, L., & Salyers, M. (1994). Measuring the fidelity of implementation of a mental health program model. Journal of Consulting and Clinical Psychology, 62, 670-678.

Meyers, J. L., & Well, A. D. (1991). Research design and statistical analysis. New York: HarperCollins.

Mintzberg, H. (1973). The nature of managerial work. New York: Harper & Row.

Mulkern, V., Left, S., Green, R. S., & Newman, E L. (1995). Performance indicators for a consumer-oriented mental health report card: Liter- ature review and analysis. In V. Mulkern (Ed.), Stakeholder perspectives on mental health performance indicators (Section 2). Cambridge, MA: The Evaluation Center at HSRI.

Newman, E L. (1994). Selection of design and statistical procedures for progress and outcome assessment. In M. Maruish (Ed.), Use of psy- chological testing for treatment planning and outcome assessment (pp. 111-134). Hillsdale, NJ: Erlbaum.

Newman, E L., & Carpenter, D. (in press). A primer on test instruments. In L. J. Dickstein, M. B. Riba, & J. M. Oldham (Vol. Eds.) & J. F. Clarkin & J. Docherty (Section Eds.), American Psychiatric Press review of psychiatry (Vol. 16). Washington, DC: American Psychiatric Press.

Newman, E L., & Ciarlo, J. A. (1994). Criteria for selecting psychological instruments for progress and outcome assessment. In M. Maruish (Ed.), Use of psychological testing for treatment planning and outcome assessment (pp. 98-110). Hillsdale, N J: Erlbaum.

Newman, F. L., Griffin, B. P., Black, R. W., & Page, S. E. (1989). Linking level of care to level of need: Assessing the need for mental health care for nursing home residents. American Psychologist, 44, 1315- 1324.

Newman, F. L., & Howard, K. I. (1986). Therapeutic effort, outcome, and policy. American Psychologist, 41, 181-187.

Schmidt, E L. (1996a). APA Board of Scientific Affairs to study issue of significance testing, make recommendations. The Score Newsletter, 19, 1,6.

Schmidt, F. L. (1996b). Statistical testing and cumulative knowledge in psychology: Implications for the training of researchers. Psychological Methods, 1, 115-129.

Seligman, M. E. P. (1995). The effectiveness of psychotherapy: The Con- sumer Reports study. American Psychologist, 50, 965-974.

Shadish, W. R., & Ragsdale, K. (in press). Random versus nonrandom assignment in controlled experiments: Do you get the same answer? Journal of Consulting and Clinical Psychology,

Shoham-Salomon, V. (1991). Introduction to special section on client- therapy interaction research. Journal of Consulting and Clinical Psy- chology. 59, 203-204.

Shoham-Salomon, V., & Hannah, M. T. (1991 ). Client-treatment inter- action as a framework for research on individual differences in dif- ferential change processes. Journal of Consulting and Clinical Psy- chology, 59, 217-225.

Sorensen, J. E., Zelman, W., Hanbery, G. W., & Kucic, A. R. (1978). Managing mental health organizations with 25 performance indicators. Evaluation and Program Planning, 10, 239-247.

Stein, L. 1., & Test, M. A. (1980). An alternative to mental treatment: I. Conceptual model, treatment program, and evaluation. Archives of General Psychiatry, 37, 392-397.

Uehara, E. S., Smukler, M., & Newman, F. L. (1994). Linking resource use to consumer level of need: Field test of"LONCA" method. Journal of Consulting and Clinical Psychology 62, 695-709.

Ware. J. E., Snow, K. K., Kosinski, M., & Gandek, B. (1993). SF-36 Health Survey. Manual and interpretation guide. Boston: The Health Institute, New England Medical Center.

Weisbrod, B. A. (1983). A guide to benefit-cost analysis, as seen through a controlled experiment in treating the mentally ill. Journal of Health Politics, Policy, and Laa; 7, 808-845.

Willett, J. B., & Sayer, A. G. (1994). Using covariance structural analysis to detect correlates and predictors of change over time. Psychological Bulletin, 116, 363-381.

Yates, B. T. (1996). Analyzing costs, procedures, process, and outcomes in human services. Thousand Oaks, CA: Sage.

Yates, B. T., & Newman, E L. (1980a). Approaches to cost-effectiveness and cost-benefit analysis of psychotherapy. In G. VandenBos (Ed.), Psychotherapy: From practice to research to policy (pp. 103-162). Beverly Hills, CA: Sage.

Yates, B. T., & Newman, E L. (1980b). Findings of cost-effectiveness and cost-benefit analysis of psychotherapy. In G. VandenBos (Ed.), Psychotherapy: From practice to research to policy (pp. 163-185). Beverly Hills, CA: Sage.

O c t o b e r 1996 • A m e r i c a n P s y c h o l o g i s t 1049