PMCCPMCCPMCC

Search tips
Search criteria 

Advanced

 
Logo of nihpaAbout Author manuscriptsSubmit a manuscriptNIH Public Access; Author Manuscript; Accepted for publication in peer reviewed journal;
 
Qual Life Res. Author manuscript; available in PMC Oct 7, 2013.
Published in final edited form as:
PMCID: PMC3791923
NIHMSID: NIHMS516820
PROMIS® Parent Proxy Report Scales: an item response theory analysis of the parent proxy report item banks
James W. Varni, David Thissen, Brian D. Stucky, Yang Liu, Hally Gorder, Debra E. Irwin, Esi Morgan DeWitt, Jin-Shei Lai, Dagmar Amtmann, and Darren A. DeWalt
James W. Varni, Department of Pediatrics, College of Medicine, Department of Landscape Architecture and Urban Planning, College of Architecture, Texas A&M University, 3137 TAMU, College Station, TX 77843-3137, USA;
jvarni/at/arch.tamu.edu
Objective
The objective of the present study is to describe the item response theory (IRT) analysis of the National Institutes of Health (NIH) Patient Reported Out-comes Measurement Information System (PROMIS®) pediatric parent proxy-report item banks and the measurement properties of the new PROMIS® Parent Proxy Report Scales for ages 8–17 years.
Methods
Parent proxy-report items were written to parallel the pediatric self-report items. Test forms containing the items were completed by 1,548 parent–child pairs. CCFA and IRT analyses of scale dimensionality and item local dependence, and IRT analyses of differential item functioning were conducted.
Results
Parent proxy-report item banks were developed and IRT parameters are provided. The recommended unidimensional short forms for the PROMIS® Parent Proxy Report Scales are item sets that are subsets of the pediatric self-report short forms, setting aside items for which parent responses exhibit local dependence. Parent proxy-report demonstrated moderate to low agreement with pediatric self-report.
Conclusions
The study provides initial calibrations of the PROMIS® parent proxy-report item banks and the creation of the PROMIS® Parent Proxy-Report Scales. It is anticipated that these new scales will have application for pediatric populations in which pediatric self-report is not feasible.
Keywords: PROMIS®, Parent proxy report, Item response theory
The Patient Reported Outcomes Measurement Information System (PROMIS®) is a National Institutes of Health (NIH) Initiative, created to advance the assessment of patient-reported outcomes (PRO) in chronic diseases. Items are evaluated using item response theory (IRT) to derive scales with scores that are maximally reliable and valid along the full spectrum of the latent trait [1]. A primary objective is to develop item banks and computerized adaptive tests (CAT) across a variety of chronic disorders [2]. During the past 7 years, the PROMIS® Pediatric Cooperative Group has developed pediatric self-report item banks for ages 8–17 years across five generic health domains (physical function, pain, fatigue, emotional health, and social health) consistent with the larger PROMIS® network [3]. It was anticipated that measures of these five generic health domains would be applicable across pediatric chronic health conditions and hence were developed as generic or non-disease-specific scales [4-7]. Additionally, an asthma-specific item bank was developed and validated [8].
It has been well documented in both the adult and pediatric literature that information provided by proxy-respondents is not equivalent to that reported by the patient [9, 10]. Imperfect agreement between self-report and proxy-report, termed cross-informant variance [11], has been consistently documented in the HRQOL measurement of children with chronic health conditions and healthy children [12]. However, even as pediatric patient self-report is advocated, there remains a role for parent proxy-report in pediatric clinical trials and health services research.
While pediatric patient self-report should be considered the standard for measuring PROs, there may be circumstances when the child is too young, too cognitively impaired, too ill or fatigued to complete a PRO instrument, and parent proxy-report may be needed in such cases [13]. Further, it is typically parents’ perceptions of their children’s health and well-being that influences healthcare utilization [14-16]. Thus, instruments should be developed that measure the perspectives of both the child/adolescent and parent since these perspectives may be independently related to healthcare utilization, risk factors, and quality of care [17].
The majority of parent proxy-report scales, consistent with other clinical assessment instruments [18], have utilized classical test theory (CTT) and have rarely taken advantage of IRT analysis in the scale development process [19]. By utilizing IRT analysis, the resulting item bank can be the basis of a more customizable measure for meeting a researcher’s or clinician’s needs. Depending on the desired level of precision, the evaluator can then select the number of items to administer and obtain scores on the same metric as all other users of this item bank [19].
The objective of the present study is to address this measurement gap in the parent proxy-report literature by describing the IRT analysis of the PROMIS® parent proxy-report item banks and the measurement properties of the new PROMIS® Parent Proxy Report Scales, including investigations of scale dimensionality and sources of local dependence and differential item functioning.
Participants
Participants were recruited between May 2008 through March 2009 in hospital-based outpatient general pediatrics and subspecialty clinics. Clinic participants were identified through a review of clinic appointment rosters or while in the clinic waiting rooms according to protocols approved by the institutional review boards (IRBs) of University of North Carolina (UNC), Duke University Medical Center, University of Washington (UW), Children’s Memorial Hospital, Chicago (CMH), and Children’s Hospital at Scott and White (S&W) in Texas. Pediatric patients within the appropriate age range who had clinic appointments and their caregivers were recruited while waiting for their clinic appointments. The UNC, Duke, UW, CMH and S&W general pediatric clinics were representative of health issues for which children have physician office visits (e.g., well child visits, acute illnesses, and some chronic illnesses). The specialty clinics included Pulmonology, Allergy, Gastroenterology, Rheumatology, Nephrology, Obesity, Rehabilitation, Dermatology, and Endocrinology. Children with asthma were over sampled during recruitment because asthma-specific items were tested.
To be eligible to participate in the large-scale testing survey, all participants were required to meet the following inclusion criteria: able to speak and read English; and able to see and interact with a computer screen, keyboard, and mouse. Children enrolled were between the ages of 8 and 17 years and with their parents/guardians formed a dyad (for convenience, we refer to the dyad as simply parent/child). Both members of the dyad were required to individually complete the items. Children completed the self-report version, and parents completed the proxy-report version.
Parents signed an informed consent document, and children signed an informed assent document that outlined the following: purpose of the study, participation requirements, potential benefits and risks of participation, and the measures implemented to protect participant privacy. Both the informed assent and the informed consent were administered in English, so parents were also required to read and speak English. Each participant received a $10 gift card in return for their time and effort. The study protocols were approved by the institutional review boards at each institution.
Pediatric self-report item bank development
The PROMIS® Pediatric item banks were developed using a strategic item generation methodology adopted by the PROMIS® Network [2]. Six phases of item development were implemented: identification of existing items, item classification and selection, item review and revision, focus group input on domain coverage, cognitive interviews with individual items, and final revision before field testing. The final pediatric self-report item banks contained 165 items across the 5 generic health domains (physical function, pain, fatigue, emotional health, and social health) and Asthma. Because physical function includes both upper extremity and mobility item banks, emotional distress includes separate anger, anxiety and depressive symptoms item banks, and fatigue includes both fatigue and lack of energy item banks, a total of 10 content domains were tested [4-8].
Parent proxy sampling plan and item distribution
The parent proxy-report items were developed from the 10 existing pediatric self-report content domains [4-8]. The items were revised to retain their meaning, while modifying the phrasing so that all items involved parents reporting on their child. For example, in the pediatric self-report pain interference domain [6], children responded to the item “I had trouble sleeping when I had pain,” while parents responded to the parent proxy-report equivalent of this item, “My child had trouble sleeping when he/she had pain.”
Proxy-report short form items were selected from items that were on the pediatric self-report short forms for each domain and did not include any items that were not already on the self-report short forms. This decision was made because most researchers prefer parent proxy-report item banks that have the same item content as the pediatric selfr-eport item banks.
The 293 proxy-report items from the 10 content domains were administered to 1,548 parents of 8- to 17-year-old children also participating in the study (see Table 1 for sample demographics and Table 2 for content domains). To reduce respondent burden, a multi-form design was used in which the items were divided among nine test forms, and each parent was administered one of the nine forms. To ensure an adequate number of individuals responded to all items, each item appeared on three of the forms. This process resulted in all items being administered to at least 428 parents. A detailed description of the item administration protocol can be found in Irwin et al. [20].
Table 1
Table 1
Parent and child demographics
Table 2
Table 2
Means and standard deviations of the pediatric self-report measures in the matchec parent–child samples, and the correlations between latent variables measured by the pediatric self-report and parent proxy-report scales
All items had a 7-day recall period and used standardized 5-point response options (e.g., never, almost never, sometimes, often, almost always; or, with no trouble, with a little trouble, with some trouble, with a lot of trouble, not able to do). Of the 293 items administered, 165 were retained for analysis, as these corresponded to the final items in the pediatric self-report item banks. A complete list of the 165 items may be found in the “Appendix 1”.
Statistical and psychometric methods
The purpose of the psychometric assessment of the parent proxy-report items was to develop parent content domain short forms that could serve as proxies for their child counterpart domain short forms. Based on the methods used in the development of the pediatric item banks [4], ten content domain-specific analyses were conducted which investigated scale dimensionality and sources of local dependence (LD).
LD occurs when items retain an association after accounting for inter-item correlations with the latent variable. Researchers typically identify locally dependent item pairs or clusters either using categorical confirmatory factor analysis (CCFA) models that account for LD with correlated residuals and subfactors, or using IRT methods involving model-based pairwise residuals. We employed both methods (CCFA and IRT) to identify potential LD.
Initially, we conducted CCFAs of the inter-item polychoric correlation matrices using Mplus [21]. Because this approach is conducted using complete case data, and our sampling design assigned only subsets of items from each content domain to any given parent, multiple CCFAs (usually 2) were conducted on independent subsets of items. This approach involved fitting 15 different CCFA models in which each analysis included at least 175 parents. We began with a single-factor model and estimated additional correlated residuals and subfactors using item content and statistical relevance for guidance. To quantify the degree to which each scale is approximately unidimensional, we used the value of “explained common variance” (ECV) attributable to the first (general) factor of the final bifactor model for each scale within each combined form [22].
To further investigate potential LD, we calibrated each content domain using Samejima’s graded response model (GRM) in an IRT framework using IRTPRO [23]. Following these initial calibrations, we tested for LD between pairs and clusters of items using a standardized χ2 statistic as implemented in IRTPRO. The Chen-Thissen X2 [24] is based on a comparison between the observed and IRT-expected response frequencies for pairs of items.
These analyses provided two different assessments of LD. Our initial assessment of the validity of these tests of LD was made by comparing the magnitude of the CCFA-based residual correlation to the magnitude of the IRT-based standardized x2 statistic. In making these comparisons, we note that the two methods use slightly different respondents; the IRT analyses used the same individuals as the CCFA analyses, and additional parents. Where there was disagreement between the two methods, a team of six content experts evaluated the validity of the potential LD. The expert panel reviewed all pairs and triplets of locally dependent items that contained more than one item from the pediatric short form. If the LD χ2 statistic was greater than 10, item(s) were eliminated so that only one item from the LD pair or triplet remained. If the LD χ2 statistic was less than 10, each member of the panel voted on whether or not to delete item(s) and if so, which item to retain. Members made decisions about whether to eliminate item(s) by evaluating whether the item’s content was substantively different from others in the locally dependent group and offered unique information about the child’s experience. When members voted to delete item(s), decisions about which item to retain were based on a content analysis of which item best fit the conceptualization of the domain. Setting aside items from locally dependent pairs ensured that the IRT assumption of unidimensionality was maintained.
We tested for the presence of differential item functioning (DIF) between parents of children ages 8 through 12 and parents of adolescents ages 13 through 17, and separately between parents of male or female children. DIF was investigated using IRTPRO’s DIF module that implements Wald χ2 tests [25, 26]. In each case, the presence of DIF indicates that the relation between the item responses and latent variable differs between groups. Because many tests of DIF were conducted, we controlled for multiple comparisons using the Benjamini-Hochberg procedure [27]. A significant χ2 statistic indicates the presence of DIF. To evaluate the influence of DIF on item scores, we used graphical methods suggested by Steinberg and Thissen [28]. The goodness of fit of the IRT model to the data was examined using the S-X2 statistic [29] as implemented in IRTPRO. Non-significant S-X2 values suggest adequate fit of the model to the data.
In order to place the parent scales on the same metric as the child items, final calibrations of unidimensional domains were conducted using as the reference group mean and variance values (in Table 2) obtained from the child responses who had parents taking the same domain items; parents who did not have children taking the same domain items were included in the calibration as a separate group.
Parent proxy-report short forms were created after setting aside locally dependent items from the item subsets derived from the pediatric self-report short forms. The degree to which these short forms produce precise scores was evaluated using information functions. IRT conceptions of information allow score precision to vary across the range of the content domain. To ease interpretability, reliability is approximately one less the inverse of test information. Thus, when information is 10, reliability is 0.90.
We used two-dimensional IRT models to estimate the correlations between the latent variables measured by the pediatric self-report scales for ages 8–17 and the parent proxy-report scales for ages 8–17. In these models, one latent variable was used with the graded model to fit the pediatric self-report responses, and a second latent variable was used to fit the parent proxy-report data; the correlation between the two latent variables was estimated simultaneously with the item parameters. In terms used by traditional test theory, this correlation is an estimate of the “disattenuated” correlation between the two variables; that is, the correlation corrected for the presence of measurement error in scores. As a result, these correlations would be 1.0 if the pediatric self-report and parent proxy-report scales measured the same constructs.
Table 3 summarizes the results of the CCFA analyses of dimensionality, tabulating the value of ECV for the general factor for each scale for each form or pair of forms for which the sample size was sufficient. The scales are strongly univocal, with 82–100% of the common variance explained by the general factor. Some degree of local dependence was detectable as statistically significant in these analyses, and the item pairs or clusters involved are noted in the tables in the “Appendix 1”; however, the principle result of the factor analytic examination of the scales was that there is so little residual common variance from a single factor that for practical purposes, unidimensional modeling may proceed.
Table 3
Table 3
Explained common variance (ECV) for the general factor for each of the parent proxy scales, for the form combinations that were administered to parents
Tables 4, ,5,5, ,6,6, ,7,7, ,88 and and99 in the “Appendix 1” list the items (arranged by the magnitude of the IRT discrimination parameter), IRT parameter estimates, item fit, and DIF statistics for the 165 items retained for analysis and 10 content domains. The notes at the bottom of each table indicate with superscripts items that may exhibit LD and items that were set aside from each domain’s short form. Items marked with the same superscript letter within a domain remain in the item banks, but may exhibit LD, and so users are cautioned to use only one item from each such lettered set in any custom form or CAT. Because many CCFA and IRT models were used to assess the properties of the items, we summarize the results by content domain.
Table 4
Table 4
Emotional distress item parameters
Table 5
Table 5
Fatigue item parameters
Table 6
Table 6
Physical functioning item parameters
Table 7
Table 7
Pain interference item parameters
Table 8
Table 8
Peer relations item parameters
Table 9
Table 9
Asthma impact item parameters
Emotional Distress: LD and DIF
Emotional distress (Table 4) comprises depressive symptoms, anxiety, and anger. No significant DIF by sex or age was identified for these content domains. Among the depressive symptoms short form items, three pairs of items were identified as being locally dependent. From the pairs “My child felt lonely” and “My child felt alone”, and “My child felt sad” and “My child felt unhappy”, the second item was set aside from each LD pair, respectively, for the proxy short form. From the pair “My child felt everything in his/her life went wrong” and “My child felt like he/she couldn’t do anything right”, neither item was set aside after content review. There were three additional pairs of potentially locally dependent items identified in the anxiety domain; however, statistical tests of LD were mixed regarding these pairs, and no further items were set aside from the short form after item content review suggested that significant LD may be spurious.
Short forms were then made for the depressive symptoms (6 items), anxiety (8 items), and anger (5 items) domains using the items from the PROMIS® pediatric self-report short forms, setting aside two items on the depressive symptoms scales that exhibited LD. Figure 1 indicates that the short forms for depressive symptoms and anxiety produce information values greater than 10 from the mean of the latent variable to more than three standard deviations above the mean. The anger scale, containing just 5 items, produces scores with reliability greater than 0.80 from about two standard deviations below the mean to more than three standard deviations above the mean.
Fig. 1
Fig. 1
Test information functions for the parent proxy PROMIS® domain short forms for the depressive symptoms, anxiety, and anger content domains
Fatigue: LD and DIF
Results for fatigue (the “Lack of Energy” and “Tired” scales; Table 5) indicate two and potentially three locally dependent clusters of items, respectively. After reviewing the content of each LD cluster of items, the review panel concluded that each item provided a unique substantive contribution to the final scale; no items were set aside. Additionally, one item “My child had enough energy to do the things s/he likes to do” had significant DIF with respect to sex, and an additional item “My child needed to sleep during the day” had significant DIF related to age. Though the final short form includes these items, investigators may wish to exclude them if age and sex variables are used to differentiate levels of fatigue. The resulting 8 and 10 item short forms (Lack of Energy and Tired, respectively) both provide reliable scores from about one-half of a standard deviation below the mean to about three deviations above the mean (see Fig. 2).
Fig. 2
Fig. 2
Test information functions for the pain, peer relationships, asthma symptoms, upper extremity/dexterity, mobility, tiredness, and lack of energy content domains
Physical Functioning: LD and DIF
Results for physical functioning (upper extremity and mobility; Table 6) indicate a single instance of LD between the item pair “My child could do sports and exercise that other kids his/her age could do” and “My child has been physically able to do the activities he/she enjoys most”. The content review panel concluded that each item provides sufficient unique contribution to warrant inclusion in the final short form. Additionally, there was evidence of both age and sex DIF throughout both scales, but this was largely due to missing or sparsely endorsed response categories at the extreme ends of the distribution, and no items were set aside. The resulting 8-item short forms produce reliable scores between about one standard deviation to three standard deviations below the mean (see Fig. 2).
Pain Interference: LD and DIF
There were two potentially locally dependent pairs of items in the pain interference domain, marked with superscripts a and b in Table 7. All four of these items were considered to be substantively unique, and neither item was set aside from the final short form. Additionally, the item “It was hard for my child to pay attention when he/she had pain” exhibited evidence of DIF by sex; the item was not set aside because it had only 6 and 3 parents of male children endorse the two most extreme response categories, as compared to 20 and 15 parents of female children; hypothesis testing using such sparse cell counts is untrustworthy. The resulting 8-item short form produced scores with information greater than 10 from about one standard deviation below the mean to about two and half standard deviations above the mean (see Fig. 2).
Peer Relationships: LD and DIF
There were two potentially locally dependent pairs of items in the Peer Relationships domain (Table 8). From the item pair “My child was able to count on his/her friends” and “My child was able to talk about everything with his/her friends”, the latter item was set aside. Neither item was set aside from the other LD pair after content review suggested a unique contribution from each item. There were no items with significant DIF. The remaining 7 items on the short form produce scores with reliability greater than 0.90 between the mean and about three standard deviations below the mean (see Fig. 2).
Asthma Impact: LD and DIF
There was no evidence of LD or DIF for the Asthma Impact domain (Table 9). The 8-item short form produced scores with information values greater than 10 between about one standard deviation below the mean to nearly three standard deviations above the mean (see Fig. 2).
Parent/child correlations
The correlations between the latent variables measured by the pediatric self-report scales and the parent proxy-report scales are in Table 2.
Summed scores
For the convenience of users who prefer to use summed scores, scoring tables that translate summed scores on the parent proxy-report short forms into IRT scaled scores are given in Tables 10, ,1111 [30]. Computation of these scales used the means and standard deviations in Table 2. Because those means and standard deviations refer back to the original calibration samples for the PROMIS® pediatric self-report scales, the effect is that it is as though these parent proxy-report scales had also been calibrated, or “normed”, on that same sample. Because the correlations (in Table 2) between the pediatric self-report scales and the parent proxy-report scales with the same names range from moderate to low, this does not mean that parent proxyreport scores are comparable with pediatric self-report scores. However, it does mean that the average “level” of the scores (on the PROMIS® T-score scales) has the same meaning with respect to “average” and “one standard deviation above average,” and so on.
Table 10
Table 10
Scoring tables for the depressive symptoms, anxiety, anger, lack of energy, and tired parent proxy scales
Table 11
Table 11
Scoring tables for the upper extremity/dexterity, mobility, pain interference, peer relations, and asthma impact parent proxy scales
This study describes the development and calibrations of the NIH PROMIS® Parent Proxy Report Scales based on an iterative series of IRT analyses regarding scale dimensionality, item LD, and DIF. After determining scale dimensionality, items with LD and DIF were next identified, and some were removed from the recommended short forms.
The potential advantages of utilizing IRT analysis in scale development include greater flexibility in selecting items from the existing parent proxy-report item banks tailored to the objectives of a particular clinical research investigation. Further, scales that have been developed with CTT may have gaps in their ability to measure the full spectrum of the latent construct, while with IRT calibrated items, one can construct a measure that is useful across the full continuum of the latent variable [19]. Thus, this analytic methodology provides clinical researchers the opportunity to select the most meaningful items for their study design and hypotheses. In this study, we proposed short forms measuring each of the content domains; however, a smaller subset of items from the item banks can also be used and scored on the same metric as the larger set using a more dynamic CAT algorithm.
Our finding that parent proxy-report demonstrated moderate to low agreement with pediatric self-report is consistent with the extant literature [12], suggesting that information provided by proxy-respondents is not equivalent to that reported by the patient. In the HRQOL literature, imperfect agreement between self-report and proxyreport has been consistently documented, typically demonstrating higher correlations for more observable domains (i.e., physical functioning) and lower correlations for less observable or internal symptoms such as emotional functioning, pain, and fatigue [31]. Our findings are consistent with this larger literature.
By administering the items spread over several test forms, we were unable to perform factor analyses across the entire item bank for six of the ten content domains. For each of these six domains, factor analysis was conducted on two separate sets of items. It is possible that factor analyses would turn out differently if all the items within each content domain were analyzed as a single set. However, because the items were created to fill content from qualitative work and then were randomly allocated to each test form, the different test forms can be viewed as replications. By having replicated factor analyses, our impressions of multidimensionality, when repeated across forms, increased our confidence in the factor analytic results.
We recruited participants from clinics across five sites to achieve a sample with diverse experiences in terms of health outcomes, but also cultural and ethnic influences. This study does not report on using the items in languages other than English or in children living in other countries, as such, we cannot assume that the scales would have the same test characteristics in those other populations.
Future research with other samples may reveal other sources of DIF for the items; an advantage of IRT as a method is that it can detect item-level DIF, and “flag” items to be used only with caution for comparisons across levels of a variable for which DIF exists. Although analysis of DIF led to smaller item banks, we believe this approach will ultimately yield a more broadly applicable measure for comparing results across populations.
In conclusion, this study provides initial IRT calibrations of the PROMIS® parent proxy-report item banks and the creation of the PROMIS® Parent Proxy Report Scales which address an important gap in the current literature. Further research is indicated on construct validity and tests of the responsiveness of these scales and item banks in larger samples of parents of pediatric patients with chronic health conditions.
Acknowledgments
This work was funded by the National Institutes of Health through the NIH Roadmap for Medical Research, Grant U01AR052181. Information on the Patient-Reported Outcomes Measurement Information System (PROMIS®) can be found at http://nihroadmap.nih.gov/ and http://www.nihPROMIS.org.
Abbreviations
PROMIS®Patient Reported Outcomes Measurement Information System
FDAFood and drug administration
HRQOLHealth-related quality of life
NIHNational Institute of Health

Appendix: item parameters and short form scoring tables
Contributor Information
James W. Varni, Department of Pediatrics, College of Medicine, Department of Landscape Architecture and Urban Planning, College of Architecture, Texas A&M University, 3137 TAMU, College Station, TX 77843-3137, USA.
David Thissen, Department of Psychology, University of North Carolina, at Chapel Hill, Chapel Hill, NC, USA.
Brian D. Stucky, RAND Corporation, Santa Monica, CA, USA.
Yang Liu, Department of Psychology, University of North Carolina, at Chapel Hill, Chapel Hill, NC, USA.
Hally Gorder, Department of Psychology, University of North Carolina, at Chapel Hill, Chapel Hill, NC, USA.
Debra E. Irwin, Department of Epidemiology, University of North Carolina, at Chapel Hill, Chapel Hill, NC, USA.
Esi Morgan DeWitt, Department of Pediatrics, Division of Rheumatology, Cincinnati Children’s Hospital Medical Center, Cincinnati, OH, USA.
Jin-Shei Lai, Department of Medical Social Sciences, Northwestern, University Feinberg School of Medicine, Chicago, IL, USA.
Dagmar Amtmann, Department of Rehabilitation Medicine, University of Washington, Seattle, WA, USA.
Darren A. DeWalt, Division of General Medicine and Clinical Epidemiology, Cecil G. Sheps Center for Health Services Research, University, of North Carolina at Chapel Hill, Chapel Hill, NC, USA.
1. Ader DN. Developing the patient-reported outcomes measurement information system (PROMIS) Medical Care. 2007;45(Suppl 1):S1–S2. [PubMed]
2. Reeve BB, Hays RD, Bjorner JB, Cook KF, Crane PK, Teresi JA, et al. Psychometric evaluation and calibration of health-related quality of life item banks: Plans for the patient-report outcomes measurement information system (PROMIS) Medical Care. 2007;45(Suppl 1):S22–S31. [PubMed]
3. Cella D, Yount S, Rothrock N, Gershon R, Cook K, Reeve B, et al. The patient-reported outcomes measurement information system (PROMIS): Progress of an NIH roadmap cooperative group during its first two years. Medical Care. 2007;45(Suppl 1):S3–S11. [PMC free article] [PubMed]
4. Irwin DE, Stucky, B D, , Thissen D, DeWitt EM, Lai JS, Yeatts K, et al. Sampling plan and patient characteristics of the PROMIS pediatrics large-scale survey. Quality of Life Research. 2010;19:585–594. [PMC free article] [PubMed]
5. Irwin DE, Stucky BD, Langer MM, Thissen D, DeWitt EM, Lai JS, et al. An item response analysis of the pediatric PROMIS anxiety and depressive symptoms scales. Quality of Life Research. 2010;19:595–607. [PMC free article] [PubMed]
6. Varni JW, Stucky BD, Thissen D, DeWitt EM, Irwin DE, Lai JS, et al. PROMIS pediatric pain interference scale: An item response theory analysis of the pediatric pain item bank. Journal of Pain. 2010;11:1109–1119. [PMC free article] [PubMed]
7. DeWitt EM, Stucky BD, Thissen D, Irwin DE, Langer M, Varni JW, Lai JS, Yeatts KB, DeWalt DA. Construction of the eight-item patient-reported outcomes measurement information system pediatric physical function scales: built using item response theory. Journal of Clinical Epidemiology. 2011;64(7):794–804. [PMC free article] [PubMed]
8. Yeatts K, Stucky BD, Thissen D, Irwin DE, Varni JW, DeWitt EM, et al. Construction of the pediatric asthma impact scale (PAIS) for the patient-reported outcomes measurement information system (PROMIS) Journal of Asthma. 2010;47:295–302. [PMC free article] [PubMed]
9. Sprangers MAG, Aaronson NK. The role of health care providers and significant others in evaluating the quality of life of patients with chronic disease: A review. Journal of Clinical Epidemiology. 1992;45:743–760. [PubMed]
10. Achenbach TM, McConaughy SH, Howell CT. Child/adolescent behavioral and emotional problems: Implications of cross-informant correlations for situational specificity. Psycho-logical Bulletin. 1987;101:213–232. [PubMed]
11. Varni JW, Katz ER, Seid M, Quiggins DJL, FriedmanBender A, Castro CM. The pediatric cancer quality of life inventory (PCQL): I Instrument development, descriptive statistics, and cross-informant variance. Journal of Behavioral Medicine. 1998;21:179–204. [PubMed]
12. Upton P, Lawford J, Eiser C. Parent-child agreement across child health-related quality of life instruments: A review of the literature. Quality of Life Research. 2008;17:895–913. [PubMed]
13. Varni JW, Limbers CA, Burwinkle TM. Parent proxy-report of their children’s health-related quality of life: An analysis of 13, 878 parents’ reliability and validity across age subgroups using the PedsQL™ 4.0 Generic Core Scales. Health and Quality of Life Outcomes. 2007;5(2):1–10. [PMC free article] [PubMed]
14. Campo JV, Comer DM, Jansen-McWilliams L, Gardner W, Kelleher KJ. Recurrent pain, emotional distress, and health service use in childhood. Journal of Pediatrics. 2002;141:76–83. [PubMed]
15. Janicke DM, Finney JW, Riley AW. Children’s health care use: A prospective investigation of factors related to care-seeking. Medical Care. 2001;39:990–1001. [PubMed]
16. Varni JW, Setoguchi Y. Screening for behavioral and emotional problems in children and adolescents with congenital or acquired limb deficiencies. American Journal of Diseases of Children. 1992;146:103–107. [PubMed]
17. Varni JW, Burwinkle TM, Lane MM. Healthrelated quality of life measurement in pediatric clinical practice: An appraisal and precept for future research and application. Health and Quality of Life Outcomes. 2005;3(34):1–9. [PMC free article] [PubMed]
18. Reise SP, Waller NG. Item response theory and clinical measurement. Annual Review of Clinical Psychology. 2009;5:27–48. [PubMed]
19. Embretson SE, Reise SP. Item response theory for psychologists. Erlbaum; Mahwah, NJ: 2000.
20. Irwin DE, Gross HE, Stucky BD, Thissen D, DeWitt EM, Lai JS, Amtmann D, Khastou L, Varni JW, DeWalt DA. Development of the PROMIS® pediatrics proxy-report item banks. 2011. Manuscript under review. [PMC free article] [PubMed]
21. Muthén LK, Muthén BO. Mplus user’s guide [Computer Software] 5th ed. Muthén & Muthén; Los Angeles, CA: 2007.
22. Reise SP, Moore TM, Haviland MG. Bifactor models and rotations: Exploring the extent to which multidimensional data yield univocal scale scores. Journal of Personality Assessment. 2010;92:544–559. [PMC free article] [PubMed]
23. Cai L, du Toit SHC, Thissen D. IRTPRO: Flexible, multidimensional, multiple categorical IRT modeling [Computer software] Scientific Software International; Chicago, IL: inpress.
24. Chen WH, Thissen D. Local dependence indexes for item pairs using item response theory. Journal of Educational and Behavioral Statistics. 1997;22:265–289.
25. Lord FM. A study of item bias using item characteristic curve theory. In: Portinga YH, editor. Basic problems in crosscultural psychology. Swets and Zeitlinge; Amsterdam: 1977. pp. 19–29.
26. Cai L. SEM of another flavour: Two new applications of the supplemented EM algorithm. British Journal of Mathematical and Statistical Psychology. 2008;61:309–329. [PubMed]
27. Benjamini Y, Hochberg Y. Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society. 1995;57:289–300.
28. Steinberg L, Thissen D. Using effect sizes for research reporting: Examples using item response theory to analyze differential item functioning. Psychological Methods. 2006;11:402–415. [PubMed]
29. Orlando M, Thissen D. Further investigation of the performance of S-X2: An item fit index for use with dichotomous item response theory models. Applied Psychological Measurement. 2003;27:289–298.
30. Thissen D, Nelson L, Rosa K, McLeod LD. Item response theory for items scored in more than two categories. In: Thissen D, Wainer H, editors. Test scoring. Lawrence Erlbaum Associates; Mahwah, NJ: 2001. pp. 141–186.
31. Eiser C, Morse R. Can parents rate their child’s health-related quality of life? Results from a systematic review. Quality of Life Research. 2001;10:347–357. [PubMed]