Search tips
Search criteria 


Logo of nihpaAbout Author manuscriptsSubmit a manuscriptHHS Public Access; Author Manuscript; Accepted for publication in peer reviewed journal;
JAMA. Author manuscript; available in PMC 2012 March 1.
Published in final edited form as:
PMCID: PMC3290863

Association between Stroke Center Hospitalization for Acute Ischemic Stroke and Mortality

Ying Xian, MD, PhD, Robert G. Holloway, MD, MPH, Paul S. Chan, MD, MSc, Katia Noyes, PhD, MPH, Manish N. Shah, MD, MPH, Henry H. Ting, MD, MBA, Andre R. Chappel, PhDc, Eric D. Peterson, MD, MPH, and Bruce Friedman, PhD, MPH



Although stroke centers are widely accepted and supported, little is known about their impact on patient outcomes.


To examine the association between admission to stroke centers for an acute ischemic stroke and mortality.

Design, Setting, and Participants

Observational study using data from the New York Statewide Planning and Research Cooperative System. We compared mortality for patients admitted with acute ischemic stroke (n=30,947) between 2005 and 2006 at designated stroke centers and non-designated hospitals using differential distance to hospitals as an instrumental variable to adjust for potential pre-hospital selection bias. Patients were followed for mortality for 1 year after the index hospitalization through 2007. To assess whether our findings were specific to stroke, we also compared mortality for patients admitted with gastrointestinal hemorrhage (n=39,409) or acute myocardial infarction (n=40,024) at designated stroke centers and non-designated hospitals.

Main Outcome Measure

Thirty-day all-cause mortality.


Among 30,947 patients with acute ischemic stroke, 15,297 (49.4%) were admitted to designated stroke centers. Using the instrumental variable analysis, admission to designated stroke centers was associated with greater use of thrombolytic therapy (4.8% vs. 1.7%; adjusted difference 2.2%, 95% CI, 1.6% to 2.8%; P<0.001) and lower 30-day all-cause mortality (10.1% vs. 12.5%; adjusted mortality difference: −2.5%, 95% CI, −3.6% to −1.4%; P<0.001). Differences in mortality also were observed at all time points, including at 1-day, 7-day, and 1-year follow-up. Moreover, the outcome differences were specific to stroke, as stroke centers and non-stroke centers had similar 30-day all-cause mortality rates among those with acute myocardial infarction (adjusted mortality difference: +0.3%, 95% CI, −0.5% to 1.0%; P=0.50) and/or gastrointestinal hemorrhage (adjusted mortality difference: +0.1%, 95% CI, −0.9% to 1.1%; P=0.83).


Admission to a designated stroke center for acute ischemic stroke was associated with more frequent use of thrombolytic therapy and lower mortality.

Keywords: quality improvement, outcome research, acute ischemic stroke

Stroke is the leading cause of serious long-term disability and the third leading cause of mortality in the United States.1 Responding to the need for improvements in acute stroke care, the Brain Attack Coalition (BAC) published recommendations for the establishment of Primary Stroke Centers in 2000.2 In December 2003, the Joint Commission began certifying stroke centers based on BAC criteria.3 Now, nearly 700 of the 5,000 acute care hospitals in the U.S. are Joint Commission-certified stroke centers.4 Some states, such as New York, Massachusetts, and Florida have established their own designation programs using the BAC core criteria.

Despite widespread support for the stroke center concept, there is limited empiric evidence demonstrating that admission to a stroke center is associated with lower mortality. Prior studies have largely focused on stroke processes of care, such as treatment timeline and use of thrombolytic therapy.58 There is comparably less information on whether or not better care at stroke centers affects acute or long-term mortality.9 Therefore, our goal was to evaluate the association between admission to stroke centers for acute ischemic stroke and mortality.


Study Population

The primary data source was the New York Statewide Planning and Research Cooperative System (SPARCS), a comprehensive reporting system that collects patient-level data from every hospital admission in New York State. We identified 33,090 hospitalized patients, 18 years of age or older, with a principal diagnosis of acute ischemic stroke between January 1, 2005 and December 31, 2006. An ischemic stroke diagnosis was verified through the International Classification of Diseases, Ninth Revision, Clinical Modification (ICD-9-CM) codes 433.x1, 434.x1 and 436. We limited our study sample to only patients presenting with an initial stroke admission during the study period. We excluded 548 (1.7%) patients who lived outside of New York State and 123 (0.4%) patients with missing data. To avoid a bias against non-designated hospitals, we also excluded 1,472 (4.4%) patients for whom the distance from their home residence to the admitting hospital was greater than 20 miles, since these patients would be less likely to receive thrombolytic therapy. Consistent with the Centers for Medicare and Medicaid Services (CMS), all transfer patients were assigned to the transferring hospital. The final sample included 30,947 patients.

New York State Stroke Center Designation

The New York State Stroke Center Designation program is collaboration between the New York State Department of Health (NYSDOH), the American Heart Association (AHA), and the New York State Quality Improvement Organization.7,10 Beginning in 2004, all New York hospitals were invited to apply to the NYSDOH for stroke center designation if they met the BAC criteria. These criteria are organized around 11 aspects of stroke care, specifically: acute stroke teams, written care protocols, emergency medical services (EMS), emergency department, stroke units, neurosurgical services, commitment and support of the medical organization, neuro-imaging services, laboratory services, outcome and quality improvement activities, and continuing medical education.2 Hospitals were evaluated for stroke center designation with an initial hospital survey, followed by an on-site review and inspection, to ensure hospital compliance with the BAC criteria and preparedness to operate as a stroke center.

Of 244 New York hospitals, 104 (42.6%) became state-designated stroke centers by the end of 2006 (seeeFigure). Considering some hospitals became stroke centers during the study period, we assigned stroke center status for each patient based on the hospital’s designation at the time of admission.

Patient Follow-up and Outcome Measures

Evaluation of in-hospital mortality may be confounded by different lengths of stay between stroke centers and non-designated hospitals. Moreover, the CMS is considering including 30-day ischemic stroke mortality as one of its publicly-reported measures of hospital quality of care.11 Therefore, as our primary outcome, we examined 30-day all-cause mortality among those who were and were not admitted to a stroke center. As secondary outcomes, we evaluated 1- and 7-day, and 1-year all-cause mortality for a sensitivity analysis. Follow-up ended on the date of death or 1 year after the index hospitalization through 2007, whichever came first. Mortality post-discharge was determined through the Social Security Administration Death Master File. Finally, we explored how the use of thrombolytic therapy (ICD-9-CM procedure code 99.10 and/or diagnosis-related group 559), discharge to skilled nursing facilities, and all-cause readmission within 30 days of the index hospital discharge differed by whether or not a patient was admitted to a designated stroke center. Patients who died during the index hospitalization were excluded from the readmission analyses.

Other Study Variables

The SPARCS reporting system provided data on patient characteristics, including socio-demographics (age, sex, race/ethnicity, and insurance status) and comorbidities (differentiated from complications using a present-on-admission indicator). Comorbidities included prior myocardial infarction, congestive heart failure, atrial fibrillation, peripheral vascular disease, diabetes mellitus with or without complications, renal insufficiency, cancer, metastatic carcinoma, liver disease, chronic obstructive pulmonary disease, dementia, connective tissue disease disorder, and peptic ulcer disease. These comorbidities were used to construct a modified version of the Charlson Comorbidity Index (CCI) tailored for ischemic stroke.12 Hospital characteristics, such as size and academic affiliation, were obtained from the NYSDOH and the American Hospital Association Annual Survey. Finally, we determined whether a patient lived in a rural or urban area by applying the Rural-Urban Commuting Area Codes classification system to the patient’s residential zipcode.13

Instrumental Variable

Because it would be impractical to randomize patients with acute ischemic stroke to designated stroke centers or non-designated hospitals, researchers must rely on observational data to assess the impact of stroke centers on mortality. However, both measured and unmeasured confounding inherent in observational studies may lead to selection bias for treatment. For example, EMS may systematically transport more severely-ill patients to stroke centers. Standard statistical approaches, such as multivariate logistic regression or propensity score analysis, cannot account for unmeasured confounding because they can only adjust for measured covariates.14,15 One solution is to use instrumental variable (IV) analysis (an econometric method) to minimize unmeasured confounding.16,17

The key notion behind instrumental variable analysis is that the instrument is highly correlated with the treatment (stroke center vs. non-designated hospital), but is otherwise unrelated to observed or unobserved prognostic risk factors so that it does not directly or indirectly affect patient outcome except through the treatment itself.16,17 This is similar to a randomized controlled trial where the randomization process assigns patients to treatment groups but the randomization itself is not directly associated with outcomes. In the case of stroke center admission, we used differential distance, which is an instrumental variable that has been used in prior studies of acute myocardial infarction and trauma.1620 Differential distance was calculated as the difference between the straight-line distance from a patient's residence to the nearest stroke center minus the straight-line distance from this patient's residence to the nearest hospital of any type. The differential distance is the additional distance, if any, beyond the nearest hospital to reach a stroke center. The choice of differential distance as an effective instrumental variable is based on two assumptions: (1) it is logical to assume that patient transported by a private vehicle will go to the nearest hospital. Importantly, the New York State Stroke Protocol requires EMS to transport stroke patients to the nearest stoke center if the pre-hospital time is less than 2 hours.21 As one would expect, those who live close to a stroke center are more likely to be transported to the stroke center; and (2) patients cannot predict if and when they will have a stroke and, therefore, they do not choose their residence based on proximity to a given hospital. Thus, distance to each type of hospital is highly predictive of whether the patient was admitted to a stroke center but is not associated with disease characteristics such as stroke severity.

Statistical Analyses

Baseline characteristics were compared between patients admitted to designated and non-designated hospitals using the standardized difference. This method has been used to assess the comparability of study subjects in the literature.22 An absolute standardized difference greater than 10 (approximately equivalent to p<0.05) indicates significant imbalance of a baseline covariate, while a smaller value supports the balance assumption between groups. We then assessed whether admission to a designated stroke center was associated with lower mortality using an instrumental variable analysis estimated by a simultaneous two-equation bivariate probit model.23 The first equation estimated the probability of stroke center admission as a function of differential distance and other covariates. The second equation assessed the association of stroke center admission with mortality, adjusted for other patient and hospital factors. Estimating two equations jointly using a bivariate probit approach provides consistent estimates of the treatment effect.23, 24 The IV-adjusted mortality estimate (technically, the average marginal effect) can be interpreted as the mean predicted difference in the probability of death for stroke patients who received treatment at designated stroke centers because they lived relatively closer to stroke centers versus patients who received treatment at non-designated hospitals because they lived further away.

We examined the robustness of our findings in several ways. First, we sought to determine whether admission to a designated stroke center was associated with lower 1- and 7-day, and 1-year all-cause mortality by repeating the analyses for these time points. Second, because the majority of stroke centers are located in New York City (see eFigure), we performed subgroup analyses for patients living in the New York metropolitan area and for those in upstate New York. Third, whites and minorities often live in different neighborhoods and may have systematically used different hospitals. We stratified the analysis by race and checked whether the effect of stroke centers varied by race/ethnicity group. Finally, to determine if our mortality findings were specific to stroke, we compared mortality among patients admitted at designated and non-designated hospitals for 2 other acute life-threatening conditions—gastrointestinal (GI) hemorrhage and acute myocardial infarction (AMI). Both conditions are quality indicators recommended by the Agency for Healthcare Research and Quality to assess a hospital’s quality of care.25 If mortality was lower for either of these 2 conditions in designated stroke centers, this would suggest that lower stroke mortality would be due to these hospitals’ overall commitment to quality improvement, rather than to these hospitals’ implementation of actions specific to stroke.

All tests were evaluated at a 2-sided significance level of p<0.05. The analyses were performed using SAS 9.2 (SAS Institute, Cary, North Carolina) and STATA 11 (StataCorp, College Station, Texas) software. This study was approved by the University of Rochester’s institutional review board, with waiver of informed consent.


Among 30,947 stroke patients, 15,297 (49.4%) were admitted to designated stroke centers and 15,650 (50.6%) to non-designated hospitals. Table 1 compares baseline characteristics of the study cohort. Patients admitted to stroke centers were more frequently younger, non-Hispanic black, less likely to live in a rural area, and more likely to be admitted at a hospital with more beds and an academic affiliation. Patients admitted to stroke centers were relatively healthier with respect to the prevalence of comorbidities, although none of the differences were statistically significant.

Table 1
Baseline Characteristics of Study Cohort

As mentioned above, instrumental variables carry with them several assumptions. First, one assumes that the instrumental variable is highly correlated with the variable of interest. We found from a logistic regression model that the differential distance was highly predictive of whether a patient was admitted to a stroke center (C-statistic=0.88). Second, one assumes that the instrumental variable does not independently affect patient outcomes so that it is not associated with other potential confounders of the outcome. Although it is not possible to examine the unmeasured stroke severity, we were able to check the balance of observed health status according to the differential distance to a stroke center. Table 2 shows baseline characteristics relative to whether a patient lived closer to a stroke center (differential distance=0 or >0 mile). Although there were small differences in certain measures, age and the prevalence of most comorbidities were far more similar than were the groups in Table 1, as reflected by diminishing standardized differences. Despite the similarity in observed health status, the differential distance groups differed substantially in their probability of being admitted to a stroke center. The high correlation between differential distance and stroke center, as well as the balance in observable health status, provide validation of the key instrumental variable assumptions.

Table 2
Patient Characteristics by Differential Distance to a Designated Stroke Center*

Stroke Mortality and Other Outcomes

Mortality rates among patients admitted to stroke centers and non-designated hospitals are summarized in Table 3. The overall 30-day all-cause mortality rate was 10.1% for patients admitted at stroke centers and 12.5% for patients admitted at non-designated hospitals (unadjusted mortality difference, −2.4%, P<0.001). Using instrumental variable analysis, we found admission to a designated stroke center was associated with a 2.5% absolute reduction in 30-day all-cause mortality (adjusted mortality difference, −2.5%; 95% CI, −3.6% to −1.4%; P<0.001).

Table 3
Mortality at Designated Stroke Centers and Non-Designated Hospitals

Use of thrombolytic therapy was 4.8% (739/15,297) for patients admitted at stroke centers and 1.7% (266/15,650) for patients admitted at non-designated hospitals (P<0.001). Admission to a stroke center was associated with the increased use of thrombolytic therapy (adjusted difference in thrombolysis use of 2.2%, 95% CI, 1.6% to 2.8%, P<0.001). However, further adjustment for use of thrombolysis in our instrumental variable models did not substantially alter the association of stroke center admission with lower 30-day mortality: adjusted mortality difference −2.7% (95% CI −3.8% to −1.6%, P<0.001). Among those surviving to hospital discharge, there was no difference in rates of 30-day all-cause readmission (14.8% vs. 14.2%, adjusted difference 1.1%, 95% CI −0.3% to 2.6%, p=0.12) and discharge to a skilled nursing facility (24.6% vs. 28.5%, adjusted difference −0.5%, 95% CI −2.1% to 1.2%, p=0.56).

Sensitivity Analyses

Our sensitivity analysis examined whether or not the effect of stroke center on mortality varied by race, location, or time points. Lower mortality was observed within the first hospital day and at 7 days. Importantly, the all-cause mortality difference also was sustained at 1-year after the index hospitalization (see Table 3). Subgroup analyses of the New York metropolitan area, upstate New York, and stratified analyses by race/ethnicity, found similar results of lower all-cause mortality at designated stroke centers (see Table 4).

Table 4
Sensitivity Analysis: 30-Day Mortality at Designated Stroke Centers and Non-Designated Hospitals

Specificity Analyses

To examine whether the lower mortality at designated stroke centers was specific to stroke, we examined mortality rates for patients admitted with GI hemorrhage and AMI at stroke centers and non-designated hospitals. Thirty-day all-cause mortality for GI hemorrhage was comparable for patients admitted to stroke centers and non-designated hospitals (5.0% vs. 5.8%; adjusted mortality difference, +0.3%; 95% CI, −0.5% to 1.0%; P=0.50). Similarly, 30-day all-cause mortality for AMI did not significantly differ between the two groups (10.5% vs. 12.7%; adjusted mortality difference, +0.1%; 95% CI, −0.9% to 1.1%; P=0.83). For these 2 conditions, there also were no differences in 1-day or 7-day mortality (Table 5). Lastly, given our sample size and observed mortality rates, a retrospective power analysis indicated that our study had more than 90% statistical power to detect a 0.1% mortality difference for AMI and 70% power to detect a 0.3% mortality difference for GI hemorrhage.

Table 5
Specificity Analysis: Mortality for Gastrointestinal Hemorrhage and Acute Myocardial Infarction at Designated Stroke Centers and non-Designated Hospitals


Reduced mortality and increased use of acute stroke therapies are two expected benefits of primary stroke centers.2 Nevertheless, limited empiric evidence supports the benefits of stroke centers—in particular, outcome-based quality measures.9 In this large observational study, we found that patients admitted to stroke centers were more likely to receive thrombolytic therapy and less likely to die when compared to patients admitted to non-designated hospitals. This survival benefit was sustained for up to one year after stroke occurrence and was independent of patient and hospital characteristics. Importantly, the lower mortality at designated stroke centers was specific to stroke and was not found for other acute life-threatening conditions, which suggests that the mortality benefit was related to stroke center designation, rather than to overall quality improvement efforts at designated stroke centers. Collectively, our study provides evidence that the implementation and establishment of a BAC-recommended stroke system of care was associated with improved outcomes for patients with acute ischemic stroke.

Previous evaluations of stroke center quality performance have primarily focused on process measures with limited information on patient outcomes.58 To date, only one study in Finland has reported lower 1-year stroke case-fatality associated with stroke centers.26 Our study extends the findings from this prior study, as systems of stroke care in the U.S. may differ substantially from other national healthcare systems (especially those with universal health coverage). Significantly, we were able to report both short-term and 1-year mortality outcomes. Finally, we were able to demonstrate that lower mortality was specific to stroke at designated stroke centers.

Importantly, geographic patterns of stroke triage are likely to be non-random. Designated stroke centers and non-designated hospitals may treat different groups of patients in terms of demographics and disease severity. For instance, it is possible that EMS may systematically transport more severely-ill patients to stroke centers,4 which is consistent with our finding of a greater IV-adjusted mortality difference (in absolute value) compared to the unadjusted difference (e.g. 2.5% vs. 2.4% for 30-day all-cause mortality). Moreover, prior studies have reported that stroke centers are more likely to admit patients with hemorrhagic strokes, which are associated with higher mortality as compared to strokes with an ischemic etiology.7,8 Indeed, we found a similar pattern, in which nearly 60% (4,193/7,243) of hemorrhagic stroke patients in New York were admitted to a stroke center during our study period. In the absence of randomized controlled trials, controlling for treatment patterns is often difficult and assessments of mortality outcomes may be biased given the presence of treatment selection. Our analysis sought to address these concerns by using an instrumental variable analysis to control for the selection bias (both measured and unmeasured) inherent in observational studies. After adjusting for patient and hospital characteristics and the potential for unmeasured selection bias with the instrumental variable analysis, we found that admission to a stroke center for an acute ischemic stroke was associated with a 2.5% absolute reduction in 30-day all-cause mortality.

The BAC recommendations serve as the cornerstone for the establishment of primary stroke centers. Previous studies have shown reduced mortality among patients who were treated by neurologists or who received organized care in a stroke care unit.2729 Although we cannot determine which individual components of the BAC criteria for stroke center designation were most important for the lower mortality observed in this study, it is likely that the BAC criteria cannot be examined as individually isolated units. Rather, the 11 core criteria, combined, establish the infrastructure and define a paradigm for optimizing care for acute ischemic stroke. By emphasizing an integrated and organized system of care with EMS, hospital emergency departments, acute stroke teams, stroke units, and neuro-imaging services, the BAC criteria facilitate rapid transportation, evaluation, and treatment. Moreover, availability of stroke protocols standardizes acute stroke care and minimizes protocol violation. These efforts are further enhanced by the BAC criteria’s emphasis on surveillance of outcomes, quality initiatives, and continuing educational programs. Of equal importance are the improved performance measures which may also impact downstream care and outcomes. Studies have found association between process of care performance measures and mortality outcomes for patients with cardiovascular disease and stroke.30, 31 It is certainly possible that improved guideline-based treatment, more frequent thrombolytic therapy, enhanced secondary prevention and risk factor management, early rehabilitation and patient education programs may also be responsible for the lower mortality rates among patients treated at stroke centers. However, these efforts may not have any appreciable short-term or immediate life-saving effect, which is consistent with our findings of minimal mortality difference at day 1 and similar readmission rates at 30 days compared to greater survival benefit at the end of 1 year follow-up. Collectively, it is likely that the combinations of these efforts elevate the structure and process of stroke care and subsequently lead to improved patient outcomes.

Since stroke center certification is voluntary, it is possible that hospitals were already committed to quality improvement and would have achieved these results regardless of designation. A recent evaluation of the Joint Commission Certified Primary Stroke Centers found that certified hospitals had better outcomes than non-certified hospitals even before the certification program began.32 Based on our data, we cannot definitively establish if the designation program resulted in reduced mortality or if higher quality hospitals participated in designation. However, this concern is largely mitigated by our specificity analysis, in which we examined mortality rates at designated and non-designated hospitals for 2 other life-threatening conditions—GI hemorrhage and AMI. Hospitals committed to quality improvement prior to stroke center designation would be expected to demonstrate lower mortality for other medical conditions, as well as for stroke. Nevertheless, the lower mortality observed in this study was specific to stroke, thus suggesting that the lower stroke mortality could not be explained simply by the fact that hospitals who received stroke designation were more likely to implement hospital-wide quality improvement.

Our study should be interpreted in the context of the following limitations. First, the SPARCS database did not include information on stroke severity. The differences in mortality may be due to patient case-mix, as opposed to variation in the quality of acute stroke care. This being said, selection bias is more likely to be against admission to a stroke center rather than favoring stroke centers.4 Second, we were unable to assess other performance and outcomes such as eligibility and contradiction of thrombolytic therapy, thrombolysis related hemorrhage, quality of life, neurological and functional status at discharge, since these measures were not collected in the SPARCS database, nor were we able to assess cause-specific mortality. Nonetheless, our study was able to report on the relationship between stroke center admission and all-cause mortality—an outcome which has not been routinely reported. Third, while our sensitivity and specificity analyses suggest that lower mortality associated with stroke center admission may be due to the implementation of the BAC criteria as part of stroke center designation, other quality improvement initiatives (e.g., the AHA Get With The Guidelines-Stroke program), economic incentives from pay for performance, and public reporting could also impact stroke care and outcomes. Fourth, our study only included data from New York. The generalizability of our findings to other states and agencies certifying stroke centers remains to be established. Fifth, we were unable to assess acute treatments other than thrombolytic therapy, including the use of life-sustaining interventions and end-of-life care which may affect short-term or intermediate survival.33 Sixth, many hospitals were transitioning to stroke center during our study period. Defining a stroke center based on designation status on the admission date may have underestimated the mortality differences with stroke center admission. Nonetheless, our conservative approach still demonstrates a lower risk of death associated with stroke centers. Finally, the instrumental variable approach assumes that differential distance has no independent effect on patient outcomes except through its impact on the likelihood of receiving treatment at the designated stroke center. The assumption by its very nature is unproven. However, this assumption would generally be satisfied if a patient's residence is not associated with stroke severity, which appears reasonable. Moreover, differential distance has been widely and successfully used as an instrument to control for selection bias in a variety of clinical settings.1620

In conclusion, we found that admission to designated stroke centers in New York State was associated with a lower risk of death for patients with an acute ischemic stroke. The lower mortality in designated stroke centers was specific to stroke. Our findings suggest that a rigorous process of designating stroke centers has the potential to improve the quality of stroke care and reduce stroke mortality.

Supplementary Material



We acknowledge the insightful contributions of Laine Thomas, PhD, Wenqin Pan, PhD and Margueritte Cox, MS (Duke Clinical Research Institute, Durham, NC) for assistance with statistical analyses and geo-mapping. We thank Zainab Magdon-Ismail, EdM, MPH (American Heart Association, Founders Affiliate), Toby I. Gropen, MD (Long Island College Hospital, New York), Nancy R. Barhydt, DrPH, RN and Anna Colello, Esq (New York State Department of Health) for assistance with data. Those who were acknowledged were not compensated for their contributions.

Financial Disclosures: Dr Holloway: consultant for Milliman Guidelines, Inc. reviewing neurology guidelines. Dr Peterson has received research grants from Schering Plough (modest), Bristol Myers Squibb (modest), Merck/Schering Plough (modest), Sanofi Aventis (modest), Saint Judes, Inc (modest) and is a consultant/advisory board member to Pfizer (modest) and Bayer Corp (significant). Dr Shah is supported by the Paul B. Beeson Career Development Award (NIA 1K23AG028942). Dr Chan is supported by a Career Development Grant Award (K23HL102224) from the NHLBI.

Funding/Support: This study was funded in part by a predoctoral fellowship 0815772D from the American Heart Association (AHA) Founders Affiliate. This study received infrastructure support from the Agency for Healthcare Research and Quality (AHRQ), grant number U18HS016964.



Dr Xian had full access to all of the data in the study and takes responsibility for the integrity of the data and the accuracy of the data analysis.

Study concept and design: Xian, Friedman, Holloway.

Acquisition of data: Xian, Friedman.

Analysis and interpretation of data: Xian, Friedman, Holloway, Noyes, Shah, Peterson, Chan, Ting, Chappel.

Drafting of the manuscript: Xian, Chan.

Critical revision of the manuscript for important intellectual content: Xian, Chan, Peterson, Holloway, Friedman, Noyes, Shah, Ting, Chappel.

Statistical analysis: Xian.

Obtained funding: Xian, Friedman, Holloway.

Administrative, technical, or material support: Xian, Friedman, Holloway.

Study supervision: Xian, Friedman, Holloway, Noyes, Shah.

Previous Presentation: This study was presented in part at the 2010 American Heart Association Quality of Care and Outcomes Research Conference, Washington, DC, May 20, 2010

Role of the Sponsor: The funding organization had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; or preparation of the manuscript. The content is solely the responsibility of the authors and does not necessarily represent the official views of the AHA and AHRQ.

Disclaimer: This study used a linked SPARCS-SSADMF database. The interpretation and reporting of these data are the sole responsibility of the authors.


1. Lloyd-Jones D, Adams RJ, Brown TM, et al. Heart disease and stroke statistics–2010 update: a report from the American Heart Association. Circulation. 2010;121(7):e46–e215. [PubMed]
2. Alberts MJ, Hademenos G, Latchaw RE, et al. Recommendations for the establishment of primary stroke centers. JAMA. 2000;283(23):3102–3109. [PubMed]
3. The Joint Commission Primary Stroke Center Certification. [Accessed Mar 07, 2010];
4. Alberts MJ. Stroke centers: proof of concept and the concept of proof. Stroke. 2010;41(6):1100–1101. [PubMed]
5. Lattimore SU, Chalela J, Davis L, et al. Impact of establishing a primary stroke center at a community hospital on the use of thrombolytic therapy: the NINDS Suburban Hospital Stroke Center experience. Stroke. 2003;34(6):55e–57e. [PubMed]
6. Douglas VC, Tong DC, Gillum LA, et al. Do the Brain Attack Coalition's criteria for stroke centers improve care for ischemic stroke? Neurology. 2005;64(3):422–427. [PubMed]
7. Gropen TI, Gagliano PJ, Blake CA, et al. Quality improvement in acute stroke: The New York State Stroke Center Designation Project. Neurology. 2006;67(1):88–93. [PubMed]
8. Stradling D, Yu W, Langdorf ML, et al. Stroke care delivery before vs after JCAHO stroke center certification. Neurology. 2007;68(6):469–470. [PubMed]
9. Fonarow GC, Gregory T, Driskill M, et al. Hospital Certification for Optimizing Cardiovascular Disease and Stroke Quality of Care and Outcomes. Circulation. 2010;122(23):2459–2469. [PubMed]
10. New York State Department of Health Designated Stroke Centers. [Accessed Mar 07, 2010];
11. The Centers for Medicare and Medicaid Services. [Accessed May 21, 2010];Hospital compare.
12. Goldstein LB, Samsa GP, Matchar DB, Horner RD. Charlson Index Comorbidity adjustment for ischemic stroke outcome studies. Stroke. 2004;35(8):1941–1945. [PubMed]
13. United States Department of Agriculture. [Accessed June 01, 2010];Measuring rurality: rural-urban commuting area codes.
14. Rosenbaum PR, Rubin DB. Reducing bias in observational studies using subclassification on the propensity score. J Am Stat Assoc. 1984;79(387):516–524.
15. Stukel TA, Fisher ES, Wennberg DE, Alter DA, Gottlieb DJ, Vermeulen MJ. Analysis of observational studies in the presence of treatment selection bias: effects of invasive cardiac management on AMI survival using propensity score and instrumental variable methods. JAMA. 2007;297(3):278–285. [PMC free article] [PubMed]
16. McClellan M, McNeil BJ, Newhouse JP. Does more intensive treatment of acute myocardial infarction in the elderly reduce mortality? Analysis using instrumental variables. JAMA. 1994;272(11):859–866. [PubMed]
17. Newhouse JP, McClellan M. Econometrics in outcomes research: the use of instrumental variables. Annu Rev Public Health. 1998;19(1):17–34. [PubMed]
18. Beck CA, Penrod J, Gyorkos TW, Shapiro S, Pilote L. Does aggressive care following acute myocardial infarction reduce mortality? Analysis with instrumental variables to compare effectiveness in Canadian and United States patient populations. Health Serv Res. 2003;38(6p1):1423–1440. [PMC free article] [PubMed]
19. McConnell KJ, Newgard CD, Mullins RJ, Arthur M, Hedges JR. Mortality benefit of transfer to Level I versus Level II trauma centers for head-injured patients. Health Serv Res. 2005;40(2):435–458. [PMC free article] [PubMed]
20. Pracht EE, Tepas JJ, 3rd, Celso BG, Langland-Orban B, Flint L. Survival advantage associated with treatment of injury at designated trauma centers: a bivariate probit model with instrumental variables. Med Care Res Rev. 2007;64(1):83–97. [PubMed]
21. New York State EMT/AEMT BLS Stroke Protocols. [Accessed Apr 11, 2010];2008
22. Austin PC. Using the Standardized Difference to Compare the Prevalence of a Binary Variable Between Two Groups in Observational Research. Communications in Statistics - Simulation and Computation. 2009;38(6):1228–1234.
23. Greene WH. Econometric Analysis. 5 ed. Upper Saddle River, NJ: Prentice Hall; 2003.
24. Bhattacharya J, Goldman D, McCaffrey D. Estimating probit models with self-selected treatments. Stat Med. 2006;25(3):389–413. [PubMed]
25. AHRQ quality indicators. Rockville, MD: Agency for Healthcare Research and Quality; 2006. Feb, [Accessed Apr 11, 2010]. Inpatient Quality Indicators Overview.
26. Meretoja A, Roine RO, Kaste M, et al. Effectiveness of primary and comprehensive stroke centers: PERFECT Stroke: a nationwide observational study from Finland. Stroke. 2010;41(6):1102–1107. [PubMed]
27. Stroke Unit Trialist’s Collaboration. Organised inpatient (stroke unit) care for stroke. Cochrane Database Syst Rev. 2007;(4) CD000197. [PubMed]
28. Mitchell JB, Ballard DJ, Whisnant JP, Ammering CJ, Samsa GP, Matchar DB. What role do neurologists play in determining the costs and outcomes of stroke patients? Stroke. 1996;27(11):1937–1943. [PubMed]
29. Goldstein LB, Matchar DB, Hoff-Lindquist J, Samsa GP, Horner RD. VA Stroke Study: neurologist care is associated with increased testing but improved outcomes. Neurology. 2003;61(6):792–796. [PubMed]
30. Peterson ED, Roe MT, Mulgund J, et al. Association between Hospital Process Performance and Outcomes among Patients with Acute Coronary Syndromes. JAMA. 2006 April 26;295(16):1912–1920. 2006. [PubMed]
31. Bravata DM, Wells CK, Lo AC, et al. Processes of Care Associated With Acute Stroke Outcomes. Archives of Internal Medicine. 2010;170(9):804–810. [PubMed]
32. Lichtman JH, Allen NB, Wang Y, Watanabe E, Jones SB, Goldstein LB. Stroke patient outcomes in US hospitals before the start of the Joint Commission Primary Stroke Center certification program. Stroke. 2009;40(11):3574–3579. [PMC free article] [PubMed]
33. Holloway RG, Quill TE. Mortality as a measure of quality: implications for palliative and end-of-life care. JAMA. 2007;298(7):802–804. [PubMed]