As of June 11, 2009, a total of 17,855 probable or confirmed cases of 2009 pandemic influenza A (H1N1) had been reported in the United States. Risk factors for transmission remain largely uncharacterized. We characterize the risk factors and describe the transmission of the virus within households.
Probable and confirmed cases of infection with the 2009 H1N1 virus in the United States were reported to the Centers for Disease Control and Prevention with the use of a standardized case form. We investigated transmission of infection in 216 households — including 216 index patients and their 600 household contacts — in which the index patient was the first case patient and complete information on symptoms and age was available for all household members.
An acute respiratory illness developed in 78 of 600 household contacts (13%). In 156 households (72% of the 216 households), an acute respiratory illness developed in none of the household contacts; in 46 households (21%), illness developed in one contact; and in 14 households (6%), illness developed in more than one contact. The proportion of household contacts in whom acute respiratory illness developed decreased with the size of the household, from 28% in two-member households to 9% in six-member households. Household contacts 18 years of age or younger were twice as susceptible as those 19 to 50 years of age (relative susceptibility, 1.96; Bayesian 95% credible interval, 1.05 to 3.78; P = 0.005), and household contacts older than 50 years of age were less susceptible than those who were 19 to 50 years of age (relative susceptibility, 0.17; 95% credible interval, 0.02 to 0.92; P = 0.03). Infectivity did not vary with age. The mean time between the onset of symptoms in a case patient and the onset of symptoms in the household contacts infected by that patient was 2.6 days (95% credible interval, 2.2 to 3.5).
The transmissibility of the 2009 H1N1 influenza virus in households is lower than that seen in past pandemics. Most transmissions occur soon before or after the onset of symptoms in a case patient.
Growing use of imaging procedures in the United States has raised concerns about exposure to low-dose ionizing radiation in the general population.
We identified 952,420 non-elderly adults in 5 healthcare markets across the United States between July 1, 2005 and December 31, 2007. Utilization data were used to determine cumulative effective doses of radiation from imaging procedures in millisieverts (mSv) and to calculate population-based rates of “moderate” (>3 to 20 mSv per year), “high” (>20 to 50 mSv per year) and “very-high” (>50 mSv per year) doses.
During the study period, 655,613 (68.8%) individuals underwent at least 1 imaging procedure associated with radiation exposure. The mean effective dose from imaging procedures was 2.4 mSv per person per year (std dev, 6.0 mSv); however, a wide distribution was noted with a median effective dose of 0.1 mSv per person per year (interquartile range, 0.0 to 1.7). Overall, the annual rate for moderate effective doses in the study population was 193.8 per 1000 enrollees, while high and very-high doses occurred at annual rates of 18.6 per 1000 enrollees and 1.9 per 1000 enrollees, respectively. In general, effective doses of radiation from imaging procedures increased with advancing age and were higher in women. Computed tomography and nuclear medicine scans accounted for 75.4% of the total effective dose and 81.8% occurred in non-hospitalized settings.
Imaging procedures are an important source of ionizing radiation in the United States and can lead to high radiation doses in patients.
Myeloablative allogeneic hematopoietic stem-cell transplantation is curative in children
with sickle cell disease, but in adults the procedure is unduly toxic. Graft rejection and
graft-versus-host disease (GVHD) are additional barriers to its success. We performed
nonmyeloablative stem-cell transplantation in adults with sickle cell disease.
Ten adults (age range, 16 to 45 years) with severe sickle cell disease underwent
nonmyeloablative transplantation with CD34+ peripheral-blood stem cells, mobilized by
granulocyte colony-stimulating factor (G-CSF), which were obtained from HLA-matched siblings. The
patients received 300 cGy of total-body irradiation plus alemtuzumab before transplantation, and
sirolimus was administered afterward.
All 10 patients were alive at a median follow-up of 30 months after transplantation
(range, 15 to 54). Nine patients had long-term, stable donor lymphohematopoietic engraftment at
levels that sufficed to reverse the sickle cell disease phenotype. Mean (±SE)
donor–recipient chimerism for T cells (CD3+) and myeloid cells
(CD14+15+) was 53.3±8.6% and 83.3±10.3%,
respectively, in the nine patients whose grafts were successful. Hemoglobin values before
transplantation and at the last follow-up assessment were 9.0±0.3 and 12.6±0.5 g per
deciliter, respectively. Serious adverse events included the narcotic-withdrawal syndrome and
sirolimus-associated pneumonitis and arthralgia. Neither acute nor chronic GVHD developed in any
A protocol for nonmyeloablative allogeneic hematopoietic stem-cell transplantation that
includes total-body irradiation and treatment with alemtuzumab and sirolimus can achieve stable,
mixed donor–recipient chimerism and reverse the sickle cell phenotype.
From 2001 through March 2006, Planned Parenthood health centers throughout the United States provided medical abortion (abortion by means of medication) principally by a regimen of oral mifepristone followed 24 to 48 hours later by vaginal misoprostol. In response to concern about serious infections, in early 2006 Planned Parenthood changed the route of misoprostol administration from vaginal to buccal and required either routine provision of antibiotics or universal screening and treatment for chlamydia; in July 2007, Planned Parenthood began requiring routine treatment with antibiotics for all medical abortions.
We performed a retrospective analysis assessing the rates of serious infection after medical abortion during a time when misoprostol was administered vaginally (through March 2006), as compared with rates after a change to buccal administration of misoprostol and after initiation of additional infection-reduction measures.
Rates of serious infection dropped significantly after the joint change to buccal misoprostol from vaginal misoprostol and to either testing for sexually transmitted infection or routine provision of antibiotics as part of the medical abortion regimen. The rate declined 73%, from 0.93 per 1000 abortions to 0.25 per 1000 (absolute reduction, 0.67 per 1000; 95% confidence interval [CI], 0.44 to 0.94; P<0.001). The subsequent change to routine provision of antibiotics led to a further significant reduction in the rate of serious infection — a 76% decline, from 0.25 per 1000 abortions to 0.06 per 1000 (absolute reduction, 0.19 per 1000; 95% CI, 0.02 to 0.34; P = 0.03).
The rate of serious infection after medical abortion declined by 93% after a change from vaginal to buccal administration of misoprostol combined with routine administration of antibiotics.
The long-term renal consequences of kidney donation by a living donor are attracting increased appropriate interest. The overall evidence suggests that living kidney donors have survival similar to that of nondonors and that their risk of end-stage renal disease (ESRD) is not increased. Previous studies have included relatively small numbers of donors and a brief follow-up period.
We ascertained the vital status and lifetime risk of ESRD in 3698 kidney donors who donated kidneys during the period from 1963 through 2007; from 2003 through 2007, we also measured the glomerular filtration rate (GFR) and urinary albumin excretion and assessed the prevalence of hypertension, general health status, and quality of life in 255 donors.
The survival of kidney donors was similar to that of controls who were matched for age, sex, and race or ethnic group. ESRD developed in 11 donors, a rate of 180 cases per million persons per year, as compared with a rate of 268 per million per year in the general population. At a mean (±SD) of 12.2±9.2 years after donation, 85.5% of the subgroup of 255 donors had a GFR of 60 ml per minute per 1.73 m2 of body-surface area or higher, 32.1% had hypertension, and 12.7% had albuminuria. Older age and higher body-mass index, but not a longer time since donation, were associated with both a GFR that was lower than 60 ml per minute per 1.73 m2 and hypertension. A longer time since donation, however, was independently associated with albuminuria. Most donors had quality-of-life scores that were better than population norms, and the prevalence of coexisting conditions was similar to that among controls from the National Health and Nutrition Examination Survey (NHANES) who were matched for age, sex, race or ethnic group, and body-mass index.
Survival and the risk of ESRD in carefully screened kidney donors appear to be similar to those in the general population. Most donors who were studied had a preserved GFR, normal albumin excretion, and an excellent quality of life.
The hemolytic–uremic syndrome consists of the triad of microangiopathic hemolytic anemia, thrombocytopenia, and renal failure. The common form of the syndrome is triggered by infection with Shiga toxin–producing bacteria and has a favorable outcome. The less common form of the syndrome, called atypical hemolytic–uremic syndrome, accounts for about 10% of cases, and patients with this form of the syndrome have a poor prognosis. Approximately half of the patients with atypical hemolytic–uremic syndrome have mutations in genes that regulate the complement system. Genetic factors in the remaining cases are unknown. We studied the role of thrombomodulin, an endothelial glycoprotein with anticoagulant, antiinflammatory, and cytoprotective properties, in atypical hemolytic–uremic syndrome.
We sequenced the entire thrombomodulin gene (THBD) in 152 patients with atypical hemolytic–uremic syndrome and in 380 controls. Using purified proteins and cell-expression systems, we investigated whether thrombomodulin regulates the complement system, and we characterized the mechanisms. We evaluated the effects of thrombomodulin missense mutations associated with atypical hemolytic–uremic syndrome on complement activation by expressing thrombomodulin variants in cultured cells.
Of 152 patients with atypical hemolytic–uremic syndrome, 7 unrelated patients had six different heterozygous missense THBD mutations. In vitro, thrombomodulin binds to C3b and factor H (CFH) and negatively regulates complement by accelerating factor I–mediated inactivation of C3b in the presence of cofactors, CFH or C4b binding protein. By promoting activation of the plasma procarboxypeptidase B, thrombomodulin also accelerates the inactivation of anaphylatoxins C3a and C5a. Cultured cells expressing thrombomodulin variants associated with atypical hemolytic–uremic syndrome had diminished capacity to inactivate C3b and to activate procarboxypeptidase B and were thus less protected from activated complement.
Mutations that impair the function of thrombomodulin occur in about 5% of patients with atypical hemolytic–uremic syndrome.
Artemisinin-based combination therapies are the recommended first-line treatments of falciparum malaria in all countries with endemic disease. There are recent concerns that the efficacy of such therapies has declined on the Thai–Cambodian border, historically a site of emerging antimalarial-drug resistance.
In two open-label, randomized trials, we compared the efficacies of two treatments for uncomplicated falciparum malaria in Pailin, western Cambodia, and Wang Pha, northwestern Thailand: oral artesunate given at a dose of 2 mg per kilogram of body weight per day, for 7 days, and artesunate given at a dose of 4 mg per kilogram per day, for 3 days, followed by mefloquine at two doses totaling 25 mg per kilogram. We assessed in vitro and in vivo Plasmodium falciparum susceptibility, artesunate pharmacokinetics, and molecular markers of resistance.
We studied 40 patients in each of the two locations. The overall median parasite clearance times were 84 hours (interquartile range, 60 to 96) in Pailin and 48 hours (interquartile range, 36 to 66) in Wang Pha (P<0.001). Recrudescence confirmed by means of polymerase-chain-reaction assay occurred in 6 of 20 patients (30%) receiving artesunate monotherapy and 1 of 20 (5%) receiving artesunate–mefloquine therapy in Pailin, as compared with 2 of 20 (10%) and 1 of 20 (5%), respectively, in Wang Pha (P = 0.31). These markedly different parasitologic responses were not explained by differences in age, artesunate or dihydroartemisinin pharmacokinetics, results of isotopic in vitro sensitivity tests, or putative molecular correlates of P. falciparum drug resistance (mutations or amplifications of the gene encoding a multidrug resistance protein [PfMDR1] or mutations in the gene encoding sarco–endoplasmic reticulum calcium ATPase6 [PfSERCA]). Adverse events were mild and did not differ significantly between the two treatment groups.
P. falciparum has reduced in vivo susceptibility to artesunate in western Cambodia as compared with northwestern Thailand. Resistance is characterized by slow parasite clearance in vivo without corresponding reductions on conventional in vitro susceptibility testing. Containment measures are urgently needed. (ClinicalTrials.gov number, NCT00493363, and Current Controlled Trials number, ISRCTN64835265.)
Mansonella perstans infection is common in areas of Africa where Wuchereria bancrofti, a causative agent of lymphatic filariasis, is endemic. M. perstans is refractory to standard antifilarial therapies. The recent discovery of bacterial endosymbionts (e.g., wolbachia) in most filarial species, including M. perstans, provides new therapeutic options for reducing microfilaremia.
In an open-label, randomized trial, we recruited subjects with M. perstans microfilaremia, with or without concomitant W. bancrofti infection, from four villages in Mali and randomly assigned them to receive doxycycline, at a dose of 200 mg daily for 6 weeks (106 subjects), or no treatment (110). At 6 months, subjects who were co-infected with W. bancrofti underwent a second random assignment, to treatment with a single dose of albendazole (400 mg) and ivermectin (150 μg per kilogram of body weight) or no treatment. Subjects were monitored daily during the first 6-week study period for adverse events. M. perstans and W. bancrofti microfilarial levels were assessed at 6, 12, and 36 months.
At 12 months, 67 of 69 subjects who had received treatment with doxycycline only (97%) had no detectable M. perstans microfilariae per 60 μl of blood, as compared with 10 of 63 subjects who had received no treatment (16%) (relative risk, 6.18; 95% confidence interval, 3.63 to 11.89; P<0.001). At 36 months, M. perstans microfilaremia remained suppressed in 48 of 64 subjects who had received treatment with doxycycline only (75%), a finding that was consistent with a macrofilaricidal effect of doxycycline. Vomiting was more frequent in the doxycycline-treated group than in the untreated group (17% vs. 4%).
These results are consistent with previous findings that M. perstans harbors the intracellular endosymbiont, wolbachia, and suggest that doxycycline is an effective therapy for M. perstans infection.
Five children from two consanguineous families presented with epilepsy beginning in infancy and severe ataxia, moderate sensorineural deafness, and a renal salt-losing tubulopathy with normotensive hypokalemic metabolic alkalosis. We investigated the genetic basis of this autosomal recessive disease, which we call the EAST syndrome (the presence of epilepsy, ataxia, sensorineural deafness, and tubulopathy).
Whole-genome linkage analysis was performed in the four affected children in one of the families. Newly identified mutations in a potassium-channel gene were evaluated with the use of a heterologous expression system. Protein expression and function were further investigated in genetically modified mice.
Linkage analysis identified a single significant locus on chromosome 1q23.2 with a lod score of 4.98. This region contained the KCNJ10 gene, which encodes a potassium channel expressed in the brain, inner ear, and kidney. Sequencing of this candidate gene revealed homozygous missense mutations in affected persons in both families. These mutations, when expressed heterologously in xenopus oocytes, caused significant and specific decreases in potassium currents. Mice with Kcnj10 deletions became dehydrated, with definitive evidence of renal salt wasting.
Mutations in KCNJ10 cause a specific disorder, consisting of epilepsy, ataxia, sensorineural deafness, and tubulopathy. Our findings indicate that KCNJ10 plays a major role in renal salt handling and, hence, possibly also in blood-pressure maintenance and its regulation.
Exposure to fine particulate matter air pollution has been associated with increased risk of cardiopulmonary and lung cancer morbidity and mortality, suggesting that sustained reductions in pollution exposure should result in improved life expectancy. This study directly evaluates changes in life expectancy associated with differential changes in fine particulate air pollution that occurred in the U.S. during the 1980s and 1990s.
Life expectancy, socio-economic, and demographic data were compiled for 217 counties in the 51 U.S. metropolitan areas with matching fine particulate air pollution data for the late 1970s/early 1980s and the late 1990s/early 2000s. Regression models were used to estimate the association between reductions in pollution and changes in life expectancy, controlling for changes in socio-economic and demographic variables and for proxy indicators of cigarette smoking.
A decrease of 10 μg/m3 of fine particulate concentration was associated with an estimated increase in life expectancy equal to 0.77 (SE = 0.17) years. The estimated effect of reduced pollution exposure on life expectancy was not highly sensitive to controlling for changes in socio-economic, demographic, and proxy smoking variables or to restricting observations to relatively large counties and central metro counties. The effect of reductions in air pollution on life expectancy in the study areas was as much as 18% of the overall increase.
Reducing exposure to ambient fine particulate matter air pollution contributed to significant and measurable improvements in life expectancy in the U.S.
life expectancy; air pollution; particulate matter
Surgical ventricular reconstruction is a specific procedure designed to reduce left ventricular volume in patients with heart failure caused by coronary artery disease. We conducted a trial to address the question of whether surgical ventricular reconstruction added to coronary-artery bypass grafting (CABG) would decrease the rate of death or hospitalization for cardiac causes, as compared with CABG alone.
Between September 2002 and January 2006, a total of 1000 patients with an ejection fraction of 35% or less, coronary artery disease that was amenable to CABG, and dominant anterior left ventricular dysfunction that was amenable to surgical ventricular reconstruction were randomly assigned to undergo either CABG alone (499 patients) or CABG with surgical ventricular reconstruction (501 patients). The primary outcome was a composite of death from any cause and hospitalization for cardiac causes. The median follow-up was 48 months.
Surgical ventricular reconstruction reduced the end-systolic volume index by 19%, as compared with a reduction of 6% with CABG alone. Cardiac symptoms and exercise tolerance improved from baseline to a similar degree in the two study groups. However, no significant difference was observed in the primary outcome, which occurred in 292 patients (59%) who were assigned to undergo CABG alone and in 289 patients (58%) who were assigned to undergo CABG with surgical ventricular reconstruction (hazard ratio for the combined approach, 0.99; 95% confidence interval, 0.84 to 1.17; P = 0.90).
Adding surgical ventricular reconstruction to CABG reduced the left ventricular volume, as compared with CABG alone. However, this anatomical change was not associated with a greater improvement in symptoms or exercise tolerance or with a reduction in the rate of death or hospitalization for cardiac causes. (ClinicalTrials.gov number, NCT00023595.)
The full complement of DNA mutations that are responsible for the pathogenesis of acute myeloid leukemia (AML) is not yet known.
We used massively parallel DNA sequencing to obtain a very high level of coverage (approximately 98%) of a primary, cytogenetically normal, de novo genome for AML with minimal maturation (AML-M1) and a matched normal skin genome.
We identified 12 acquired (somatic) mutations within the coding sequences of genes and 52 somatic point mutations in conserved or regulatory portions of the genome. All mutations appeared to be heterozygous and present in nearly all cells in the tumor sample. Four of the 64 mutations occurred in at least 1 additional AML sample in 188 samples that were tested. Mutations in NRAS and NPM1 had been identified previously in patients with AML, but two other mutations had not been identified. One of these mutations, in the IDH1 gene, was present in 15 of 187 additional AML genomes tested and was strongly associated with normal cytogenetic status; it was present in 13 of 80 cytogenetically normal samples (16%). The other was a nongenic mutation in a genomic region with regulatory potential and conservation in higher mammals; we detected it in one additional AML tumor. The AML genome that we sequenced contains approximately 750 point mutations, of which only a small fraction are likely to be relevant to pathogenesis.
By comparing the sequences of tumor and skin genomes of a patient with AML-M1, we have identified recurring mutations that may be relevant for pathogenesis.