|Home | About | Journals | Submit | Contact Us | Français|
Amyotrophic lateral sclerosis is a devastating neurodegenerative disease caused by loss of motor neurons. Its pathophysiology remains unknown, but progress has been made in understanding its genetic and biochemical basis. Clinical trialists are working to translate basic science successes into human trials with more efficiency, in the hope of finding successful treatments. In the future, new preclinical models, including patient-derived stem cells may augment transgenic animal models as preclinical tools. Biomarker discovery projects aim to identify markers of disease onset and progression for use in clinical trials. New trial designs are reducing study time, improving efficiency and helping to keep pace with the increasing rate of basic and translational discoveries. Ongoing trials with novel designs are paving the way for amyotrophic lateral sclerosis clinical research.
Amyotrophic lateral sclerosis (ALS) is a devastating neurodegenerative disease caused primarily by motor neuron loss resulting in progressive weakness, disability and eventually death. The incidence of ALS is approximately 2/100,000 . Although this incidence is comparable to that of multiple sclerosis, the prevalence of ALS is much lower at 5/100,000 because of its rapid progression . Average age of onset is 55–65 years, mean survival is 3–5 years and onset before 30 years of age is uncommon .
Amyotrophic lateral sclerosis is diagnosed and monitored based on the clinical history, physical examination, disease course, electrodiagnostic findings and elimination of competing diagnoses. Consensus diagnostic criteria, known as the revised El Escorial Criteria, are typically used for study enrollment . These criteria have been valuable to the field as diagnostic criteria; however, early diagnosis remains challenging. At the time of diagnosis, less than 60% fulfill El Escorial Criteria for definite or probable ALS and the average duration between symptom onset and diagnosis is approximately 1 year [4–6].
The pathophysiology of ALS is not understood. Even whether neurons or glial cells are the primary site of pathogenesis is the subject of debate. Current theories about pathogenesis center primarily on abnormal protein aggregation (e.g., neurofilament heavy chain aggregates, superoxide dismutase 1 (SOD1) aggregates and the inclusion of cystatin C in Bunina bodies), abnormal RNA processing (evidenced by the presence of causative mutations in genes for RNA-binding proteins, such as transactive response DNA-binding protein 43 [TDP-43] and fused in sarcoma/translated in liposarcoma), abnormal mitochondrial function, glutamate excitotoxicity and neuroinflammation . Thus, candidate drugs targeting these mechanisms have received the most attention from trialists.
Until the late 1980s and early 1990s, very litle was known about the etiology of ALS. As a result, few clinical trials for ALS were conducted. However, in the early 1990s a series of events led to an increase in the number of ALS clinical trials. The SOD1 gene encoding Cu/Zn superoxide dismutase became the first known gene causing familial ALS (FALS), riluzole gained US FDA approval for treatment of ALS, the first transgenic ALS animal models were developed and gains were made in supportive care for those with ALS. These breakthroughs led to optimism in the scientific community and fostered the development of regional clinical research networks.
In 1991, linkage analysis localized a FALS gene to the long arm of chromosome 21 , and subsequently identified it to be in the SOD1 gene . Numerous distinct SOD1 mutations have been described – substitution of alanine for valine at position 4 (SOD1A4V) is the most common in the USA and is associated with a rapid disease progression .
Since the discovery of the SOD1 gene mutation, multiple causative mutations have been identified [11–23], and the pace of these discoveries continues to increase due to the introduction of powerful new techniques for gene discovery. However, the vast majority of ALS cases (>90%) are still sporadic, without known genetic contribution.
In 1994, the first transgenic mouse model of ALS was reported . Initial investigations of humans and transgenic mice carrying SOD1 mutations suggested a reduction in SOD1 quantity in the cerebrospinal fluid (CSF) and activity in erythrocytes [25–27]. Subsequent investigations revealed an ALS phenotype in transgenic mice with an elevation in SOD1 concentration  and normal SOD1 enzymatic activity , while SOD1-deficient mice did not develop an ALS phenotype . An effort to treat humans and transgenic sheep expressing the mutant SOD1 gene with recombinant wild-type SOD1 via intrathecal infusion failed to show disease improvement despite marked elevations in SOD1 . This evidence suggests that SOD1 mutations confer a toxic gain-of-function, rather than a loss-of-function, fitting with the predominantly autosomal dominant inheritance pattern of SOD1 FALS. Potential treatments for ALS are often screened in the G93A mouse model prior to human trials.
However, the disconnect between the frequent, yet small, successes in transgenic animal models and the negative human trial results is a topic of much debate. The G93A mouse has been shown to produce mutant human SOD1 mRNA at concentrations many times higher than wild type mouse SOD1 , which has led to hypotheses that overexpression could be an important driver of the difference between responses in mouse models and humans. This overexpression may confer different response to treatment in the mouse models than in humans. Furthermore, the presymptomatic span in rodents is very short, and once symptomatic the transgenic mice rapidly progress to paralysis and death. The early onset and rapidity of the decline may also explain the difference in response between mice and humans. Most patients have sporadic ALS, the pathophysiology of which may differ in critical ways from that of FALS. Even FALS due to different mutations may respond differently to potential therapies. Thus, the wisdom of screening potential therapies in the SOD1 model has come under scientific scrutiny and new screening models are needed.
Riluzole is the only FDA-approved treatment for ALS. In two separate trials, at 12 or 18 months, the risk of tracheostomy or death was significantly decreased [32,33], although there was no effect on strength, vital capacity (VC) or functional scales. Subsequently, riluzole was shown to extend survival in SOD1G93A transgenic mouse models by 10–15 days .
Ciliary neurotrophic factor, brain-derived neurotrophic factor and IGF were each tested in large clinical trials in the mid 1990s . None of these trials were positive, but they led to the validation of a number of clinical end points. The most influential has been the ALS functional rating scale (ALSFRS), developed by the ciliary neurotrophic factor Study Group .
In 2008, it was reported that treatment with lithium reduced spinal motor neuron loss, delayed symptom onset, prolonged survival after disease onset, conferred a 36% overall survival benefit in SOD1G93A transgenic mice and, in a small trial in humans, improved survival, slowed VC loss and preserved strength over 15 months . Many ALS patients were eager to be treated with lithium and clinicians were likely to oblige, creating a challenging environment for further trials of lithium.
Two trials used innovative designs to study lithium for ALS quickly, especially given the off-label availability of lithium. First, a Northeast ALS (NEALS) Consortium trial guaranteed participants open-label treatment at the end of the study and employed a time-to-event design to speed the study . Terminal study event was defined as a ALSFRS-revised (R) decrease of 6-points. Interim analyses with futility stopping rules were preplanned. The trial enrolled quickly and was stopped for futility at the first interim analysis. Thus, an innovative design overcame a specific challenge (recruitment) and provided important data. The time-to-event analysis sped enrollment at the expense of a decreased statistical power and ability to detect only large changes in the ALSFRS-R rate of decline.
Second, the Western ALS (WALS) Study Group performed an open-label, historical placebo, Phase II trial of lithium for ALS . The placebo arm subjects from a prior trial of minocycline  acted as historical controls and were well matched. ALSFRS-R scores declined more rapidly in those receiving lithium. Eliminating the placebo arm sped enrollment and provided power to detect small changes, albeit at the expense of true randomization. The use of control patients from a prior clinical trial holds promise.
Challenges in ALS research include disease rarity and heterogeneity, a lack of pathophysiological understanding and the absence of reliable biomarkers. As discussed, animal models are a frequently used tool, but their use has not translated to success in humans.
Amyotrophic lateral sclerosis is a rare and fatal disease, and even in multicenter trials, patient enrollment can be slow. Multidisciplinary ALS clinics confer survival and quality-of-life benefits [41–44] and may facilitate patient enrollment in trials. In future, telemedicine may allow immobile patients to participate in multidisciplinary clinics for care and research participation. For now, however, recruitment remains a looming obstacle for all clinical trials.
The rapidity of progression varies widely in ALS. Average survival is approximately 3 years , but some patients progress rapidly, succumbing in less than a 1 year, whilst others survive for more than a decade . Survival is longer in patients with a longer delay between symptom onset and diagnosis and in younger patients, and is shorter in patients with bulbar onset of symptoms [46,47]. El Escorial Criteria stage at the time of diagnosis is inconsistently correlated with rate of disease progression [46,48]. No specific markers of disease onset or progression have been reproducibly identified.
Studies of homogeneous subpopulations within a disease can increase statistical power and decrease trial duration. Unfortunately, clinically evident subgroups are either not predictive of disease course or are not obvious until late in the disease. Causative mutations can confer reliable phenotypes; for example, rapid progression is a feature of the SOD1 A4V mutation and lower motor neuron predominance is a hallmark of fused in sarcoma (FUS) mutations. However, these mutations are rare, and different mutations in a single gene may have dissimilar clinical features . Thus, genotyping has not been used to characterize subgroups for study, although some trialists have begun to collect and store DNA for future sequencing and subgroup analysis.
The NIH defines a biomarker as ‘a characteristic that is objectively measured and evaluated as an indicator of normal biologic processes, pathogenic processes or pharmacologic responses to a therapeutic intervention’ . There are ongoing biomarker discovery efforts , and such markers could hasten study entry or serve as surrogate outcome measures in clinical trials. In the absence of such markers, trialists have depended upon more traditional outcome measures discussed below, leading to the need for large, prolonged and expensive efficacy trials.
Effective ALS clinical trial design depends upon the appropriate choice of outcome measure (Table 1). Over a decade of work by ALS clinical trialists has defined a useful set of outcome measures that are the current mainstays of ALS clinical trial design. At present, the FDA requires that the primary outcome of a registration trial for a medication, one used to support the approval of a medication for a given indication, must provide evidence of clinically meaningful change, such as survival or functional measures. The European Agency for the Evaluation of Medicinal Products provides more specific guidance on the use of end points in clinical trials, suggesting the use of two of the following three items as primary end points: survival, disability (including ALSFRS-R) and strength .
One of the most universally accepted outcome measures is tracheostomy-free survival. Since the early/mid-1990s, common surrogate measures have included VC , the ALSFRS-R [54–57] and quantitative measures of strength testing such as hand-held dynamometry . Secondary outcome measures such as the Modified Ashworth Spasticity Scale and numerous measures of cognitive ability and quality-of-life questionnaires are sometimes used to supplement primary outcome measures.
Each of these outcome measures has strengths, but none is perfect . Tracheostomy-free survival is undeniably clinically relevant. At the same time, it increases trial duration and cost, and provides no information about quality of life or disability. It can be confounded by the nonuniform application of life-extending interventions, such as noninvasive ventilation, feeding tube, riluzole therapy, and participation in a multidisciplinary ALS clinic. Successful randomization and blinding are necessary to overcome these potential confounders.
VC declines with ALS progression, is clinically relevant, related to survival and can be easily followed in the clinic. Many trials have used a pre-determined decline in forced VC as the primary outcome measure. However, VC decline can be a late feature in patients with limb-onset disease and is known to be affected early in patients with bulbar-onset ALS . Furthermore, patients with prominent bulbar symptoms often have difficulty performing these tests due to facial weakness, leading to inaccuracies in measurement.
ALSFRS-R has become one of the most widely accepted measures of ALS progression. It captures many clinically relevant features of disease progression, focuses on disability caused by ALS, has been shown to be reproducible and is easily administered. However, despite all its potential benefits, it remains a subjective score. In addition, statistical handling of mortality in conjunction with ALSFRS-R has proven stubbornly complicated.
Hand-held dynamometry is a quantitative means to assess reduction in muscle strength, the primary effect of ALS. It is objectively measured, easily transported and reproducible. However, the test is subject to error introduced by examiner technique; thus, extensive outcome training is required to ensure intra- and inter-rater reliability, particularly when it is used as an outcome measure in multicenter studies. Furthermore, like other quantitative measures of strength, its results can vary based on patient effort.
In 1979, Lasagna described numerous challenges in general clinical trial design . These included a failure to publish negative results, recruitment and retention issues, statistical handling of subjects lost-to-follow-up, the use of inherently flawed outcome measures, a poor understanding of these limitations and the relative lack of clinical relevance of trial outcomes. Over 30 years later, these challenges in clinical research remain. However, in the field of ALS, new tools now allow clinical researchers to successfully plot a course through these once unnavigable waters. ALS clinical research, once typified by setbacks and obstacles, is now characterized by increasingly rapid basic and translational scientific discoveries, flourishing biomarker discovery projects, growing biorepositories, innovative clinical trial designs and an international and collaborative research community. These developments have led to a recent resurgence in the number of ALS clinical trials (Figure 1).
Since the discovery of the gene, causative mutations in a host of other genes have been reported, including TARDBP , FUS/TLS [17,22], ALS2 , DCTN1 , ANG, VAPB , SPG11 , OPTN , FIG4 , CHMP2B, SETX , VCP , UBQLN22  and C9ORF72 [60,61] (Table 2). New causative mutations are being described with increasing frequency owing to powerful new techniques such as whole exome capture. However, limitations in our knowledge of the genetics of ALS are still evident. The list of known causative genes is not exhaustive and the pathophysiology linking any given mutation to the ALS phenotype remains uncertain.
If the field of ALS genetics is young, then epigenetic research in ALS is in its infancy. In 2010, intermediate-length polyglutamine repeat expansion in the ataxin-2 gene was shown to confer increased risk of ALS in patients with TARDMP mutations, thus making it the first disease-modifying gene characterized in ALS . The search for genetic mutations that confer disease risk or susceptibility to an environmental toxin has also begun. Numerous studies of the paroxonase genes (lipid oxidation and organophosphate detoxification) have suggested a role of polymorphisms in ALS susceptibility [63–66], although others have failed to demonstrate this association [67–69], and still others have concluded that some polymorphisms in these genes represent causative ALS mutations .
Investigation into gene expression and the role of miRNA processing is similarly nascent, but is gaining interest. Both TDP-43- and FUS-encoded proteins associate with the Drosha complex, which is required to process primary miRNA transcripts . The Dicer1 enzyme provides further miRNA processing, and Dicer1-deficient mice have been shown to develop spinal motor neuron degeneration, suggesting a role of miRNA in this process . However, since miRNAs often mediate multiple RNA pathways, these results require further investigation. If the culprit miRNAs are identified, then preclinical and clinical research can focus on a means to alter the miRNA machinery.
The number of animal models has grown alongside the number of ALS gene mutations. Current transgenic animals can be broadly categorized into models of protein misfolding, intermediate filament disorganization, microtubule transport disruption and a newer category of RNA processing disruption in the form of TDP-43 mice [73,74]. Thus far, however, no model has been as robust or widely applied for preclinical testing as the SOD1G93A model.
The process for transforming human embryonic stem cells into motor neurons for study is well-established and numerous protocols now exist . In vitro investigations have revealed a susceptibility of embryonic stem cell-derived motor neurons to factors secreted by cocultured glia from transgenic mice expressing SOD1 gene mutations, giving some insight into noncell autonomous factors in the disease .
Induced pluripotent stem (iPS) cells can be created from fibroblasts isolated from human skin biopsies , a technology that is being increasingly applied in the study of ALS. This approach to stem cell procurement eliminates some ethical debate surrounding the use of embryonic stem cells, but also allows for the derivation of stem-cells from ALS patients. iPS cells, like embryonic stem cells, can be differentiated into motor neurons for use as a tool in the laboratory [75,78]. iPS-derived motor neurons from patients with ALS are currently being created. As a preclinical disease model, they could be instrumental in both the understanding of ALS pathophysiology and high-throughput candidate drug screening that might complement, or supplant, mouse model screening.
A great deal of work has gone into developing novel surrogate outcome measures and biomarkers. Useful biomarkers might be found in biofluids (abnormal protein configuration or concentration, altered small molecule concentration or dysregulated RNA), imaging studies (MRI or MR spectroscopy), quantitative strength testing or new electrodiagnostic techniques.
Patients with ALS can show mild global brain atrophy and hyperintensity in the corticospinal tracts on T2-weighted MRI of the brain; however, this finding is neither sensitive nor specific enough to be useful diagnostically . In fact, traditional MRI of the brain and spinal cord is currently only used in the diagnostic workup of ALS to exclude disease mimics. Newer MRI techniques, including diffusion tensor imaging, task-based functional MRI (fMRI), resting state–fMRI and magnetic resonance spectroscopy (MRS) are being increasingly studied in ALS. Previously published clinico-pathological studies had demonstrated the focal onset and spread of the disease in the CNS, thus, these studies did not reveal previously undiscovered dimensions of the disease [80,81]. However, because MRI technology is widely available, non-invasive and can be repeated, MRI-based biomarkers are attractive for use in clinical trials.
Diffusion tensor imaging is, perhaps, the best characterized novel MRI technique, and has the potential to facilitate diagnosis and disease monitoring; however, confirmatory studies are necessary before the technique can be reliably implemented [82,83]. Task-based fMRI has shown increased activation of the contralateral supplemental motor area, basal ganglia and cerebellum, but is not sensitive enough to confer diagnostic benefit . Resting state fMRI use has been explored preliminarily in ALS, and in combination with tractography techniques such as diffusion tensor imaging, may identify changes in brain region connectivity in ALS . Relaxation rate of the corticospinal tracts and deep grey matter have been used to show tract disruption and iron accumulation, suggesting a potential role of iron-induced oxidative damage . Proton MRS has been used to show a reduction in cortical N-acetyl aspartate, a neuronal marker .
Biofluid and tissue biomarker development in ALS has focused primarily on proteomics (the study of the protein complement in a cell or organism) [86–89], metabolomics (the study of the complement of small molecules present in a cell or organism reflecting the metabolic state of the organism) , and gene expression analysis (the study of the type and amount of RNA transcripts in a cell or organism) .
Biomarkers of ALS may be most effectively identified from the CSF because of its proximity to the motor neurons and limited protein content, which is well-characterized in healthy control subjects [92,93]. Cross-sectional analysis will help identify diagnostic biomarkers, but CSF markers of disease severity must be identified through serial sampling of biofluids and intra-subject analyses.
To this end, the Northeast ALS consortium has an ongoing study aimed at collecting, cataloging, and storing samples of serum, plasma, DNA and CSF. Serum and plasma are being collected serially and collection of serial CSF samples is just beginning.
The accurate test of limb isometric strength is a device under development to test 12 muscle groups in the limbs using a fixed, wireless load cell, which does not require the examiner to provide counterforce [94–96]. The accurate test of limb isometric strength testing generates force curves whose shape provides information about patient effort. The device remains in development, but confirmatory testing is planned and could provide a detailed functional outcome measure for clinical trials.
Electrical impedance myography is a technique for investigating the electrical properties of a localized area of muscle to evaluate the muscle tissue health. A high-frequency, low-intensity electrical stimulation is applied transdermally and recorded by a nearby electrode to derive muscle impedance, which changes as muscle health declines . It is highly reproducible and requires little training to perform . In the SOD1G93A rat model, electrical impedance myography phase slope correlated strongly with survival and motor unit number estimation .
As the basic and preclinical science discoveries in ALS have improved, ALS clinical trialists have implemented an increasing number of innovations to find ways to maximize trial efficiency and power, investigate potential new therapies and implement new assessment tools.
The evolving understanding of ALS genetics and pathophysiology, and an improving preclinical research tool set has led to a growing pipeline of potential therapies for ALS. Some of these potential therapies come from repurposing an FDA-approved medication, while others are novel compounds. Traditionally prescribed rigorous, but dogmatic, testing follows three clinical phases: Phase I trials are safety trials. Phase II trials are somewhat larger and provide further safety information and dose-selection. Large Phase III trials evaluate efficacy. The stalwart performance of the three-phase paradigm is admirable but lacks efficiency. Between each phase, the trial must be halted, databases locked, data cleaned and analyzed and results presented and published. As opportunities to move promising compounds from the laboratory bench into clinical research emerge, clinical researchers have begun to apply novel trial designs to improve efficiency. Unique trends have emerged in each of the trial phases .
Phase I trials evaluate the safety and tolerability of a study drug. Phase II trials extend safety testing and allow dose-ranging and selection. Most trial sponsors also include secondary outcomes aimed at evaluating efficacy in Phase II designs, but the trials lack statistical power to do so. This creates a conundrum for ALS clinical trialists – full evaluation of a promising preclinical drug requires Phase I–III testing, yet limited resources exist for testing many promising therapies. The development of new preclinical tools, biomarkers and surrogate outcome measures might reduce trial duration. Innovative designs can streamline trials, preserve safety and help trialists quickly evaluate more candidate drugs.
Phase I trials evaluate safety and provide information about human pharmacokinetics. Safety testing is usually accomplished by establishing the maximum tolerated dose (MTD) in humans, but when reactions are irreversible or cannot be monitored Phase I trials might instead test predesignated doses, define the drug half-life or assess for the minimum effective dose . Phase I trials can be performed in healthy volunteers, but most Phase I trials in ALS are performed in patients. The ethical considerations are complex . To improve efficiency, some ALS Phase I trials now allow subjects to enroll in multiple dosage cohorts. For some novel therapies, investigators or regulatory agencies choose to include a placebo group in Phase I trials, although there is no strong statistical argument for doing so.
Traditional designs use a fixed dose-escalation scheme to reach MTD. Once a predetermined number of patients has tolerated a given dose, the dose is increased for the next group. The continual reassessment model (CRM), applies mathematical modeling and incorporates real-time data to modify dose escalation [103,104]. If more toxicity is encountered, the dose escalation is slowed or stopped. This method has been shown to improve trial efficiency and MTD estimation accuracy [103,104]. However, CRM is statistically complicated and can lead to rapid dose escalation, which has hampered its use . Most CRMs in current use include rules that modify the underlying model to prevent very rapid dosage escalation. Given the potential benefits, CRM designs hold promise for Phase I ALS trials.
Phase II designs fill numerous roles. For investigators they provide dose ranging to identify a dosage level for a Phase III trial. For patients, they may be seen as providing access to investigational therapy, for regulatory bodies they provide additional safety data prior to launching a Phase III trial, and finally, for sponsors, these trials may provide preliminary hope of efficacy. Both selection and futility designs aim to improve Phase II trial efficiency.
Futility trials are designed to abandon large-scale efficacy testing if a given threshold for benefit is not met. The design may lead investigators to erroneously designate treatments with small benefits as futile (type II error). In light of this, it is less frequently employed by pharmaceutical company sponsors. However, academic investigators have used futility designs to evaluate potential therapies whose efficacy has been suggested. One example is co-enzyme Q10 for ALS . A preliminary dose-ranging stage was followed by a futility design to quickly assess the promise of a larger trial. Ultimately, the drug failed to surpass futility thresholds and further trials were abandoned .
Phase II selection design trials run multiple different drugs, or multiple drug/dosing combinations against one another with the goal of carrying the most promising one forward to a Phase III trial. Selection design trials eliminate placebo arms and test candidate drugs against one another. They do not assess whether the selected treatment is beneficial relative to placebo, but a secondary non-inferiority test against placebo or historical control can partially address this issue. However, without a true placebo control, data from selection design trials cannot be used to infer treatment benefit, only to select a candidate drug for further study.
The first selection design study in ALS compared minocycline 100 mg + creatine 10 g twice daily to celecoxib 400 mg + creatine 10 g twice daily for ALS and included a non-inferiority test against historical controls [108,109]. The trial demonstrated a more favorable response to celecoxib–creatine and non-inferiority to placebo using a pool of just 60 participants. As understanding of ALS grows and the number of candidate drugs for its treatment proliferates, the benefits of selection design trials become increasingly important and may grow in popularity.
For a decade, Phase III trials have embraced well-established surrogate outcomes such as ALSFRS-R or VC in place of, or more commonly in addition to, tracheostomy-free survival. Surrogate outcomes, for the purposes of Phase III trials, must be easily measured, reliable, valid and clinically relevant. Trialists are eager to apply such surrogate outcomes in clinical trials when they become available. Collaboration among trial centers and multinational clinical trials are increasingly being used to speed up Phase III trial enrollment.
The ALSFRS-R, now a trusted surrogate outcome for trialists and regulatory bodies, can be a challenge for statistical analysis. Slopes of ALSFRS-R decline can be compared using linear mixed effects modeling, but this method does not take death into account. The development of the ALSFRS-R Joint Rank  has allowed for the simultaneous analysis of survival and ALSFRS-R decline in a combined end point. The Joint Rank statistic ranks study participants in each treatment group, first by survival and then by ALSFRS-R score. The Joint Rank can increase power relative to analysis of either ALSFRS-R or survival analysis alone in some circumstances, for example when mortality rates are high. It has preliminarily been accepted by oversight committees as an acceptable outcome measure for Phase III registration trials.
There is now a Center for Disease Control (CDC)-funded National ALS Registry, which could become a critical resource for trial recruitment and epidemiologic studies in ALS, but patient confidentiality concerns will have to be addressed first. Disease-specific networks, such as the Great Lakes ALS Group, WALS Study Group, the Canadian ALS Consortium, NEALS Consortium and European ALS Consortium have mobilized ALS clinical trialists, prepared centers to participate in multicenter trials, provided support for investigators designing multi-center trials and improved multicenter trial efficiency tremendously. Even more centralized networks using federated (centralized) institutional review board processes and standard contracts for all sites could provide additional efficiency. In 2010, dextromethorphan/quinidine was approved under the brand name Nuedexta® for the treatment of pseudobul-bar affect due to ALS or MS [111,112]. Clinical trials of the medication employed the use of ALS trial networks to perform these trials efficiently.
In addition to improving design to add efficiency to ALS clinical trials, trialists are now focusing on international opportunities to increase the number of potential centers. With proper site training, this approach has already met with significant success in trials and the trend is expanding.
The traditional placebo-controlled trial design continues to be critical for Phase III clinical trials, to enable determination of efficacy and safety. It is also the only type of Phase III trial currently accepted by regulatory agencies. Some trialists routinely use 2:1 randomization schemes to entice potential participants to enroll. Others use traditional 1:1 designs, but include an open-phase extension at the end of the trial to allow all participants an opportunity to receive the study drug and thus encourage enrollment.
Genetic discoveries in ALS have provided potential opportunities to explore gene therapy in humans in the form of antisense oligonucleotides (ASOs) and RNA-silencing (small interfering RNA). Both reduce gene expression in a targeted fashion through Watson–Crick base-pair binding to a specific mRNA and have shown promise in preclinical research . Gene therapy using siRNA has shown mixed results in ALS [114,115], but is plagued by two major challenges; delivery and off-target effects. First, the short half-life and modest cellular uptake have led researchers to use various viral vectors to transport the siRNA to cells. However, viral vectors may themselves carry risk. Second, off-target effects are common, perhaps because siRNA disrupts cellular mechanisms used to process naturally occurring miRNA. ASOs do not cross the blood–brain barrier, so must be delivered to the CNS directly, but cause few side effects and, for now, appear more promising as a potential gene therapy for ALS. The relative lack of experience with gene therapy in humans has led oversight organizations to take a conservative approach in most cases, which has tended to slow clinical trial progress.
The use of stem cells as a preclinical model has been addressed, but stem cell transplantation into the spinal cord or motor cortex might also be used as a therapy. Transplanted stem cells might provide neurotrophic support to motor neurons and slow ALS progression. Motor neuron replacement is a more complex and distant goal of stem cell therapy . Other authors argue that cell-replacement therapy will not reverse the inflammatory changes or decrease in autophagy seen in ALS . For now, the first Phase I trial of human embryonic stem cell transplantation in patients with ALS has begun (trial description in ‘Examples of innovative ongoing clinical trials’) and additional groups have begun to explore stem cell transplantation in ALS patients. In the near future, safety concerns will likely mean that meticulously designed and implemented trials with a strong focus on safety evaluation will take precedent over efficiency in the field of stem cell transplantation for ALS.
Never before have such a number of promising clinical trials for ALS been underway at the same time. Trials of drugs aimed at a number of putative pathologic processes are being tested. Each of these selected ongoing trials represents a step forward in ALS clinical trials by implementing a novel design concept or tool.
This trial compared creatine (30 g/day) to two dosages of tamoxifen (40 mg and 80 mg twice daily). The selection design was used to rapidly select the most promising of two potential ALS therapies for further testing.
A pilot study published in 2010 investigated the effect of creatine on brain chemistry using MRS and revealed that at 5 and 10 g twice daily (doses tested in prior ALS trials ), no changes were seen on MRS . At 15 g twice daily, the MRS revealed increased creatine (8%; p = 0.06) and decreased glutamate and glutamine levels (−17%; p = 0.039). This provided a small, unpublished study suggesting that tamoxifen can slow disease in a dose-dependent fashion.
Progression of ALS may be due, at least in part, to glutamate excitotoxicity, and in preclinical trials ceftriaxone upregulated EAAT2, the glial glutamate transporter responsible for clearing glutamate from the synaptic cleft [120,121]. On these strengths, a clinical trial of ceftriaxone for the treatment of ALS was designed and is ongoing. In order to hasten the study, while preserving its power to detect efficacy and monitor safety, a multistage, nonstop, sequential design was adopted. The total study length from first enrolled subject was planned to be approximately 53 months. In stages 1 and 2, 60 subjects were to be enrolled, with 8 months of enrollment for stage 1 followed by 20 weeks of follow-up in stage 2. These initial 60 subjects would continue into stage 3. Stage 3 would include an additional 540 subjects, for a total of 600 subjects at up to 50 sites in the USA and Canada .
As noted above, the mechanism by which mutant SOD1 injures motor neurons is unknown, but preclinical models have demonstrated a dose–response relationship of mutant SOD1 protein to disease severity . Thus, decreasing levels of the mutant protein with a drug such as ISIS 333611 (Isis Pharmaceuticals) may be therapeutic. ISIS 333611 is a modified ASO to SOD1. Currently, a Phase I trial to examine the safety and tolerability of a single 12 h intrathecal infusion of ISIS 333611 is underway. The trial utilizes a dose-escalation design with four dosage level cohorts, each consisting of eight subjects, two of whom receive placebo. This trial is making important inroads in designing trials to test gene therapy in ALS. Given the rarity of the patient population, study subjects have been permitted to participate in more than one dosecohort, providing that they meet eligibility criteria at the time of re-enrollment and that at least five half-lives of the study drug have passed, to ensure complete washout. This design feature offers increased trial efficiency in a study that draws from a small patient population .
NP001 (Neuraltus Pharmaceuticals) is a small-molecule drug aimed at altering macrophage activation patterns with the goal of reducing inflammatory activation in the CNS postulated to perpetuate the disease. A single-dose, Phase I trial of NP001 was conducted in 32 ALS participants in 2010, and a Phase II trial is now underway. Randomization is 1:1:1 placebo, low-dose and high-dose study drug. Study drug is administered in monthly infusion cycles. The trial of NP001 marks an important investigation into the role of neuroinflammation in ALS and the ability to alter this phenomenon to impact disease course. This type of innovative approach to the treatment of a well-described disease is too seldom investigated, given the significant investment required for a clinical trial .
The first human trial of stem cell therapy in the USA is underway (Neuralstem Inc.). It is an open-label Phase I trial of human spinal stem cells (HSSC) engineered from the spinal cord of a single fetus and injected into the spinal cord using a proprietary device. The primary outcome of the trial is safety and tolerability. Five cohorts will be enrolled. The first has been completed and consisted of six participants with advanced ALS who were nonambulatory. Three received unilateral lumbar cord injections and the next three bilateral injections. The following cohorts will have three ambulatory participants each and will involve injections in the lumbar spine (cohorts 2 and 3), cervical spine (cohort 4), and cervical and lumbar spine (cohort 5). Investigators enrolled participants with late-stage disease first and performed lumbar injections before cervical to minimize risk .
The landscape of clinical research in ALS is now flush with clinical trials. Preclinical tools that are just coming into existence, such as iPS-derived motor neurons, new transgenic animals and novel genetic sequencing tools, have the potential to facilitate high-throughput screening of candidate drugs. Once identified, imaging and/or biofluid biomarkers could become indispensible tools in clinical research.
New and innovative study designs will continue to increase efficiency of clinical trials in ALS. Trials of lithium, dexpramipexole and ceftriaxone for ALS incorporate nonstandard designs and are leading a wave of novel trial designs that will continue to evolve and improve clinical trial efficiency. The creation of clinical trial networks will be extended to national and international networks with central IRB, contracts and study coordination resources to reduce study startup time. Finally, clinical trialists will incorporate telemedicine and create forums for better patient communication. Social networking resources may become important tools for trial publicity, including subject recruitment and retention.
J Berry is supported by an NIH training grant (T32NS048005) and Muscular Dystrophy Association Clinical Research Training Grant. M Cudkowicz has worked as a consultant for Shire, Trophos, Cytokinetics and GlaxoSmithKline. She has received industry-sponsored grants from Neuralstem Inc., ISIS Pharmaceuticals Inc., Knopp Biosciences and Biogen Idec. She has also received funding from NIH grants: UO1 NS049640, UO1 NS052592, RC1 NS069395 and RC1 NS068179.
Financial & competing interests disclosure
The authors have no other relevant affiliations or financial involvement with any organization or entity with a financial interest in or financial conflict with the subject matter or materials discussed in the manuscript apart from those disclosed.
No writing assistance was utilized in the production of this manuscript.
Papers of special note have been highlighted as:
of considerable interest