Home | About | Journals | Submit | Contact Us | Français |

**|**HHS Author Manuscripts**|**PMC2884971

Formats

Article sections

- Abstract
- Introduction
- Model specification in the CRM
- Choosing dose levels based on indifference intervals
- Calibrating δ in finite-sample settings
- General Applications
- Applications
- Discussion
- References

Authors

Related links

Clin Trials. Author manuscript; available in PMC 2010 June 14.

Published in final edited form as:

PMCID: PMC2884971

NIHMSID: NIHMS203347

Department of Biostatistics, Mailman School of Public Health, Columbia University.

Corresponding author: Shing M. Lee, Department of Biostatistics, Mailman School of Public Health, Columbia University, 722 West 168th Street, New York, New York 10032, U.S.A. ude.aibmuloc@4112lms

The publisher's final edited version of this article is available at Clin Trials

See other articles in PMC that cite the published article.

The continual reassessment method (CRM) is an adaptive model-based design used to estimate the maximum tolerated dose in dose finding clinical trials. A way to evaluate the sensitivity of a given CRM model including the functional form of the dose-toxicity curve, the prior distribution on the model parameter, and the initial guesses of toxicity probability at each dose is using indifference intervals. While the indifference interval technique provides a succinct summary of model sensitivity, there are infinitely many possible ways to specify the initial guesses of toxicity probability. In practice, these are generally specified by trial and error through extensive simulations.

By using indifference intervals, the initial guesses used in the CRM can be selected by specifying a range of acceptable toxicity probabilities in addition to the target probability of toxicity. An algorithm is proposed for obtaining the indifference interval that maximizes the average percentage of correct selection across a set of scenarios of true probabilities of toxicity and providing a systematic approach for selecting initial guesses in a much less time consuming manner than the trial and error method. The methods are compared in the context of two real CRM trials.

For both trials, the initial guesses selected by the proposed algorithm had similar operating characteristics as measured by percentage of correct selection, average absolute difference between the true probability of the dose selected and the target probability of toxicity, percentage treated at each dose and overall percentage of toxicity compared to the initial guesses used during the conduct of the trials which were obtained by trial and error through a time consuming calibration process. The average percentage of correct selection for the scenarios considered were 61.5% and 62.0% in the lymphoma trial, and 62.9% and 64.0% in the stroke trial for the trial and error method versus the proposed approach.

We only present detailed results for the empiric dose toxicity curve, although the proposed methods are applicable for other dose toxicity models such as the logistic.

The proposed method provides a fast and systematic approach for selecting initial guesses of probabilities of toxicity used in the CRM that are competitive to those obtained by trial and error through a time consuming process, thus, simplifying the model calibration process for the CRM.

The estimation of the maximum tolerated dose (MTD) is considered the main objective of phase I clinical trials. The MTD is defined as the dose at which a specified proportion of the patients will experience dose limiting toxicity (DLT). Various methods for estimating the MTD have been proposed throughout the years among them the continual reassessment method (CRM; [1]). This method uses a single parameter dose toxicity model to describe the relationship between dose and probability of DLT. Using the observed toxicity data from each patient, the model parameter is estimated and the dose associated with the DLT probability closest to a prespecified target is administered to the next patient. The CRM, being a Bayesian model-based method, requires the specification of (i) the functional form of dose-toxicity model, (ii) the prior distribution of its parameter, and (iii) the ‘dose levels’. While Chevret [2] studies extensively the impacts of the functional form of the model and the prior distribution of its parameter on the performance of the method, the selection of the ‘dose levels’ has not been addressed in the literature. When using the CRM, the dose levels are not the doses administered, but rather they are obtained via backward substitution of the initial guesses of the DLT probabilities in the model. Ideally, these initial guesses correspond to the ‘desirable range of toxicity probabilities for testing’ [1]. However, in most circumstances that information is not available and selection of the initial guesses is performed in an ad hoc manner. In practice, they are generally chosen by trial and error based on extensive simulations to examine the operating characteristics of various sets of initial guesses. This process is complicated since there are as many initial guesses as the number of doses.

The challenges in the calibration of CRM are presented using two motivating examples. The first one is a dose finding trial in patients with previously untreated diffuse large B cell or mantle cell non-Hodgkin’s lymphoma [3]. The main objective of the trial was to determine the MTD of VELCADE when administered in combination with CHOP + Rituximab (CHOP-R). DLT was defined as life threatening or disabling neurologic toxicity, very low platelet count or symptomatic non-neurologic or non-hematologic toxicity requiring intervention. The target probability of DLT was 0.25. Eighteen patients were treated for six 21-day cycles (126 days). The standard dose for CHOP-R was administered every 21 days. There were five doses of VELCADE with the third dose level being the starting dose. Dose escalation was conducted according to the time-to-event CRM (TITE-CRM; [4]). In this paper, however, we will focus on the dose-toxicity modeling issue and disregard the time-to-event component.

The second motivating example is a dose finding trial of short-term high dose lovastatin for the treatment of stroke within 24 hours of symptom onset [5]. The MTD was defined as the dose associated with a target probability of DLT of 0.10, where DLT was defined as liver or muscle toxicity based on changes in laboratory markers or symptoms. The treatment was administered to 33 patients for 30 days, with dose escalation for lovastatin occurring in the first 3 days and all patients receiving 20 mg/day for the following 27 days. The five doses of lovastatin that were evaluated up to 30 days after treatment were 1, 3, 6, 8 and 10 mg/kg. Dose escalation for the stroke trial was done in two stages. The first 3 patients were assigned the first dose level. If no DLT was observed, dose escalation was done as described in Table 1 per the coherence principle [10]. Once a DLT is observed, the trial would be switched to the TITE-CRM.

The second author served as the statistician on both trials. Initial guesses of the DLT probabilities were chosen based on model sensitivity [7] and simulated operating characteristics under a pre-specified set of dose-toxicity configurations. As there was no systematic way to choose the initial guesses for evaluation, many sets were tested resulting in a time-consuming ad hoc process. In both trials, the prior MTD was set at the third dose; thus the initial DLT rate at dose level 3 was set at the corresponding pre-specified target. Therefore, the search of the remaining initial guesses were done in a four-dimensional space. This could be a daunting task that deters applied statisticians from using the CRM. The motivation of this paper is to provide a systematic approach for choosing the initial guesses of the probabilities of DLT and consequently the dose levels. We propose selecting initial guesses used in the CRM by specifying a range of acceptable toxicity probabilities instead of a target probability of toxicity. An algorithm is provided for obtaining the range of acceptable toxicity probabilities by maximizing the average percentage of correct selection across a set of scenarios of true probabilities of toxicity. The algorithm provides a fast and systematic approach for selecting near optimal initial guesses in a much less time consuming manner than the trial and error method. Thus, it simplifies the design calibration process, and enhances the use of the CRM.

Suppose that we are interested in estimating the dose level associated with a target DLT probability *p _{T}*. Let

Empiric: *F*(*d*, β) = *d*^{β}

One-parameter logistic: *F*(*d*, β) = {1 + exp(−*a* − β*d*)}^{−1} where a is a fixed constant

As the CRM was originally proposed in a Bayesian framework, a prior distribution π(·) on the model parameter β is also assumed. Given the prior distribution and the data accrued up to the first *n* patients (i.e. the doses assigned to the patients and their corresponding toxicity outcomes), β can be estimated by the posterior mean (denoted as _{n}). The dose level recommended for the (*n* + 1)st patient is the dose with the model-based DLT probability closest to *p _{T}*, i.e.,

$$\text{arg}\phantom{\rule{thinmathspace}{0ex}}\underset{{d}_{j}}{\text{min}\phantom{\rule{thinmathspace}{0ex}}}|F({d}_{j};{\widehat{\beta}}_{n})-{p}_{T}|.$$

This process is continued until a pre-specified number of patients.

Here, we note an important distinction between the dose levels *d*_{1}, *d*_{2},…,*d _{K}* and the doses administered (e.g. 1 mg/kg, 3 mg/kg, etc.). In the context of the CRM, the dose level

Chevret suggests via simulations that the performance of the CRM as measured by the percentage of correct selection depends on the choice of *F*, with one-parameter logistic with *a* = 3 being superior; see Table II in [2]. This conclusion is based on an exponential prior π(β) = exp(−β) and a particular choice of the initial guesses of DLT probabilities. In what follows, we take a different approach by which the initial guesses are chosen by first fixing *F* and π, and demonstrate that by selecting the appropriate initial guesses, the empiric dose toxicity model can have comparable performance to the one-parameter logistic with *a* = 3.

Assume that we have a dose finding trial with *K* doses (*d*_{1}…*d _{K}*) and the target probability of DLT is

While these consistency conditions are theoretically revealing, their practical use is limited because the conditions involve the unknown true DLT probabilities µ_{k}’s. There fore, Cheung and Chappell suggested converting the intervals in the parameter space for β to intervals on the probabilty scale. Then, the indifference interval for a given correct dose level *l* was defined as an interval of DLT probabilities associated with the neighboring doses such that these neighboring doses may be selected instead of the true MTD (*l*). In notation, the indifference interval for the MTD (*l*) can be expressed as (NA, *F*(*d*_{l+1}, *b*_{l+1})) for *l* = 1, (*F*(*d*_{l−1}, *b _{l}*),

For example, in the lymphoma trial *p _{T}* is 0.25 and the dose toxicity model is assumed to be empiric where β [−5, 5]. Assume that the initial guesses of the probabilities of DLT at each dose are 0.05, 0.12, 0.25, 0.40 and 0.55, respectively. Then the sets are

In this article, instead of comparing different dose toxicity models using indifference intervals, we propose to calibrate the CRM model by first specifying the target length of an indifference interval. Specifying the indifference interval in turn facilitates the selection of the initial guesses of the probabilities of DLT due to the correspondence between indifference intervals and the initial guesses of the probabilities of DLT, as follows. Given the prior MTD (*v*), the total number of dose levels (K), the target probability of DLT (*p _{T}*) and the dose toxicity model (

$$F({d}_{v-1},{b}_{v})+F({d}_{v},{b}_{v})=2{p}_{T}\phantom{\rule{thinmathspace}{0ex}}\text{and}\phantom{\rule{thinmathspace}{0ex}}F({d}_{v},{b}_{v+1})+F({d}_{v+1},{b}_{v+1})=2{p}_{T}$$

(1)

$$F({d}_{v-1},{b}_{v})={p}_{T}-\delta \phantom{\rule{thinmathspace}{0ex}}\text{and}\phantom{\rule{thinmathspace}{0ex}}F({d}_{v+1},{b}_{v+1})={p}_{{\rm T}}+\delta $$

(2)

Equations (1) are based on the definition of *b _{k}* for,

Additionally, it can be shown that given the regularity conditions specified on the dose toxicity model, as δ increases, the distance between *p _{i}* and

Theorem 1: Let *F*(*d*, β) be the dose toxicity function such that it is strictly increasing in *d* for all β and strictly monotone in the parameter β in the same direction for all *d*, then |*p _{v+i}*(δ)−

The proof for Theorem 1 is provided in Appendix 2. From Theorem 1, we can infer that as δ increases, the set of initial guesses of probabilities of toxicity will be steeper. Simulations were also performed to examine the effect of δ on the prior probability of choosing a dose, *P*(β *H _{k}*). For example, in the lymphoma trial, if δ = 0.03 then

For large sample sizes, the method will guarantee that the target probability of DLT will fall in the specified indifference interval, however, it is necessary to provide some guidelines for the selection of δ in the finite sample setting. We expect that for finite samples the optimal δ value that yields the highest percentage of correct dose selection (PCS) will depend on the other design parameters (*p _{T}, K, v, F*) as well as the sample size. We propose an algorithm to calibrate the CRM by selecting the δ that yields the highest PCS given a calibration set of true probabilities of DLT.

Algorithm

- Let δ = 0.01
- Calculate the initial guesses of the probabilities of toxicity at each dose (
*p*_{1},…,*p*). If_{K}*F*(*d*, β) is the empiric model,$${p}_{i+1}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{T}\phantom{\rule{thinmathspace}{0ex}}+\phantom{\rule{thinmathspace}{0ex}}\delta )\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{i})}{\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{T}\phantom{\rule{thinmathspace}{0ex}}-\phantom{\rule{thinmathspace}{0ex}}\delta )}\right)\phantom{\rule{thinmathspace}{0ex}}\text{for}\phantom{\rule{thinmathspace}{0ex}}i=v,\dots ,K-1.$$(3)If$${p}_{i-1}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{T}\phantom{\rule{thinmathspace}{0ex}}-\phantom{\rule{thinmathspace}{0ex}}\delta )\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{i})}{\text{log}\phantom{\rule{thinmathspace}{0ex}}({p}_{T}\phantom{\rule{thinmathspace}{0ex}}+\phantom{\rule{thinmathspace}{0ex}}\delta )}\right)\phantom{\rule{thinmathspace}{0ex}}\text{for}\phantom{\rule{thinmathspace}{0ex}}i=2,\dots ,v.$$(4)*F*(*d*, β) is the logistic model,$$\begin{array}{ll}{d}_{i+1}\hfill & =\phantom{\rule{thinmathspace}{0ex}}\frac{(\text{log}(\frac{({p}_{T}+\delta )}{1-({p}_{T}+\delta )})-a))\phantom{\rule{thinmathspace}{0ex}}{d}_{i}}{\text{log}(\frac{({p}_{T}-\delta )}{1-({p}_{T}-\delta )})-a}\phantom{\rule{thinmathspace}{0ex}}\text{for}\phantom{\rule{thinmathspace}{0ex}}\mathrm{i}=v,\dots ,K-1\hfill \\ {d}_{i-1}\hfill & =\phantom{\rule{thinmathspace}{0ex}}\frac{(\text{log}\phantom{\rule{thinmathspace}{0ex}}(\frac{({p}_{T}-\delta )}{1-({p}_{T}-\delta )})-a))\phantom{\rule{thinmathspace}{0ex}}{d}_{i}}{\text{log}\phantom{\rule{thinmathspace}{0ex}}(\frac{({p}_{T}+\delta )}{1-({p}_{T}+\delta )})-a}\phantom{\rule{thinmathspace}{0ex}}\text{for}\phantom{\rule{thinmathspace}{0ex}}\mathrm{i}=2,\dots ,v.\hfill \\ \hfill {p}_{i}& =\phantom{\rule{thinmathspace}{0ex}}\frac{\text{exp}(a+{d}_{i})}{1+\text{exp}(a+{d}_{i})}\phantom{\rule{thinmathspace}{0ex}}\text{for}\phantom{\rule{thinmathspace}{0ex}}\mathrm{i}=1,\dots ,K.\hfill \end{array}$$ - Specify the calibration set and assume the K different scenarios of true probabilities of DLT follow the plateau configuration where
*p*=_{i}*p*for_{L}*i*<*l*,*p*=_{i}*p*for_{U}*i*>*l*and*p*=_{i}*p*for_{T}*i*=*l*where*l*= 1,…,*K*, ψ = 2,*p*=_{U}*p*ψ/(1 −_{T}*p*(1 − ψ)) and_{T}*p*=_{L}*p*/(ψ +_{T}*p*(1 − ψ))_{T} - Perform simulations using the CRM under each of the scenarios indicated in step 3 and obtain the corresponding PCS.
- Average the PCS across all
*K*scenarios of the calibration set. - Repeat steps 1–5 for δ values of 0.02 to the minimum of
*p*−0.005 and the upper limit of the desired indifference interval on a discrete domain with a grid width of 0.01._{T} - Select the δ that yields the highest average PCS.
- Calculate the initial guesses of the probabilities of toxicity at each dose (
*p*_{1},…,*p*) for the δ selected in step 7 and use these as the initial guesses of the probabilities of toxicity for the CRM._{K}The following steps are optional: - Repeat steps 1 through 8 with ψ = 3 and ψ = 5
- For step 7, in addition to selecting the δ that yields the highest average PCS, select the range of δ that yields PCS within 1 percentage point of the highest average PCS for each value of ψ.
- If the intersection of the ranges of δ for the various values of ψ is non empty, select the δ in the intersection that yields the highest average PCS across all values of ψ.
- If the intersection of the ranges of δ for the various values of ψ is empty, select the δ values that yield the highest average PCS for each ψ and perform a sensitivity analysis using the validation set to select the optimal value of δ.

- Calculate the initial guesses of the probabilities of toxicity at each dose (
*p*_{1},…,*p*) for the δ value or values selected in step 10 and use these as the initial guesses of the probabilities of toxicity for the CRM._{K}

The plateau configuration was selected as the calibration set because it provides the most conservative scenario for the CRM. The performance of the CRM will improve for scenarios when the true probabilities of DLT are steeper. The size of the jump is specified using the odds ratio, ψ. The *p _{L}* and

The proposed algorithm is used to obtain the δ that yields the highest average PCS for various scenarios of target probability of DLT, sample size, number of doses and prior MTD, given calibration sets using an odds ratio of 2. The target probabilities of DLT selected are 0.10, 0.20, 0.25 and 0.33. The sample sizes are 20, 25, 30, 35 and 40. The number of doses ranges from 4 to 7 and the prior MTD ranges from the lowest dose to the median dose. These were selected based on the most common scenarios encountered in practice. The dose toxicity model is assumed to be empiric with the prior distribution for β being normal with µ = 0 and σ^{2} = 1.34 in all cases. The model was chosen based on prior publications on the CRM [9]. We ran 2000 simulations for each scenario except for a sample size of 20 with 7 doses as on average there would be less than 3 patients per dose. The CRM did not allow dose skipping during escalation. The CRM also did not allow for dose escalation immediately after a DLT was observed [10]. However, that dose can be assigned after outcomes from more patients are obtained.

For each scenario, Table 3–Table 6 displays the δ that yields the highest average PCS as well as the range of δ for which the average PCS was within 1% of the optimal δ. Table 3, ,4,4, ,5,5, and and66 display the optimal δ corresponding to target probability of DLT of 0.10, 0.20, 0.25 and 0.33, respectively. It can be observed that the value of the optimal δ decreases as sample size and the number of doses increases. Additionally, the value of the optimal δ increases as a function of the prior MTD.

To illustrate the application of the tables to a particular study, we calibrate the hypothetical example presented in Chevret. The example had a sample size of 25 with a target probability of DLT of 0.25 (*p _{T}* = 0.25). It had six doses (K=6) with the third dose level being the prior MTD (

The validation set of true probabilities of DLT included the six scenarios originally included in Chevret. For each scenario of the validation set, we obtain the distribution of the MTD based on 2000 simulations using the initial guesses of probabilities of DLT used by Chevret (i.e. 0.05, 0.10, 0.25, 0.35, 0.50, 0.70) along with a logistic with a=3 and a=1 as the dose toxicity model and then using the initial guess corresponding to a δ of 0.08 along with an empiric dose toxicity model. For each scenario of the validation set, Table 7 displays the percentage with which each dose level is recommended (% Recommendation), the average absolute difference between the true probability of the dose selected (d*) and *p _{T}* (Average |

The Extended Algorithm is applied to calibrate the two motivating examples. The calibration is based on 2000 simulations. For both of the motivating examples, the dose toxicity model is assumed to be empiric (*F*(*d*; β) = *d*
^{exp(β)}) with the prior distribution for β being normal with µ = 0 and σ^{2} = 1.34. In addition, we performed 2000 simulations using the CRM to compare the initial guesses of probabilities of DLT obtained using the Extended Algorithm to those selected previously based on extensive simulations. The comparisons are done using a validation set of true probabilities of DLT. This is the set of true probabilities of DLT that were originally used when designing the trial. The CRM does not allow dose skipping during escalation. In addition, the CRM does not allow for dose escalation immediately after a DLT is observed.[10]

For the lymphoma trial the initial guesses of the probabilities of toxicity at each dose were 0.05, 0.12, 0.25, 0.40 and 0.55, respectively. These were selected based on extensive simulations evaluating the operating characteristics of various initial guesses of the probabilities of DLT based on 2000 simulations using the validation set. Table 8 shows the validation set of the true probabilities of DLT. The other design parameter values are: *p _{T}* = 0.25,

Lymphoma Trial: Operating characteristics of the extensive simulation, Extended Algorithm (δ=0.10), the optimal δ for ψ=5 (δ=0.12) and a δ value of 0.03 given *p*_{T} = 0.25 and the prior MTD is dose level 3. *p*(*d*) **...**

We calibrate the lymphoma trial using the Extended Algorithm. For each calibration set, we iterate the δ values between 0.01 and 0.24. Each δ value corresponds to a set of initial probabilities of DLT. For example, given a δ value of 0.05 which corresponds to an indifference interval of (0.20, 0.30), we can calculate the initial guesses of the probabilities of DLT using equations (3) and (4) as follow:

$$\begin{array}{l}{d}_{1}={p}_{1}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}(0.20)\text{log}(0.157)}{\text{log}(0.30)}\right)=0.084\\ {d}_{2}={p}_{2}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}(0.20)\text{log}(0.25)}{\text{log}(0.30)}\right)=0.157\\ {d}_{4}={p}_{4}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}(0.30)\text{log}(0.25)}{\text{log}(0.30)}\right)=0.355\\ {d}_{5}={p}_{5}=\text{exp}\phantom{\rule{thinmathspace}{0ex}}\left(\frac{\text{log}(0.30)\text{log}(0.355)}{\text{log}(0.20)}\right)=0.460\end{array}$$

Thus, the initial guesses of the probabilities of DLT at each dose are 0.084, 0.157, 0.250, 0.355, 0.460 respectively, and they yield an indifference interval of (0.20, 0.30).

The results are displayed in Figure 1. As expected, PCS increases with increasing odds ratio. However, the range of δ that yields high PCS varies very little depending on the odds ratio. For an odds ratio (ψ) of 2, the highest average PCS is 48.8% when δ is 0.09. The average PCS was within 1% for δ values between 0.06 and 0.10. For an odds ratio of 3, δ of 0.09 also yielded the highest average PCS of 64.6%. The range of δ is 0.09 to 0.13. The highest average PCS for an odds ratio of 5 was 79.0% which occurred when δ is 0.12. The range of δ is 0.10 to 0.14. Thus, across the various odds ratio a δ of 0.10 is the only δ value that gives an average PCS that is within 1% of the highest average PCS. The initial guesses of the probabilities of DLT at each dose for a δ of 0.10 are 0.011, 0.082, 0.25, 0.464, 0.654. We compare the performance of the initial guesses of the probabilities of DLT corresponding to a δ value of 0.10 to the initial guesses based on extensive simulations using the validation set. To evaluate the sensitivity of the performance to the δ selected based on different odds ratios, we also compared the above initial guesses to the initial guesses of the probabilities of DLT corresponding to a δ value of 0.12 which is the optimal for an odds ratio of 5. The initial guesses of the probabilities of DLT at each dose for a δ of 0.12 are 0.003, 0.058, 0.25, 0.509, 0.719. To evaluate the effect of δ on the operating characteristics, we also compared the above δ values to a δ value of 0.03.

The validation set of true probabilities of DLT included four scenarios. For each scenario of the validation set, Table 8 displays the percentage with which each dose level is recommended (% Recommendation), the average absolute difference between the true probability of the dose selected (d*) and *p _{T}* (Average |

For the stroke trial, the initial guesses of the probabilities of DLT based on extensive simulations were 0.02, 0.06, 0.10, 0.18 and 0.30 for each dose respectively. These were selected similarly to the ones for the lymphoma trial. Table 9 shows the validation set for the stroke trial. The other design parameter values are: *p _{T}* = 0.10,

Stroke Trial: Operating characteristics of the extensive simulation versus the Extended Algorithm given *p*_{T} = 0.10. The initial guesses obtained using the Extended Algorithm correspond to δ = 0.04. An initial design is used prior to observing the **...**

We compare the performance of the initial guesses of the probabilities of DLT corresponding to a δ value of 0.04 to the initial guesses previously chosen using the validation set. The validation set of true probabilities of DLT for the stroke trial included five scenarios. For each scenario of the validation set, Table 9 displays the percentage with which each dose level is recommended (% Recommendation), the average absolute difference between the true probability of the dose selected (d*) and *p _{T}* (Average |

To estimate the MTD using the CRM, it is necessary to specify the initial guesses of the probability of DLT at each dose. The literature on the CRM generally uses the same set of initial guesses of the probabilities of DLT that was used in the original paper published by O’Quigley et. al. [1] without any justification. The task of selecting the initial guesses can be daunting given the high dimensionality of the problem and the extensive number of simulations involved. We believe this deters many applied statisticians from using the CRM. The focus of this paper is to provide a systematic and simple approach for choosing the initial guesses of the probabilities of DLT. We provide the optimal δ for the most common scenarios encountered assuming an empiric dose toxicity model along with an example that illustrates the way to use the tables. By simplifying the selection of the initial guesses of the probabilities of DLT in practice, we hope to enhance the use of the CRM. All simulations in this paper can be easily performed using the getprior and the crmsim function in the R package ‘dfcrm’ [11], [12]. The getprior function calculates the initial guesses of the probabilities of DLT at each dose given δ, *p _{T}*, K,

The method proposed to select the initial guesses of the probabilities of DLT is based on the concept of indifference intervals. It can be argued that indifference interval is an asymptotic concept and sample sizes are small in dose finding trials. However, it has been demonstrated that poor asymptotic behavior implies poor finite sample operating characteristics. In cases when the investigator can specify the range of acceptable probabilities of toxicity, it is only necessary to calculate the initial guesses of the probability of toxicity at each dose based on that indifference interval and examine the operating characteristics given under scenarios of true probabilities of toxicity. In cases when the investigator can not specify a desired indifference interval, the proposed algorithm can be used to calibrate the CRM and to screen for model specifications. In addition, we propose an algorithm that chooses the optimal δ and provide the optimal δ for some common real life scenarios based on simulations examining the operating characteristics in the finite sample setting.

For all the examples and simulations, the dose toxicity model was assumed to be empiric as we generally use the empiric model in practice. The optimal δ will be different if the dose toxicity model is assumed to be logistic. These are not provided, but can be obtained using the approach described in the paper for the motivating examples. A logistic dose toxicity model can be used for the ‘dfcrm’ package.

- F is strictly increasing in
*d*for all β. - F is strictly monotone in the parameter β in the same direction for all
*d*. - Given any
*p*(0,1), for each_{T}*d*, there exists β in the interior of Θ, the parameter space, such that*F*(*d*; β)=*p*._{T} *F*(*d*, β) is bounded away from 0 and 1 for all_{k}*k*and β Θ.*F*′(*d*, β) =*F*(*d*, β)/β is uniformly bounded in β.- For each 0 <
*t*< 1 and each d, the functionis continuous and strictly monotone in β.$$t\frac{F\prime (d,\beta )}{F(d,\beta )}+(1-t)\frac{-F\prime (d,\beta )}{1-F(d,\beta )}$$

Theorem 1: Let *F*(*d*, β) be the dose toxicity function such that it is strictly increasing in *d* for all β and strictly monotone in the parameter β in the same direction for all *d*, then |*p _{v+i}*(δ) −

Without loss of generality, assume that *F*(*d*, β) is strictly decreasing in β for all *d, i* > 0 and δ_{1} < δ_{2}. Thus *p _{T}* <

For *i* = 1, we need to show that *p*
_{v+1}(δ_{1}) < *p*
_{v+1}(δ_{2}). Since *F*(*d*,β) is decreasing in β, ${F}_{{d}_{v}}^{-1}({p}_{T}-{\delta}_{1})<{F}_{{d}_{v}}^{-1}({p}_{T}-{\delta}_{2})$. This implies that, ${F}_{\beta ({\delta}_{1})}^{-1}({p}_{T}+{\delta}_{1})<{F}_{\beta ({\delta}_{2})}^{-1}({p}_{T}+{\delta}_{2})$ where *F*(*dv*,β(δ_{k})) = *p _{T}* − δ

Now, suppose that *p _{v+j}*(δ

Thus, by induction, *p _{v+i}*(δ

nsim<−2000

n<−25

target<−0.25

K<−6

nu<−3

# using the delta from calibration

bestdelta<−0.08

prior<−getprior(bestdelta, target, nu, K)

ptrue<−c(0.09, 0.16, 0.27, 0.38, 0.57, 0.75)

new<−crmsim(ptrue, prior, target, n, nu, nsim=nsim, count=FALSE)

1. O’quigley J, Pepe M, Fisher L. Continual Reassessment Method: A practical design for Phase I clinical Trials in Cancer. Biometrics. 1990;46:33–48. [PubMed]

2. Chevret S. The continual reassessment method in cancer phase I clinical trials: A simulation study. Statistics in Medicine. 1993;12:1093–1108. [PubMed]

3. Leornard JP, Furman RR, Cheung YKK, et al. Phase I/II trial of botezomib plus CHOP-Rituximab in diffuse large B cell (DLBCL) and mantle cell lymphona (MCL): Phase I results. Blood. 2005;106(11):147A–147A. 491 Part1.

4. Cheung YK, Chappell R. Sequential designs for phase I clinical trials with late-onset-toxicities. Biometrics. 2000;56(4):1177–1182. [PubMed]

5. Elkind MS, Sacco RL, MacArthur RB, et al. The Neuroprotection with Statin Therapy for Acute Recovery Trial (NeuSTART): an adaptive design phase I dose-escalation study of high-dose lovastatin in acute ischemic stroke. Int J Stroke. 2008;3(3):210–218. [PMC free article] [PubMed]

6. Cheung YK. Coherence principles in dose-finding studies. Biometrika. 2005;92(4):863–873.

7. Cheung YK, Chappell R. A simple technique to evaluate model sensitivity in the Continual Reassessment Method. Biometrics. 2002;58:671–674. [PubMed]

8. Shen L, O’Quigley J. Consistency of the continual reassessment method under model mispecification. Biometrika. 1996;83:395–405.

9. O’Quigley J, Shen L. Continual reassessment method: A likelihood approach. Biometrics. 1996;52:673–684. [PubMed]

10. Cheung YK. Coherence principles in dose-finding studies. Biometrika. 2005;92:863–873.

11. R Development Core Team. R Foundation for Statistical Computing. Vienna, Austria: 2008. R: A language and environment for statistical computing. ISBN 3-900051-07-0, URL http://www.R-project.org.

12. Cheung YK. dfcrm: Dose-finding by the continual reassessment method. 2008. R package version 0.1–1. http://www.columbia.edu/~yc632.

PubMed Central Canada is a service of the Canadian Institutes of Health Research (CIHR) working in partnership with the National Research Council's national science library in cooperation with the National Center for Biotechnology Information at the U.S. National Library of Medicine(NCBI/NLM). It includes content provided to the PubMed Central International archive by participating publishers. |