PMCCPMCCPMCC

Search tips
Search criteria 

Advanced

 
Logo of nihpaAbout Author manuscriptsSubmit a manuscriptHHS Public Access; Author Manuscript; Accepted for publication in peer reviewed journal;
 
J Data Sci. Author manuscript; available in PMC 2010 April 5.
Published in final edited form as:
J Data Sci. 2008 October 1; 6(4): 515–531.
PMCID: PMC2849171
NIHMSID: NIHMS181404

A Solution to Separation and Multicollinearity in Multiple Logistic Regression

Abstract

In dementia screening tests, item selection for shortening an existing screening test can be achieved using multiple logistic regression. However, maximum likelihood estimates for such logistic regression models often experience serious bias or even non-existence because of separation and multicollinearity problems resulting from a large number of highly correlated items. Firth (1993, Biometrika, 80(1), 27–38) proposed a penalized likelihood estimator for generalized linear models and it was shown to reduce bias and the non-existence problems. The ridge regression has been used in logistic regression to stabilize the estimates in cases of multicollinearity. However, neither solves the problems for each other. In this paper, we propose a double penalized maximum likelihood estimator combining Firth’s penalized likelihood equation with a ridge parameter. We present a simulation study evaluating the empirical performance of the double penalized likelihood estimator in small to moderate sample sizes. We demonstrate the proposed approach using a current screening data from a community-based dementia study.

Keywords: Logistic regression, maximum likelihood, penalized maximum likelihood, ridge regression, item selection

1. Introduction

In dementia studies involving large community-based cohorts of elderly participants, screening tests are often administered to study participants in order to obtain a measure of cognitive function. The scores of the screening tests are then used to determine whether a study participant should undergo more intensive clinical assessment for diagnosis of dementia. Such screening tests are usually constructed using a number of items which aim at measuring various domains of cognitive function. In practical epidemiological or clinical research, however, there is always a need to shorten existing screening instruments in order to more efficiently gather the needed information in a limited amount patient time.

The selection of test items for such shortened screening tests has often involved univariate comparisons of percentage of correct responses between demented and normal subjects. Items are retained if they show significant difference in responses between the two groups of subjects. However, the univariate method only considers each individual item and ignores that responses for multiple items are often highly correlated. Combining items with the largest differences in item responses between demented and normal groups may not lead to the most discriminating overall test.

A natural alternative to the univariate method would be multivariate item selection using a multiple logistic regression model, which multivariately models the relationship between dementia outcomes and the screening items. The multiple logistic regression model serves several purposes. One is that the model offers the collective predictive accuracy of the items. The second appeal is that the model also provides a linear combination of all test items which may be used as a score to predict the dementia outcomes. Lastly, the standardized parameter estimates from the logistic model also offer the rankings of test items in terms of strength of association with dementia.

Item selection using multiple logistic regression often encounters serious estimation problems when applied to screening data in dementia. Separation and multicollinearity are the two common problems in the logistic regression. The problems become exasperated in the dementia screening data because the two problems frequently occur together. These problems in logistic regression have led to aborted attempts by many investigators to shorten instruments for screening purposes using multiple logistic regression.

Separation in logistic regression frequently occurs when the binary outcome variable can be perfectly separated by a single covariate or by a non-trivial linear combination of the covariates (Albert and Anderson (1984)). Heinze and Schemper (2002) demonstrated that separation can happen even when the underlying model parameters are low in absolute value. They also showed that the probability of separation depends on sample size, on the number of dichotomous covariates, the magnitude of the odds ratios and the degree of balance in their distribution. Ceiling effect resulting from a high proportion of sample having maximum score in item response also increases the probability of separation.

Separation alone can cause infinite estimates or biased estimates. In cases where finite maximum likelihood estimates are available, correction of bias in maximum likelihood estimates of logistic regression have been studied by Anderson and Richardson (1979), McLachlan (1980), Schaefer (1983), Copas (1988), McCullagh and Nelder (1989) and Cordeiro and McCullagh (1991). In situations where finite maximum likelihood estimates may not exist, Firth (1992, 1993) introduced a penalized maximum likelihood estimator which reduced the bias of maximum likelihood estimates in generalized linear models. Firth’s approach guarantees the existence of estimates by removing the first order bias at each iteration step. Bull et al. (2002) extended Firth’s approach to multinomial logistic regression with nominal response categories, comparing it to maximum likelihood estimates and to maximum likelihood estimates corrected by an estimate of the asymptotic bias. They showed that Firth’s penalized maximum likelihood estimator was superior to the other methods in small samples and Firth’s estimator was equivalent to the maximum likelihood estimator as sample size increased. Heinze and Schemper (2002) applied Firth’s approach to logistic regression with separated data sets. Their simulation results demonstrated that Firth’s penalized likelihood estimator provided an ideal solution to the separation problem in logistic regression.

Multicollinearity is not uncommon when there is a large number of covariates. It may become a serious concern in dementia data because screening items are often highly correlated. Multicollinearity can cause unstable estimates and inaccurate variances which affects confidence intervals and hypothesis tests (Hoerl and Kennard, 1970, 1988). A ridge estimator originally developed for linear regression provides a way to deal with the problems caused by multicollinearity. The ridge estimator in general shrinks estimates towards the origin. The amount of shrinkage is controlled by the ridge parameter, whose size depends on the number of covariates and the magnitude of collinearity. The mean squared error (MSE) is guaranteed to be reduced accordingly by the introduction of ridge parameter (Schaefer et. al, 1984; Hoerl and Kennard,1970,1988). le Cessie and van Houwellingen (1992) applied the ridge regression method to logistic regression to improve parameter estimates and decrease prediction errors.

In this paper, we propose a double penalized maximum likelihood estimator for logistic regression models which combines Firth’s penalized likelihood approach with a second penalty term for ridge parameter which is capable of handling both separation and multicollinearity. In section 2, we describe the proposed approach in detail. In section 3, we evaluate the empirical performance of the proposed estimator in a simulation study. In section 4, We demonstrate the proposed approach using screening item data from a community-based dementia study.

2. Proposed Approach

2.1 Double penalized likelihood estimator

Let Yi be a binary response variable, i = 1, …, n. Let xi =(xi1, …, xip) be a p-dimentional row vector of covariates. We denote the covariate matrix by X. In a logistic model, we model the probability P(Yi = 1) by

P(Yi=1)=πi=11+expxiβ

where β is a p × 1 vector of parameters corresponding to the covariates. The log-likelihood function under the logistic regression model is given by

L(β)=i=1nli(β),

where

li(β)=Yilogπi+(1Yi)log(1πi).
(2.1)

Maximizing L(β) yields the maximum likelihood estimate (MLE) of β.

Maximum likelihood estimates of β are found to be biased away from the point β = 0. The asymptotic bias of the MLE [beta] can be expressed as:

Bias(β^)=b1(β)n+b2(β)n2+.

Most bias correction methods are focused on removing the first-order bias from the asymptotic bias of [beta] by using:

β^corrected=β^b1(β^)n.

However, this method relies on the existence of the MLE [beta]. In small to medium sized data, or data with many covariates, it is not uncommon for [beta] to be infinite. Let A be the Fisher information matrix for l(β). Firth (1993) proposed the following penalized likelihood function:

L(β)=L(β)A1/2

with the penalized log-likelihood function of

l(β)=l(β)+12logA.
(2.2)

Firth (1992) noted that this penalized likelihood function is equivalent to the posterior distribution function for a canonical parameter of an exponential family model using the Jeffrey’s invariant prior (1946).

We propose to add a second penalty term to Firth’s penalized likelihood function by including a ridge parameter which forces the parameters to spherical restrictions:

l(β)=l(β)+12logAλ||Pβ||2,
(2.3)

where P is a p×p matrix which imposes linear constrains on the parameters of β and || || denotes the Euclidean norm of a parameter vector. When P is the identity matrix, all parameters in β equally shrink towards the origin. A special situation for P would be a partial diagonal matrix with 1 at some diagonal elements and 0 elsewhere and the parameters corresponding to 0 will not be subjected to any shrinkage. Without loss of generalization, we consider P to be an identity matrix in subsequent development.

In the double penalized likelihood function l**(β), λ is the ridge parameter which controls the amount of shrinkage of the norm of β. The choice of λ depends on the number of covariates and/or the degree of multicollinearity among the covariates.

In a simple and intuitive example, suppose we have just one independent variable, xi, which can take only two values: 0 or 1. We consider the simplest logistic regression of

P(Yi=1)=11+exiβ.

In other words, we consider a simple logistic regression with just one regression parameter without an intercept. Now let n11 = Σ(yi = 1|xi = 1), n01 = Σ(yi = 0|xi = 1), n1 = n11 + n01, n10 = Σ(yi = 1|xi = 0) and n00 = Σ(yi = 0|xi = 0). In addition, let p1=Prob(yi=1xi=1)=11+eβ. Since p0 = Prob(yi = 1|xi = 0) = 1/2, the likelihood function can be written as:

L=p0n10(1p0)n00p1n11(1p1)n1n11.

Now let C=p0n10(1p0)n00 and it can be seen that C is a constant not involving the parameter β. Hence the corresponding log-likelihood function can be written as:

l=logC+n11β+n1log(1p1).

The maximum likelihood estimate of β for this model is:

β^=logn11n1n11.

It is clear from the formulae above that when n1 = n11, meaning that there is no observation with the combination of xi = 1 and yi = 0, a complete separation of data will happen. Under such data separation, the maximum likelihood estimate for β is infinite, i.e. [beta] = ∞.

For this simple logistic model, Fisher’s information matrix is reduced to A = n1p1(1 − p1). Thus Firth’s penalized likelihood function for this model is written as:

L=Cp1n11+12(1p1)n1n11+12.

Parameter estimate from Firth’s penalized likelihood function is therefore:

β^=logn11+12n1n11+12.

Notice that Firth’s penalized likelihood estimate is essentially the frequently used method of adding 12 to a zero cell.

Our proposed double penalized log-likelihood function is:

l=logC+(n11+12)β+(n1+1)log(1p1)2λβ.

Parameter estimate from the double penalized likelihood approach for this simple logistic model becomes:

β^=logn11+δn1n11+(1δ),

where δ=12λ. It can be seen that when λ = 0, δ=12, thus the double penalized likelihood approach is the same as Firth’s penalized approach. From this sense, our proposed double penalized approach is an extension to Firth’s method with additional ability to choose λ to reduce multicollinearity. See Section 2.3 below for further discussion on the choice of λ.

2.2 Parameter estimation

The Newton-Raphson maximization algorithm procedure can be used to derive [beta] for β. The iterative Newton-Raphson algorithm is defined as:

β^(t+1)=β^(t)+(A)1(t)U(β^(t)),

where t stands for the number of iteration, (A**)−1(t) is the information matrix of the double penalized likelihood function in (2.3), and U**([beta](t)) is the first derivative of the log-likelihood function, i.e.

U(β)=l(β)β+β(12logA)2λβ.

A computationally convenient formulae in matrix form for β(12logA) was provided by Bull et al (2002) as:

β(12logA)=XQ(XX)vec(A1),

where Q = W [multiply sign in circle] ei, i = 1, …, n, W is a n × n diagonal matrix with element πi(1 − πi)(1 − 2πi), ei is a 1 × n vector with 1 on the ith column and 0 elsewhere, vec(A−1) is the vectorized A−1.

Since the information matrix for the double penalized likelihood function, A=U(β)β+122logAβ22λI, is difficult to derive, we propose to use

AU(β)β2λI.

In Bull et al. (2002) where Firth’s original penalized likelihood was considered, the authors proposed to use the information matrix from the likelihood function which also excluded Firth’s penality term. The approximate information matrix can also be used to provide variance estimates of the parameters at convergence. In a simulation study presented in the following section, we evaluate the approximate variance estimates using empirical variance estimates.

2.3 The choice of the ridge parameter

The ridge parameter, λ, is generally chosen to minimize prediction errors of the logistic models. There are various definitions of the prediction errors in the literature. Efron(1986), le Cessie and van Houwelingen (1992), for example, defined three measures to quantify the prediction errors:

  1. classification error (CE)
    CE={1ifYnew=1andπ^<12;orYnew=0andπ^>12,12ifπ^=12,0otherwise;
  2. squared error (SE)
    SE=(Ynewπ^)2;
  3. minus log-likelihood error (ML)
    ML={Ynewlogπ^+(1Ynew)log(1π^)}.

The advantages and disadvantages of these methods were reviewed in le Cessie and van Houwelingen (1992). Unlike the classification error approach which considers model prediction in the neighborhood of π=12, the squared error approach and minus log-likelihood error approach calculate the prediction errors in the entire range of π values. In this paper, we focus the squared error as a measure of prediction errors because of its computational convenience.

The mean squared error (MSE) is calculated as an index based on the entire data set when comparing the effects of different ridge parameters. In lack of an external validation dataset, we used cross-validated estimate of the mean squared error (MSE) which is defined as:

MSEcvλ=1ni=1n(Yiπ^iλ)2,

where π^iλ denotes the prediction of πi at a given λ based on parameter estimates obtained with the ith observation left out. The optimal λ was chosen to minimize the MSE.

For ease of computation in the cross-validation, we adopted a one-step approximation method proposed by Cook and Weisberg (1982) for the estimation of β(i)λ, the parameters with ith observation left out:

β^(i)λ=β^λ{A}1XiT(Yiπ^iλ)1hii,

where hii denotes the ith diagnonal element in the hat matrix generated from Ã**. Similar to le Cessie and van Houwelingen (1992) MSE from cross-validation can be approximated by:

MSEcvλ1ni=1n(Yiπ^iλ)2(1hii)2.

The asymptotic consistency of the double penalized likelihood estimator was shown in Gao and Shen(in revision). We investigate the empirical performances of the proposed estimator in small to moderate samples in the following section.

3. A Simulation Study

A simulation study was conducted to evaluate the empirical performance of the double penalized likelihood estimator (DPLE) in logistic regression and compared it to the maximum likelihood estimator (MLE) and Firth’s penalized likelihood estimator (PLE) with respect to mean bias and mean squared error (MSE). We considered various scenarios where there were a large number of covariates with potential high correlations. A combination of binary and continuous covariates was included in the simulation. We chose ten covariates with five binary and five continuous covariates in the simulation so that the demand of the large number of covariates was met. The sample sizes were set to be 30, 50, 80, 130 and 200 so that we could investigate how the performance changes with the change of sample sizes. We implemented the algorithm using SAS IML and macro languages (SAS 8.2). A SAS macro is available upon request.

We first generated 10 continuous variates from multivariate normal distributions with zero means, unit variances and a specified correlation matrix, followed by dichotomization at zero for the first five continuous variates to produce the five binary covariates. The correlation coefficients among the variates from the multivariate normal distributions were set to be equally as high as 0.8. The binary response variable was generated by comparing the predicted probabilities from the logistic model with true parameter values to a random number from the uniform distribution in [0, 1].

We simulated 1000 datasets for each sample size scenario. Our results confirmed that the proportion of non-existence for the regular MLEs were extremely high in small samples but decreased as sample size increased. In our simulations, none of the datasets had finite MLEs when the sample size was 30. The proportion of having finite MLEs increased as sample size increased (5%, 33%, 83% and 100% for sample size 50, 80, 130 and 200 respectively). Results of the simulations are presented in Table 1.

Table 1
Simulation results for logistic regression with 10 covariates using three estimation methods

In general, both penalized maximum likelihood estimators had significantly smaller mean biases and MSEs compared to the MLEs. The differences are larger when the sample size is small or medium (see Table 1). In Figure 1, we plotted the mean bias and the MSE of the PLE and DPLE for different sample sizes. It can be seen that DPLE has bigger bias than PLE and the MSE of DPLE was smaller compared to that of PLE. This trend is more evident when the sample size is small or moderate. As the sample size increased, the two procedures became indistinguishable with respect to the mean bias and MSE. We point out that these results are expected because the introduction of the ridge parameter in the DPLE sacrifices bias for the gain in MSE.

Figure 1
Plots of mean biases and mean squared errors for sample sizes and covariate types in logistic regression

There seem to be differences in the parameter estimates for the binary and continuous covariates between the PLE and DPLE estimators with respect to mean bias and MSE in our simulation. The ridge effect appeared to be stronger in continuous covariates than in binary covariates. For the binary covariates, the reduction of MSE in DPLE compared to PLE clearly showed a compromise in terms of increased bias. For the continuous covariates, the MSE was significantly reduced in the DPLE method. However, the mean bias was not significantly affected by the addition of the ridge parameter. DPLE and PLE differed most notably in smaller sample sizes and their differences become negligible when the sample size gets large.

We also included the percentage of estimated bias using the approximate variance estimates compared to empirical variance estimates based on the simulation in Table 1. The approximate information matrix overestimated the variances in both DPLE and PLE estimators when the sample size was small. The magnitude of overestimation is significantly reduced as sample size increased.

4. Indianapolis Dementia Study

The Indianapolis Dementia Study is a community-based prospective cohort study of Alzheimer’s disease and dementia in elderly African Americans in Indianapolis, Indiana. Data was collected during one baseline wave and four follow-up waves at 2, 5, 8 an 11 years after baseline. An enrichment sample of additional subjects was added at the third follow-up wave when blood samples were also collected from those subjects who consented for biochemical analyses. The Community Screening Interview for Dementia (CSID) was used as a screening instrument to evaluate and stratify study subjects’ cognitive function (Hall et al., 1996). Items in the CSID were selected from various well-known neuropsychological tests to measure the following functions in an interview: memory, abstract thinking, reasoning and judgement. There are 33 test questions in the CSID. At the third follow-up examination, 2628 elderly African-American subjects were screened with the CSID and stratified into three performance groups, namely poor, intermediate and good performance. Four hundred and fifty eight subjects were randomly selected with the sampling probabilities of 100%, 50% and 5%, respectively, in the poor, intermediate and good performance groups to undergo extensive clinical evaluation for dementia and Alzheimer’s disease.

In this analysis, we selected 5 binary and 5 continuous items from the CSID which were presumed predictive of dementia. We included subjects who were administered the CSID for the first time at the 8-year wave and who also had biochemical measures available. Furthermore, we excluded subjects with a diagnosis of mild cognitive impairment and subjects with incomplete responses to screening items, resulting in a sample of 59 subjects, of whom 21 were demented and 38 were normal. The items are described in details as below:

  1. Do you remember my name? (interviewers introduced themselves at the beginning of interview and asked the subject to remember their names for later recall)
  2. Who is the Mayor of the city?
  3. What is the name of the civil rights leader who was assassinated in Memphis in 1968?
  4. Who is the current president in the USA?
  5. Who is the current governor of Indiana?
  6. Place orientation. It is generated from the following 4 binary items:
    1. What is the name of the city?
    2. What are the two major streets near your home?
    3. What is the city market?
    4. What is your complete address, including your zip code?
  7. Time orientation. it is generated from the following 5 binary items:
    1. What day of the week is it?
    2. What month is it?
    3. What year is this?
    4. What season is it?
    5. Did it rain(snow)yesterday?
  8. Read a short story and then ask the subject to repeat as much of the story as he/she can.
  9. Recall the story from item 8 after a while.
  10. Ask the subject to give the names for as many different animals as he/she can in one minute.

The first 5 items are binary with correct and incorrect responses and the last 5 items are continuous. The items “Place orientation” and “Time orientation” are highly correlated (r = 0.74). Table 2 presents demographic characteristics and item responses between the demented and normal groups. All items except “Governor” have significantly different responses univariately between the demented and normal.

Table 2
Demographic characteristics and item properties in the demented and normal groups

The logistic regression models using the MLE, PLE and DPLE are presented in Table 3. Figure 2 demonstrated that the cross-validated estimate of the MSE changes as a function of the ridge parameter. The optimal ridge parameter is chosen to be 0.01 for this data set. Table 3 showed the two penalized MLEs are significantly smaller in magnitude than the regular MLEs for all items. The DPLEs are closer to zero than the PLEs. The ridge effects are clearly demonstrated in the items “Place Orientation” and “Time Orientation”, in which there are about 16% and 14% reduction of estimates respectively in the DPLE method compared to the PLE method. This is because the two items are highly correlated as mentioned early and the ridge parameter penalized the multicollinearity effect.

Figure 2
Cross-validated estimates of the mean squared errors as a function of the ridge parameter
Table 3
Parameter estimates and asymptotic standard errors in logistic regression for dementia using MLE, PLE and DPLE.

In this dataset, finite MLEs were available in the regular logistic regression model given the sample size of 59. However, based on our converged simulation results we expect the MLEs to have larger mean squared errors than either the PLEs or the DPLEs.

5. Discussion

In this paper, we proposed a double penalized maximum likelihood estimator for logistic regression model, providing a solution to the problems of separation and multicollinearity for multivariate item selection in dementia. We demonstrated with simulation results that the double penalized estimator achieves minimum mean squared error at the expense of tolerable amount of bias in small to moderate samples. Most importantly, the double penalized estimation approach yields viable estimates in cases which maximum likelihood estimates do not, therefore providing an attractive alternative to maximum likelihood estimator in logistic regression when a large number of covariates have to be considered. In addition, we acknowledge that the double penalized approach can be applied to any GLIM model though we focused on our discussion in logistic regression model.

Acknowledgments

This research was supported by NIH grants: R01 AG15813, R01 AG09956 and P30 AG10133. We thank Dr. Shelley Bull for providing us with a GAUSS program for comparison with our program.

Contributor Information

Jianzhao Shen, Division of Biostatistics, Department of Medicine, Indiana University School of Medicine, 1050 Wishard Blvd. RG4101, Indianapolis, IN 46202-2872, USA.

Sujuan Gao, Division of Biostatistics, Department of Medicine, Indiana University School of Medicine, 1050 Wishard Blvd. RG4101, Indianapolis, IN 46202-2872, USA.

References

  • Albert A, Anderson JA. On the existence of maximum likelihood estimates in logistic regression models. Biometrika. 1984;71:1–10.
  • Anderson JA, Richardson SC. Logistic discrimination and bias correction in maximum likelihood estimation. Technometrics. 1979;21:71–78.
  • Bull SB, Mak C, Greenwood CMT. A modified score function estimator for multinomial logistic regression in small samples. Computational Statistics & Data Analysis. 2002;39:57–74.
  • Cook RD, Weisberg S. Residuals and Influence in regression. Chapman and Hall; 1982.
  • Copas JB. Binary regression models for contaminated data, with discussuion. Journal of Royal Statistics Society Series B. 1988;50:225–265.
  • Cordeiro GM, McCullagh P. Bias correction in generalized linear models. Journal of Royal Statistics Society Series B. 1991;53:629–643.
  • Efron B. How biased is the apparent error rate of a prediction rule? Journal of the American Statistiscal Association. 1986;81:461–470.
  • Firth D. Bias reduction, the Jeffreys prior and GLIM. In: Fahrmeir L, Francis B, Gilchrist R, Tutz G, editors. Advances in GLIM and Statistical Modelling. Springer; 1992.
  • Firth D. Bias reduction of maximum likelihood estimates. Biometrika. 1993;80:27–38.
  • Gao S, Shen J. Asymptotic properties of a double penalized maximum likelihood estimator in logistic regression. Letters of Probability and Statistics. 2007;77:925–930.
  • Hall KS, et al. A cross-cultural community based study of dementias: methods and performance of the survey instrument Indianapolis, U.S.A and Ibadan, Nigeria. International Journal of Methods in Psychiatric Research. 1996;6:129–142.
  • Heinze G, Schemper M. A solution to the problem of separation in logistic regression. Statistics in Medicine. 2002;21:2409–2419. [PubMed]
  • Hoerl AE, Kennard RW. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics. 1970;12:55–67.
  • Hoerl AE, Kennard RW. Plus Supplement. Vol. 8. Wiley; New York: 1988. Ridge Regression, Encyclopedia of Statistical Sciences 9 vols; pp. 129–136.
  • Jeffreys H. An invariant form for the prior probability in estimation problems. Proc R Soc A. 1946;186:453–61. [PubMed]
  • le Cessie S, van Houwelingen JC. Ridge estimators in logistic regression. Applied Statistics-Journal of the Royal Statistical Society Series C. 1992;41:191–201.
  • McLachlanG J. A note on bias correction in maximum likelihood estimation with logistic discrimination. Technometrics. 1980;22:621–627.
  • McCullagh P, Nelder JA. Generalized linear models. 2. Chapman and Hall; 1989.
  • SAS Propriety Software Version 8.2. SAS institute Inc; Cary, NC: 2001.
  • Schaefer RL, Roi LD, Wolfe RA. A ridge logistic regression. Communications in Statistics - Theory and Methods. 1984;13:99–113.
  • Schaefer RL. Bias correction in maximum likelihood logistic regression. Statistics in Medicine. 1983;2:71–78. [PubMed]
  • Standardized Occupational Classification Manual. Washinton, DC: Office of Federal Statistical Policy and Standards; 1980.