Next Article in Journal
Radio-Controlled Intelligent UGV as a Spy Robot with Laser Targeting for Military Purposes
Next Article in Special Issue
Partially Linear Additive Hazards Regression for Bivariate Interval-Censored Data
Previous Article in Journal
Two Reliable Computational Techniques for Solving the MRLW Equation
Previous Article in Special Issue
Statistical Inference of Truncated Cauchy Power-Inverted Topp–Leone Distribution under Hybrid Censored Scheme with Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Estimation of the Binary Response Model

by
Muhammad Amin
1,
Muhammad Nauman Akram
1,
B. M. Golam Kibria
2,
Huda M. Alshanbari
3,
Nahid Fatima
4 and
Ahmed Elhassanein
5,6,*
1
Department of Statistics, University of Sargodha, Sargodha 40100, Pakistan
2
Department of Mathematics and Statistics, Florida International University, FIU, Miami, FL 33199, USA
3
Department of Mathematical Sciences, College of Science, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
4
Department of Mathematics & Sciences, Prince Sultan University, Riyadh 11586, Saudi Arabia
5
Department of Mathematics, College of Science, University of Bisha, P.O. Box 551, Bisha 61922, Saudi Arabia
6
Department of Mathematics, Faculty of Science, Damanhour University, Damanhour 22511, Egypt
*
Author to whom correspondence should be addressed.
Submission received: 7 October 2022 / Revised: 27 January 2023 / Accepted: 28 January 2023 / Published: 8 February 2023
(This article belongs to the Special Issue Computational Statistics & Data Analysis)

Abstract

:
The binary logistic regression model (LRM) is practical in situations when the response variable (RV) is dichotomous. The maximum likelihood estimator (MLE) is generally considered to estimate the LRM parameters. However, in the presence of multicollinearity (MC), the MLE is not the correct choice due to its inflated standard deviation (SD) and standard errors (SE) of the estimates. To combat MC, commonly used biased estimators, i.e., the Ridge estimators (RE) and Liu estimators (LEs), are preferred. However, most of the time, the traditional LE attains a negative value for its Liu parameter (LP), which is considered to be a major drawback. Therefore, to overcome this issue, we proposed a new adjusted LE for the binary LRM. Owing to numerical evaluation purposes, Monte Carlo simulation (MCS) study is performed under different conditions where bias and mean squared error are the performance criteria. Findings showed the superiority of our proposed estimator in comparison with the other estimation methods due to the existence of high but imperfect multicollinearity, which clearly means that it is consistent when the regressors are multicollinear. Furthermore, the findings demonstrated that whenever there is MC, the MLE is not the best choice. Finally, a real application is being considered to be evidence for the advantage of the intended estimator. The MCS and the application findings pointed out that the considered adjusted LE for the binary logistic regression model is a more efficient estimation method whenever the regressors are highly multicollinear.

1. Introduction

The binary LRM (BLRM) is a common RM, and it is preferred in situations where the RV is dichotomous or binary. Forecasting a dichotomous RV is vital in epidemiology and medicine [1,2]. For example, a set of risk factors such as cholesterol and blood pressure effect on the possibility of undergoing a heart attack. The method of MLE is commonly employed to estimate the binary LRM [2,3]. In the case of MC, the variance becomes inflated, and the standard errors become high, and consequently, the t-statistic and F-ratios become insignificant [3,4,5]. To overcome the issue of MC, different biased techniques are available in the writings. One of them constructing on ridge regression (RR), see [6], for the linear RM. The logistic RR was first introduced in [7] via the results of [6]. For the same purpose three biased estimators are considered [3]. Lee and Silvapulle [8] introduced the two ridge parameters for the logistic RR estimator (LRRE). As in the LRRE, the ridge parameter plays a very important role. After these studies, many researchers focused on the collection of the ridge parameter in the LRRE [9,10,11,12,13,14]. Liu [15] discussed some limitations of the RRE and proposed another alternative estimator called the LE for handling the issue of multicollinearity in a better way. The LE for the BLRM was adapted and called the logistic LE (LLE), [16]. The theoretical and numerical properties of the LLE and its comparison with the MLE and LRRE have been discussed.
In the LLE, the LP (d) plays a main character in the estimation of the LLE and should lie between 0 and 1, i.e., 0 < d < 1 . Sometimes, the Liu parameter produces negative and zero values that affect the efficiency of the LLE [17]. Several authors introduced various types of Liu estimators in the BLRM to improve the efficiency of the LE [18,19,20,21,22]. These modified Liu estimators also often produced zero values for the Liu parameter. To overcome the limitations of the available LEs, Lukman et al. [23] introduced the modified one-parameter LE for the linear RM.
In this article, we propose a new adjusted LLE (ALLE) for the BLRM. Its theoretical properties are investigated. Furthermore, the properties of the new estimator are assessed via a theoretical comparison and compared with other estimation methods. The matrix mean squared error (MMSE) and the scalar mean squared error (MSE) are used as the performance evaluation criteria. To investigate the new estimator a MCS study has been performed. The forthcoming sections are organized as follows: Section 2 presents the construction of the proposed estimator for the BLRM with a theoretical comparison. A MCS study is presented in Section 3. Section 4 is devoted to analyzing prostate cancer data. Section 5 ends with some conclusions.

2. Statistical Method

Following [7,24,25], the BLRM is defined as
y i = δ i + ε i ,   i = 1 , , n ,
where ε i has zero mean and variance ω i = δ i 1 δ i with expectation δ i of y i and are independent for each i . Considering, the ith value of RV to be distributed as Bernoulli B e δ i , with
δ i = e x p x i t ξ 1 + e x p x i t ξ ,   i = 1 , , n ,
where δ = δ i i = 1 , 2 , , n , and X = x i i = 1 , 2 , , n t is the n × p data matrix with x i = 1 , x 1 i , , x p i t ; and ξ = ξ j j = 0 , 1 , p t is considered to be the p + 1 vector of regression coefficients. The regression coefficients vector ξ is usually estimated via the MLE method. The logarithm of the likelihood function for Equation (1) is computed as
H = i = 1 n y i l o g δ i + 1 y i l o g 1 δ i .
Let δ ^ T = δ ^ i T i = 1 , 2 , , n be the estimates of δ at the Tth step, using ξ ^ T and V ^ T = d i a g δ ^ i T 1 δ ^ i T . The following iterative algorithm is well define [26],
ξ ^ T + 1 = ξ ^ T + X t V ^ X 1 X T V ^ T y δ ^ T ,
The Equation (3) can be written as
ξ ^ M L E = Υ 1 X t V ^ z ^ ,
with Υ = X t V ^ X and z ^ = z i i = 1 , 2 , , n where η i = x i t ξ , while z i = η i + y i δ i η i / δ i . ξ ^ M L E has asymptotic normal distribution as ξ ^ M L E ~ N β , Υ 1 [8].
Let be the orthogonal matrix that has the eigenvectors of Υ as columns. Then the covariance and MMSE of the ξ ^ M L E are
C o v ξ ^ M L E = Υ 1   and   M M S E ξ ^ M L E = Γ 1 t .
The scalar MSE of the ξ ^ M L E is
M S E ξ ^ M L E = E ξ ^ M L E ξ t ξ ^ M L E ξ = t r Γ 1 t = j = 1 r 1 μ j ,
where Γ = d i a g μ 1 ,   μ 2 , , μ r .
In the case of high correlated regressors, the matrix Υ becomes ill-conditioned, which causes the problem of multicollinearity. The LRRE is defined by Schaefer et al. [7] as a simple extension of Hoerl and Kennard [6] to treat multicollinearity effects as
ξ ^ L R R E = R k ξ ^ M L E ,
where R k = Υ + k I r 1 Υ with the ridge parameter k ( k > 0 ) and identity matrix I r . The bias vector (BV), covariance matrix (CM) and MMSE of (7) are
B i a s ξ ^ L R R E = k Γ k 1 ξ ,
C o v ξ ^ L R R E = Γ k 1 Γ Γ k 1 t
and
M M S E ξ ^ L R R E = R k Υ 1 R k t + B i a s ξ ^ L R R E B i a s ξ ^ L R R E t = Γ k 1 Γ Γ k 1 t + k 2 Γ k 1 ξ ξ t Γ k 1 t ,
where Γ k = d i a g μ 1 + k ,   μ 2 + k ,   , μ r + k . From (10) by applying the t r   . operator, we obtain the scalar MSE of the LRRE as
M S E ξ ^ L R R E = t r   M M S E ξ ^ L R R E = j = 1 r μ j μ j + k 2 + k 2 j = 1 r τ j 2 μ j + k 2 ,
where τ = t ξ and k (k > 0) is the Hoerl and Kennard [6] ridge parameter. Optimizing the Equation (11) with respect to k yields
k = 1 j = 1 r τ ^ j 2 .
Another estimator that treats the multicollinearity better than the LRRE was given by Liu [25] and Mansson et al. [11] for the BRLM that is named as the LLE and is defined by
ξ ^ L L E = F d ξ ^ M L E ,
where F d = Υ + I r 1 Υ + d I r , and d 0 d < 1 is the LP. The BV and CM of (13) are given respectively by
B i a s ξ ^ L L E = d 1 Γ I 1 ξ .
C o v ξ ^ L L E = Γ I 1 Γ d Γ 1 Γ d Γ I 1 t .
By using (14) and (15), the MMSE is defined as
M M S E ξ ^ L L E = F d Υ 1 F d t + B i a s ξ ^ L L E B i a s ξ ^ L L E t    = Γ I 1 Γ d Γ 1 Γ d Γ I 1 t + d 1 2 Γ I 1 ξ ξ t Γ I 1 t ,
where Γ I = d i a g μ 1 + 1 ,   μ 2 + 1 ,   , μ r + 1 , and Γ d = d i a g μ 1 + d ,   μ 2 + d ,   , μ r + d . The scalar MSE of the LLE is given by
M S E ξ ^ L L E = t r M M S E ξ ^ L L E = j = 1 r μ j + d 2 μ j μ j + 1 2 + d 1 2 j = 1 r τ j 2 μ j + 1 2 ,
where d is the LP. Optimizing the Equation (17) with respect to d yields d = j = 1 r τ j 2 1 / τ j 2 + 1 / μ j . For d = 1 , we obtain ξ ^ L L E = ξ ^ M L E .

2.1. The Proposed Estimator

In the LLE, most of the time, the value of d is negative, which affects the BLRM estimation under multicollinearity. This estimator has been proposed for the Poisson RM [27]. Therefore, following [27], we define the adjusted logistic Liu estimator (ALLE), which is given as
ξ ^ A L L E = A d   ξ ^ M L E ,
where A d = Υ + I r 1 Υ d 0 I r , and 0 < d 0 < 1 is the Liu parameter for the ALLE. The new estimator gives a genuine refinement in the efficacy of the BLRM coefficients. The BV, CM, MSE matrix and scalar MSE of the ALLE are given in Equations (19)–(22), respectively.
B i a s ξ ^ A L L E = d 0 1 Υ + I r 1 ξ .
C o v ξ ^ A L L E = Υ + I r 1 Υ d 0 I r   Υ 1 Υ d 0 I r Υ + I r 1 .
M M S E ξ ^ A L L E = C o v ξ ^ A L L E + B i a s ξ ^ A L L E B i a s ξ ^ A L L E t = Υ + I r 1 Υ d 0 I r   Υ 1 Υ d 0 I r Υ + I r 1 + d 0 1 2 Υ + I r 1 β β t Υ + I r 1 .
M S E ξ ^ A L L E = t r M M S E ξ ^ A L L E = j = 1 r μ j d 0 2 μ j μ j + 1 2 + d 0 1 2 j = 1 r τ j 2 μ j + 1 2 .

2.2. Theoretical Comparison

Proposition 2.1.
Let M > 0, i.e., M is a positive definite (p,d) matrix. Then M τ τ t 0 if and only if τ t M 1 τ 1 for some vector τ [28].
Proposition 2.2.
Let θ ^ 1 = B 1 y and θ ^ 2 = B 2 y be estimators of θ . Suppose that D = C o v ( θ ^ 1 ) C o v θ ^ 2 > 0 , where C o v ( θ ^ 1 ) and C o v θ ^ 2 represents the CM of θ ^ 1 and θ ^ 2 , respectively. Then, M M S E θ ^ 1 M M S E θ ^ 2 > 0 if and only if c 2 t D + c 2 c 2 t 1 c 2 < 1 , where c 2 is the bias whereas M M S E θ ^ j = C o v θ ^ j + c j c j t , where c j is the BV of θ ^ j [3].
Theorem 2.1.
For the BLRM, if 0 < d 0 < 1 , then ξ ^ A L L E is better than ξ ^ M L E ; that is, Δ 1 = M M S E ξ ^ M L E M M S E ξ ^ A L L E > 0 if and only if b A L L E t Υ 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 1 b A L L E < 1 , where b A L L E = d 0 1 Υ + I r 1 ξ .
Proof: 
See Appendix A.  □
Theorem 2.2.
Under the BLRM, if k > 0 and 0 < d 0 < 1 , then ξ ^ A L L E is better than ξ ^ L R R E ; that is, Δ 2 = M M S E ξ ^ L R R E M M S E ξ ^ A L L E > 0 if and only if b A L L E t Υ + k I r 1 Υ Υ + k I r 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 b A L L E < 1 , where b A L L E = d 0 1 Υ + I r 1 ξ .
Proof: 
See Appendix A.  □
Theorem 2.3.
Under the BLRM, if 0 d < 1 and 0 < d 0 < 1 , then ξ ^ A L L E is better than ξ ^ L L E ; that is, Δ 3 = M M S E ξ ^ L L E M M S E ξ ^ A L L E > 0 if and only if b A L L E t Υ + I r 1 Υ + d I r   Υ 1 Υ + d I r Υ + I r 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 b A L L E < 1 , where b A L L E = d 0 1 Υ + I r 1 ξ .
Proof: 
See Appendix A.  □

2.3. Parameter Estimation

In the literature different methods are employed to optimize the biasing parameters; see [17,20,29,30]. Computing the critical points of the Equation (22) with respect to d 0 , we obtain the jth term as
d 0 j = μ j 1 + τ j 2 1 + τ j 2 μ j .
In (23), replacing μ ^ j and τ ^ j by their unbiased estimators, we obtain
d ^ 0 j = μ ^ j 1 + τ ^ j 2 1 + τ ^ j 2 μ ^ j .
For practical considerations, we consider the minimum value of (24) as
d ^ 0 m i n = m i n μ ^ j 1 + τ ^ j 2 1 + τ ^ j 2 μ ^ j .

3. Simulation

3.1. Design

Here the simulation study is designed, and general layout is given. The comparison criteria is the simulated MSE of the estimators, i.e., MLE, LRRE, LLE and ALLE, when the regressors are multicollinear. Considering z i j to be the independent standard normal pseudo-random numbers and following [2,14,23], the EVs with various levels of correlation can be generated via the following expression:
x i j = 1 ρ 2 1 / 2 z i j + ρ z i j + 1 ,   i = 1 ,   ,   n ; j = 1 ,   ,   r   ,
where ρ is select so as ρ 2 be the correlation between any two explanatories. For simulation, we select four different values of ρ : 0.80; 0.90; 0.95; 0.99. Four different sample size are considered n :   50 ;   100 ;   200 ; 400. Moreover, we also consider different number of EVs p : 3 ; 6 ; 12 , and β is selected to be the eigenvector of the matrix F corresponding to the largest eigenvalue [14].
The Bernoulli distribution B e δ i is used to generate observations, where δ i = e x p x i t β 1 + e x p x i t β ,   i = 1 , , n such that the data matrix X = ( x i t ) i = 1 , 2 , , n . The experiment is repeated 1000 times by generating z i j . The bias and MSE values of the estimators are computed via the equation:
B i a s   β ^ = i = 1 R ξ ^ i ξ R ; M S E β ^ = i = 1 R ξ ^ i ξ t ξ ^ i ξ R .
where ξ ^ i ξ is the deviation between the parameter and its estimate for each i of the BLRM estimators and R is the number replications. The R Language is used for computations.

3.2. Discussion

The estimated biases and MSEs of the BLRM estimators for changed controlled conditions are summarized in Table 1, Table 2 and Table 3 for p = 3, 6 and 12, respectively. The simulated results in Table 1, Table 2 and Table 3 clearly demonstrated that the constructed ALLE surpassed the other estimators in terms of minimum bias and MSE. In all cases, the performance of the MLE is the worst since the estimated MSE is larger as compared to the other estimators. Based on the findings of the simulation, it can be seen that there is a direct relationship between the scalar MSEs and various levels of multicollinearity. As we change the multicollinearity level from mild to severe, the estimated MSEs increase for a given sample size and number of explanatory variables (EV). However, it is clearly observed that the proposed ALLE for the BLRM is a unique estimation method due to its minimum bias as well as minimum estimated MSEs. We also provide a graphical representation for the readers to see a clear image of the constructed and other considered estimators. Figure 1a–d demonstrates the effect of collinearity on the performance of the under-studied estimators with different sizes. Figure 1 clearly shows that the proposed ALLE performed the best for all the levels of multicollinearity. In addition, when the results are compared with respect to the sample size, then it can be noticed that the estimated MSE values of all the estimators decrease as n increases. For more details, see Figure 2a–d. When the results are interpreted with regard to the number of EVs, then it is observed from Table 1, Table 2 and Table 3 that the estimated MSEs increase by increasing the value of p. Based on the results, ALLE is the most suitable and consistent option whenever the RV is binary, and the EVs are correlated due to its minimum bias and estimated MSEs. Undoubtedly, other biased estimators also attain a lower MSE in contrast to the MLE, but the variation in the proposed ALLE is quite lower than the other estimators for all evaluated states.

4. Application

Here, a real application is considered to discuss the performance of the constructed estimator. For this purpose, prostate cancer data have been considered, which were taken from Kutner et al. [31]. However, the RV y is the seminal vesicle invasion, i.e., presence or absence of seminal vesicle invasion: the binary RV is 1 if yes; 0 otherwise. The descriptions of seven regressors are given in Table 4.
The eigenvalues of F matrix are found to be: μ 1 = 40,669.86 , μ 2 = 1960.62 , μ 3 = 1868.34 , μ 4 = 176.214 , μ 5 = 45.520 , μ 6 = 18.943 and μ 7 = 27.4104 . The multicollinearity of the regressors is assessed via a condition index (CI). The CI is evaluated by the eigenvalues of the X t V ^ X matrix without an intercept term, while the eigenvalues of the F matrix include the intercept term. We observed that the C I = μ m a x / μ m i n = 121.97 > 30 . This confirms that there is a multicollinearity issue among the regressors. Moreover, we also compute the correlations among the regressors, and the results are displayed in Figure 3. It declares the moderate correlation among CV and PSA. level, CP and PSA. level and CV and CP.
The estimated coefficients (EC) and the MSE values are presented in Table 5. The EC of the MLE, LRRE, LLE and ALLE are obtained from (4), (7), (14) and (19), respectively. However, the scalar MSE values of the MLE, LRRE, LLE and ALLE are calculated using equations (6), (12), (18) and (23), respectively. The LRRE involves a shrinkage ridge parameter k , which is found to be 0.077. Whereas the Liu parameter d for the LLE was found to be 0.5689. However, the shrinkage parameter d 0 of the proposed ALLE was found to be 0.0693. Table 5 reveals that ALLE decreases the EC and MSE value in a better way in comparison with the competitive estimators. It is also noticed that the MLE attains the largest MSE, which clearly indicates that it is the most sensitive estimator when there exists a multicollinearity among the regressors. In addition, the proposed ALLE is the most consistent option in the case of multicollinearity. Further, it can be noticed that the performance of LRRE is comparatively better than that of LLE and MLE. Further, it is also observed that the ALLE’s performance is better as compared to the other biased estimators as well as the MLE. Therefore, it is recommended to use the ALLE estimator when one considers the BLRM with correlated regressors because of its consistent and robust behavior against multicollinearity.

5. Concluding Remarks

We constructed a new adjusted logistic LE (ALLE) for the binary LRM (BLRM) to handle the multicollinearity issue. The MLE method is not a good choice due to its inflated variance and standard error whenever the regressors are multicollinear. The efficacy of the constructed estimator is judged via a MCS under various controlled conditions. The investigation has been performed for different values of correlation, sample sizes and the number of EVs. Findings showed that the ALLE’s performance is better compared to competitive ones and the MLE. Further, the MSE values of the MLE, LRRE, LLE and ALLE increase as the multicollinearity level increases. This phenomenon is particularly sharp for small sample sizes and whenever the correlation is high. However, this increase is quite smaller for the proposed ALLE. The superiority of the proposed estimator over others is also proved via a real application. The findings from both simulation and application clearly support our constructed estimator, which is a good and robust estimation method whenever there exists an imperfect but high multicollinearity among the regressors. As a result, we strongly advise practitioners to use the new estimator when estimating the unknown BLRM regression coefficients in the presence of multicollinearity.

Author Contributions

Authors contributed equally. All authors have read and agreed to the published version of the manuscript.

Funding

The author Nahid Fatima would like to acknowledge the support of Prince Sultan University for its support and help for paying the Article Processing Charges (APC) of this publication. Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R 299), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Data Availability Statement

The data is included with in the article.

Acknowledgments

The author Nahid Fatima would like to acknowledge the support of Prince Sultan University for its support and help for paying the Article Processing Charges (APC) of this publication. The authors are thankful to Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R 299), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

There is no conflict of interest.

Appendix A

Proof of Theorem 2.1
Δ 1 = Υ 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 b A L L E b A L L E t .
= Υ 1 I r Υ + I r 1 Υ d 0 I r Υ d 0 I r Υ + I r 1 b A L L E b A L L E t .
However, the scalar MSE of (24) can be written as
M S E ξ ^ M L E M S E ξ ^ A L L E =   d i a g 1 μ j μ j d 0 2 μ j μ j + 1 2 j = 1 r t b A L L E t b A L L E =   d i a g 2 μ j d 0 + 1 d 0 2 1 μ j μ j + 1 2 j = 1 r t b A L L E t b A L L E ,
where b A L L E is the bias of our proposed estimation method. We observed from Theorem 2.1 that Υ 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 is positive definite. Υ 1 I r Υ + I r 1 Υ d 0 I r Υ d 0 I r Υ + I r 1 is positive definite if 2 μ j d 0 + 1 d 0 2 1 > 0     j = 1 , , r . Thus, by using proposition 2.1 and 2.2, we conclude that M S E ξ ^ M L E > M S E ξ ^ A L L E ; the proof is ended by proposition 2.1 and 2.2.  □
Proof of Theorem 2.2
Δ 2 = Υ + k I r 1 Υ Υ + k I r 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 + b L R R E b L R R E t b A L L E b A L L E t ,
where b L R R E = k Υ + k I r 1 . While the difference in scalar MSEs of (26) can be written as
M S E ξ ^ L R R E M S E ξ ^ A L L E =   d i a g μ j μ j + k 2 μ j d 0 2 μ j μ j + 1 2 j = 1 r t + b L R R E t b L R R E b A L L E t b A L L E
=   d i a g μ j 2 μ j + 1 2 μ j d 0 2 μ j + k 2 μ j μ j + k 2 μ j + 1 2 j = 1 r t + b L R R E t b L R R E b A L L E t b A L L E ,
where b L R R E and b A L L E is the bias of LRRE and our proposed (ALLE) estimation method. We observed from Theorem 2.2 that Υ + k I r 1 Υ Υ + k I r 1 Υ + I r 1 Υ d 0 I r Υ 1 Υ d 0 I r Υ + I r 1 is positive definite if μ j 2 μ j + 1 2 μ j d 0 2 μ j + k 2 > 0     j = 1 , , r . Thus, by using proposition 2.1 and 2.2, we conclude that M S E ξ ^ L R R E > M S E ξ ^ A L L E ; the proof is ended by proposition 2.1 and 2.2.  □

References

  1. Lin, E.; Lin, C.-H.; Lane, H.-Y. Logistic ridge regression to predict bipolar disorder using mRNA expression levels in the N-methyl-D-aspartate receptor genes. J. Affect. Disord. 2022, 297, 309–313. [Google Scholar] [CrossRef] [PubMed]
  2. Afzal, N.; Amanullah, M. Dawoud–Kibria Estimator for the Logistic Regression Model: Method, Simulation and Application. Iran. J. Sci. Tech. Trans. A Sci. 2022, 46, 1483–1493. [Google Scholar] [CrossRef]
  3. Schaefer, R.L. Alternative estimators in logistic regression when the data are collinear. J. Stat. Comput. Simul. 1986, 25, 75–91. [Google Scholar] [CrossRef]
  4. Lukman, A.F.; Ayinde, K. Review and classifications of the ridge parameter estimation techniques. Hacet. J. Math. Stat. 2017, 46, 953–967. [Google Scholar] [CrossRef]
  5. Frisch, R. Statistical confluence analysis by means of complete regression systems. Econ. J. 1934, 45, 741–742. [Google Scholar]
  6. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 1970, 12, 55–67. [Google Scholar] [CrossRef]
  7. Schaefer, R.L.; Roi, L.D.; Wolfe, R.A. A ridge logistic estimator. Commun. Stat.—Theor. Methods 1984, 13, 99–113. [Google Scholar] [CrossRef]
  8. Lee, A.H.; Silvapulle, M.J. Ridge estimation in logistic regression. Commun. Stat. Simul. Comput. 1988, 17, 1231–1257. [Google Scholar] [CrossRef]
  9. Cessie, S.L.; Houwelingen, J.C.V. Ridge estimators in logistic regression. J. R. Stat. Soc. C 1992, 41, 191–201. [Google Scholar] [CrossRef]
  10. Segerstedt, B. On ordinary ridge regression in generalized linear models. Commun. Stat. Theor. Method. 1992, 21, 2227–2246. [Google Scholar] [CrossRef]
  11. Akram, M.N.; Amin, M.; Elhassanein, A.; Aman Ullah, M. A new modified ridge-type estimator for the beta regression model: Simulation and application. AIMS Math. 2022, 7, 1035–1057. [Google Scholar] [CrossRef]
  12. Kibria, B.M.G.; Månsson, K.; Shukur, G. Performance of some logistic ridge regression estimators. Comput. Econ. 2012, 40, 401–414. [Google Scholar] [CrossRef]
  13. Asar, Y. Some new methods to solve multicollinearity in logistic regression. Commun. Stat. Simul. Comput. 2017, 46, 2576–2586. [Google Scholar] [CrossRef]
  14. Hadia, M.; Amin, M.; Akram, M.N. Comparison of link functions for the estimation of logistic ridge regression: An application to urine data. Commun. Stat. Simul. Comput. 2022. [Google Scholar] [CrossRef]
  15. Liu, K. A new class of biased estimate in linear regression. Commun. Stat. 1993, 22, 393–402. [Google Scholar]
  16. Mansson, K.; Kibria, B.M.G.; Shukur, G. On Liu estimators for the logit regression model. Econ. Model. 2012, 29, 1483–1488. [Google Scholar] [CrossRef]
  17. Qasim, M.; Amin, M.; Amanullah, M. On the performance of some new Liu parameters for the gamma regression model. J. Stat. Comput. Simul. 2018, 88, 3065–3080. [Google Scholar] [CrossRef]
  18. İnan, D.; Erdoğan, B.E. Liu-type logistic estimator. Commun. Stat. Simul. Comput. 2013, 42, 1578–1586. [Google Scholar] [CrossRef]
  19. Şiray, G.Ü.; Toker, S.; Kaçiranlar, S. On the restricted Liu estimator in the logistic regression model. Commun. Stat. Simul. Comput. 2015, 44, 217–232. [Google Scholar] [CrossRef]
  20. Asar, Y.; Genc, A. New shrinkage parameters for the Liu-type logistic estimators. Commun. Stat. Simul. Comput. 2016, 45, 1094–1103. [Google Scholar] [CrossRef]
  21. Wu, J. Modified restricted Liu estimator in logistic regression model. Comput. Stat. 2016, 31, 1557–1567. [Google Scholar] [CrossRef]
  22. Wu, J.; Asar, Y. More on the restricted Liu estimator in the logistic regression model. Commun. Statist. Simul. Comput. 2017, 46, 3680–3689. [Google Scholar] [CrossRef]
  23. Lukman, A.F.; Kibria, B.M.G.; Ayinde, K.; Jegede, S.L. Modified one-parameter Liu estimator for the linear regression model. Model. Simul. Eng. 2020, 2020, 9574304. [Google Scholar] [CrossRef]
  24. Wu, J.; Asar, Y.; Arashi, M. On the restricted almost unbiased Liu estimator in the logistic regression model. Commun. Stat. Theor. Method. 2018, 47, 4389–4401. [Google Scholar] [CrossRef]
  25. Varathan, N.; Wijekoon, P. Logistic Liu estimator under stochastic linear restrictions. Stat. Pap. 2019, 60, 595–612. [Google Scholar] [CrossRef]
  26. Pregibon, D. Logistic regression diagnostics. Ann. Stat. 1981, 9, 705–724. [Google Scholar] [CrossRef]
  27. Li, Y.; Asar, Y.; Wu, J. On the stochastic restricted Liu estimator in logistic regression model. J. Stat. Comput. Simul. 2020, 90, 2766–2788. [Google Scholar] [CrossRef]
  28. Amin, M.; Akram, M.N.; Kibria, B.M.G. A new adjusted Liu estimator for the Poisson regression model. Concurr. Comput. Pract. Exper. 2021, 33, e6340. [Google Scholar] [CrossRef]
  29. Farebrother, R.W. Further results on the mean square error of ridge regression. J. R. Stat. Soc. 1976, 38, 248–250. [Google Scholar]
  30. Mustafa, S.; Amin, M.; Akram, M.N.; Afzal, N. On the performance of link functions in the beta ridge regression model: Simulation and application. Concurr. Comput. Pract. Exper. 2022, 34, e7005. [Google Scholar] [CrossRef]
  31. Kutner, M.H.; Nachtsheim, C.J.; Neter, J.; Li, W. Applied Linear Statistical Models, 5th ed.; McGraw Hill: New York, NY, USA, 2005. [Google Scholar]
Figure 1. The BLRM estimators in relation to multicollinearity: (a). n = 50 , (b). n = 100 , (c). n = 200 , (d). n = 400 .
Figure 1. The BLRM estimators in relation to multicollinearity: (a). n = 50 , (b). n = 100 , (c). n = 200 , (d). n = 400 .
Axioms 12 00175 g001
Figure 2. The BLRM estimators for various n : (a). ρ = 0.80 , (b). ρ = 0.90 , (c). ρ = 0.95 , (d). ρ = 0.99 .
Figure 2. The BLRM estimators for various n : (a). ρ = 0.80 , (b). ρ = 0.90 , (c). ρ = 0.95 , (d). ρ = 0.99 .
Axioms 12 00175 g002
Figure 3. Correlation matrix among the seven regressors of a prostate cancer data.
Figure 3. Correlation matrix among the seven regressors of a prostate cancer data.
Axioms 12 00175 g003
Table 1. Bias and MSE values of the estimators for p = 3.
Table 1. Bias and MSE values of the estimators for p = 3.
BiasMSE
n ρ LRRELLUALLEMLELRRELLUALLE
500.80−0.1164−0.0706−0.0181297.24158.62109.1122.15
0.900.11040.0454−0.0059495.09255.09187.0140.57
0.95−0.0601−0.0047−0.0572970.93498.44393.8895.04
0.990.21720.1165−0.05854195.401971.661634.97431.22
1000.800.22450.1493−0.0182141.1676.7850.449.45
0.900.12070.0485−0.0047224.26115.7780.5616.53
0.950.12880.0948−0.0192410.35202.71151.4032.77
0.99−0.0330−0.0106−0.01781858.28873.64698.43166.81
2000.800.03480.0297−0.020165.6835.7923.815.17
0.900.15450.0914−0.0048105.2754.2238.299.97
0.950.12350.05830.0107194.0595.7972.4419.80
0.990.20430.08720.0328925.09423.63348.39118.45
4000.800.05840.03940.006730.4716.6510.462.32
0.900.21810.13820.056652.0927.3417.834.48
0.950.10360.06890.018391.9547.2130.587.20
0.990.06940.03410.0212435.24213.83142.5137.29
Table 2. Bias and MSE values of the estimators for p = 6.
Table 2. Bias and MSE values of the estimators for p = 6.
BiasMSE
n ρ LRRELLUALLEMLELRRELLUALLE
500.800.28890.1538−0.0200983.86547.58350.6221.41
0.900.02480.0232−0.03431820.37982.09656.3244.56
0.950.25480.14220.00293505.321879.781249.9967.30
0.990.43580.2714−0.015417,361.149256.686295.43407.18
1000.800.16480.1086−0.0137349.73200.38114.256.74
0.900.19270.09480.0303617.50339.62195.7812.17
0.950.24780.15290.03611224.53666.51392.0124.99
0.990.04270.01110.01005899.933172.731845.47120.22
2000.800.28580.20140.0484152.7489.5148.153.12
0.900.31550.19700.0865273.10155.0484.304.79
0.950.33090.21620.0968516.23285.45156.098.09
0.990.23040.15570.07492529.831393.39756.0237.58
4000.800.17620.12650.071973.5843.0722.742.16
0.900.19340.14100.0957135.4676.2940.173.36
0.950.22360.16250.1109252.20139.1572.365.01
0.990.24500.18070.12921203.21652.75330.8918.81
Table 3. Bias and MSE values of the estimators for p = 12.
Table 3. Bias and MSE values of the estimators for p = 12.
BiasMSE
n ρ LRRELLUALLEMLELRRELLUALLE
500.800.69090.5045−0.00523448.201847.271398.0117.48
0.900.54280.34540.00741,049,972.883275.032453.7429.13
0.950.61660.45750.0142884,816.476229.804808.4663.91
0.990.40640.2876−0.01992,162,403.6830,159.2323,554.65590.93
1000.800.49060.32610.0717818.31495.73265.913.52
0.900.25660.16810.04801447.22861.72454.045.08
0.950.31700.20550.06752849.431697.90906.368.21
0.990.09220.05870.013113,816.378172.864350.7330.44
2000.800.27170.19040.1051322.54200.0094.283.01
0.900.16940.11850.0773586.96357.99169.713.78
0.950.30370.23060.14611144.84698.86330.324.79
0.990.24960.19410.12765603.543422.061606.9216.26
4000.800.36490.29260.2101143.9391.7041.902.77
0.900.29280.23940.1839257.25160.6471.302.82
0.950.28590.23500.1873498.05309.52136.673.23
0.990.28740.24430.19252399.741479.83647.966.25
Table 4. Explanation of the respected regressors.
Table 4. Explanation of the respected regressors.
NRegressor CallExplanation
1PSA levelSerum prostate- specific antigen level (mg/mL).
2Cancer volume (CV)Estimate of prostate cancer volume (cc).
3WeightProstate weight (gm)
4AgeAge of patients (years)
5Benign prostatic hyperplasia (BPH)Amount of benign prostatic hyperplasia (cm2)
6Capsular penetration (CP)Degree of capsular penetration (cm)
7Gleason score (GS)Pathologically determined grade of disease using total score of two patterns (summed scores were either 6, 7, or 8 with higher scores indicating more prognosis).
Source: Adapted in part from: Hastie, T. J.; R. J. Tibshirani; and J. Friedman. The Elements of Statistical Learning: Data Mining, Inference, and Prediction, New York; Springer. Verlag, 2001.
Table 5. Regression estimates and MSEs of the logistics regression estimators for the prostate cancer data.
Table 5. Regression estimates and MSEs of the logistics regression estimators for the prostate cancer data.
MLELRRELLEALLE
(Intercept)−10.1574−2.3825−5.82660.4431
x10.11890.10900.11340.0985
x2−0.1345−0.1146−0.1234−0.0915
x30.00010.00020.0001−0.0001
x40.09730.02200.0553−0.0169
x5−0.2281−0.2185−0.2228−0.2021
x60.64870.65130.65010.5976
x7−0.0673−0.4491−0.2799−0.4555
MSE45.68734.056915.90752.2935
Note: The best value is in bold font.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Amin, M.; Akram, M.N.; Kibria, B.M.G.; Alshanbari, H.M.; Fatima, N.; Elhassanein, A. On the Estimation of the Binary Response Model. Axioms 2023, 12, 175. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms12020175

AMA Style

Amin M, Akram MN, Kibria BMG, Alshanbari HM, Fatima N, Elhassanein A. On the Estimation of the Binary Response Model. Axioms. 2023; 12(2):175. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms12020175

Chicago/Turabian Style

Amin, Muhammad, Muhammad Nauman Akram, B. M. Golam Kibria, Huda M. Alshanbari, Nahid Fatima, and Ahmed Elhassanein. 2023. "On the Estimation of the Binary Response Model" Axioms 12, no. 2: 175. https://0-doi-org.brum.beds.ac.uk/10.3390/axioms12020175

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop