Scholarly article on topic 'Novel Harmonic Regularization Approach for Variable Selection in Cox’s Proportional Hazards Model'

Novel Harmonic Regularization Approach for Variable Selection in Cox’s Proportional Hazards Model Academic research paper on "Mathematics"

Share paper
OECD Field of science

Academic research paper on topic "Novel Harmonic Regularization Approach for Variable Selection in Cox’s Proportional Hazards Model"

Hindawi Publishing Corporation

Computational and Mathematical Methods in Medicine Volume 2014, Article ID 857398, 9 pages

Research Article

Novel Harmonic Regularization Approach for Variable Selection in Cox's Proportional Hazards Model

Ge-Jin Chu, Yong Liang, and Jia-Xuan Wang

University Hospital, State Key Laboratory of Quality Research in Chinese Medicines, Faculty of Information Technology, Macau University of Science and Technology, Macau

Correspondence should be addressed to Yong Liang;

Received 23 April 2014; Revised 13 July 2014; Accepted 25 July 2014; Published 24 November 2014

Academic Editor: Andrzej Kloczkowski

Copyright © 2014 Ge-Jin Chu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Variable selection is an important issue in regression and a number of variable selection methods have been proposed involving nonconvex penalty functions. In this paper, we investigate a novel harmonic regularization method, which can approximate nonconvex Lq (1/2 < q < 1) regularizations, to select key risk factors in the Cox's proportional hazards model using micro-array gene expression data. The harmonic regularization method can be efficiently solved using our proposed direct path seeking approach, which can produce solutions that closely approximate those for the convex loss function and the nonconvex regularization. Simulation results based on the artificial datasets and four real microarray gene expression datasets, such as real diffuse large B-cell lymphoma (DCBCL), the lung cancer, and the AML datasets, show that the harmonic regularization method can be more accurate for variable selection than existing Lasso series methods.

1. Introduction

One of the most important objectives for survival analysis is to select a small number of key risk factors from many potential predictors [1]. Commonly, the Cox proportional hazards model [2, 3] is used to study the relationship between predictor variables and survival time. Suppose a dataset has a sample size of n to study the survival time T on covariate x; we use the data form of (t1, S1t x1),..., (tn, Sn, xn) to represent the individual's sample, where the survival time tt being complete if St = 1 and right censored if St = 0. As in regression, xt = (xi1,xi2,...,xip) is a potential prediction vector.

By Cox's proportional hazards model, the hazard function is given as

h(t\p) = h0 (t) exp (xTp), (1)

where the baseline hazard function h0(t) is unspecified and ß = (ß1,ß2,..., ßp) is the regression coefficient vector of p variables. Cox's partial log-likelihood is expressed as

Kß) = !st

Xlß - log ( T exP (xfß)

where Rt denotes the set of indices of the survival individuals at time tt.

In practice, only a small number of the predictor variables actually affect the hazard rate. The goal of variable selection in Cox's proportional hazards model is to select the key risk factors. Recently a series of penalized partial likelihood methods, such as the L1 [4-7], Lq (0 < q < 1) [8] and L1/2 [9, 10] penalized methods were proposed for Cox's proportional hazards model. These penalized partial likelihood methods find important risk factors by shrinking some regression coefficients to zero.

The standard penalized methods cannot directly be applied to the nonlinear Cox model to obtain parameter estimates. Therefore, Tibshirani [11] proposed an iterative procedure to transform the Cox's partial log-likelihood function (2) to linear regression problem. Let x = (xilt.. - ,xip) , i = 1,.. ,,n, and denote the p xn predictor variable matrix, n = XTI3, p = -(dl/dt]), A = -(d2l/d^d^T), and z = q + A-^, where A- is a generalized inverse of A. Since the general quadratic programming cannot be directly solved to the cases with p » n, Gui and Li [12] applied the Choleski decomposition to obtain C = A1/2 such that CTC = A, y = Cz, and x = Cx. By the Taylor expansion, the partial log-likelihood l(p) is approximated by the quadratic form:

[y - xTpf [y - xTp) .

Thus, the regularization methods can directly solve the penalized regression problem:

also investigate the fast harmonic regularization algorithm to solve the Cox model for the high dimension low sample size problem ("large p small n problem").

The rest of the paper is organized as follows. Section 2 describes the harmonic regularization method. Section 3 gives a harmonic regularization algorithm to obtain estimates form Cox model. Section 4 evaluates our method by simulation studies and application to four real microarray datasets, such as the diffuse large B-cell lymphoma (DLBCL) datasets with the survival times and gene expression data. Section 5 concludes the paper with some useful remarks.

2. Harmonic Regularization

In general, a united framework of the regularization in machine learning has a form:

ft (X) = arg min [R (ft) + XP (ft)],

ft = arg min ( ||y - xT ftH + X^ P(ftj)

where X is the tuning parameter.

Tibshirani [5] proposed the Lasso (least absolute shrinkage and selection operator) method, which has L1 penalty P(ftj) = \ftj|, which shrinks small coefficients to zero and hence results in a sparse representation of the solution. Fan and Li [4] proposed the smoothly clipped absolute deviation (SCAD) penalty, which avoids excessive penalties on large coefficients and enjoys the oracle properties. Zhang [7] proposed the minimax concave plus (MCP) method, which is a continuous and nearly unbiased approach in high-dimensional linear regression. Zhang and Lu [6] suggested an adaptive Lassomethodwithanadaptively L1 penalty estimate the parameters, which uses the penalty P(ftj) = \ftj\l\ft'j\, where the weights 1/\ft'j\ are chosen adaptively by the data. Zou and Hastie [13] proposed an elastic net method that combines the L1 and L2 (P(ftj) = \ftj\2) penalties.

The above mentioned series of regularized regression methods were based on the L1 penalty. Recently, several works on learning sparse models have stressed the need of other penalties for achieving better sparsity profile. For instance, Rosset and Zhu [14] suggested the use of a Lq penalty, which simply consists in replacing the L1 norm with nonconvex Lq norm (0 < q < 1). Zhang [15] presented a multistage convex relaxation scheme, which can be relaxed to a smoothed Lq regularization. Mazumder et al. [16] pursued a coordinate-descent approach with nonconvex penalties (SparseNet) and study its convergence properties. Xu et al. [9, 10] further explored the properties of the Lq (0 < q < 1) penalty and revealed the extreme importance and special role of the L1/2 regularization. In our previous work [17,18], we developed several fast algorithms using the L1/2 penalty to solve the logistic regression model and the Cox model. Our computational results showed that L1/2 regularization outperforms some L1 regularization methods. In this paper, we propose a novel harmonic regularization method which approximates to the Lq (1/2 < q < 1) penalties. We

where R(p) is a loss function, P(p) is a penalty function, and A is a tuning parameter. Different A here is in correspondence with different penalized constraint to the model, so different solution is to be got, respectively. The penalized constraint is the weakest when X = 0 and becomes stronger as X increases.

Obviously a regularization (5) can be divided by two elements, the loss function R(p) and the penalty function P(fi). Moreover, different loss function and different penalty will result in different algorithm. For example, when the loss function is hinge loss and the penalty P(p) = \\p\\ , the result is a support vector machine algorithm. Let the loss function be square loss and using P(p) = \\p\\q denote the Lq regularization methods, if q = 2, it is the ridge regression [19] and can be used to solve the ill-posed problem. If q = 0, it is the subsets regression [20], which applies L0 regularization with the penalty function P(p) = (1/2)7(|^=0|).When q = 1,it is the Lasso algorithm [21], which applied L1 regularization. Lasso and its variations (or the Lasso type algorithms), such as elastic net [13], SCAD [4], MCP [7], adaptive Lasso [6], and stage-wise Lasso [22] are extensively studied and applied in recent years in the fields of statistics and machine learning.

It is well known that L0 regularization is ideal sparsest for variable selection. Unfortunately, L0 regularization is a combinatorial optimization problem, which is difficult to be solved. In contrast, L1 regularization leads to a convex optimization problem and easy to be solved, but it does not yields sufficiently sparse variable selection. Donoho et al. [23, 24] had shown that L0 regularization is equivalent to L1 regularization under certain conditions. These imposed conditions therefore characterize those problems for which no matter what L1 or L0 regularization is applied, the same sparse solutions will be produced. However, for many practical problems, the sparsity of solutions yielded through L1 and L0 regularization is far from being equivalent. Particularly, the solutions found with L1 regularization is very often less as sparse as the solutions found with L0 regularization.

In fact, when 0 < q < 1, the Lq regularization automatically performs variable selection by removing predictors with very small nonzero estimated coefficients. The smaller

the q is, the sparser the solutions found with Lq regularization will be. This leads researchers to study Lq regularization with 0 < q < 1 because it can find the more sparse solutions than those found with L1 regularization and easier to be solved than L0 regularization. For example, Zhang [15] presented a multistage convex relaxation scheme for solving problems with nonconvex objective functions. For learning formulations with sparse regularization, they analyzed the behavior of a specific multistage relaxation scheme.

Nevertheless, the applications to the Lq penalty function with 0 < q < 1 not often attracts much attention done mainly due to the reason that when 0 < q < 1, the penalty function changes from a convex function to a nonconvex one and so the corresponding optimization problem is not easy to solve. Meanwhile, another difficulty in fact is that the differential quotient of the penalty function at origin is which results in the invalidation of the ordinarily optimization algorithms.

In this paper, we propose the harmonic regularization which can approximate the Lq penalty with 1/2 < q < 1, because some research works show that the L1/2 penalty can be taken as a representative of the Lq (0 < q< 1)penalty[22]. The harmonic regularization scheme can be expressed as

P = arg

a (a- 1)

2 - a a - 1

2 - a a - 1

where 1 < a < 2. When the shrinkage parameter a is

close to 1, -((2-a)/(a-

1)) = \p\, the harmonic regularization approximates to the L1 regularization. When the parameter a is close to 2,

^^m^-1i)2-{{2-a)/{a-1)) « VW

and the harmonic regularization approximates to the L1/2 regularization. Moreover, comparing with the Lq (1/2 < q < 1) penalties, the harmonic regularization has the property that its first derivative is finite at origin, which implies that the corresponding regularization problem can be efficiently solved via the direct seeking techniques.

3. The Harmonic Regularization Algorithm for the Cox Model

In this section, we propose a generalized path seeking algorithm of the harmonic regularization for Cox's model. As mentioned in the last section, when Lq (0 < q < 1) regularization is to be applied, an inevitable difficulty is how to efficiently solve the nonconvex optimization problem caused by the Lq (0 < q < 1) regularization (It is easy to see that in the case of Lq (q > 1) regularization is applied, the penalty term becomes convex). Fortunately, direct path seeking makes it possible to overcome that difficulty. Direct path seeking, which sequentially constructs a path directly in the parameter space, closely approximates that for a penalty

function without having to repeatedly solve numerical optimization problems. Popular path seeking based on squared-error includes partial least squares regression (PLS, [23]), forward stepwise regression [22], least angle regression [25], piecewise linear path [14], and gradient boosting. Friedman [26] proposed the generalized path seeking, which can produce solutions that closely approximate those for any convex loss function and nonconvex constraints. The advantages of path seeking methods provide us a new way to solve the problem of regularization with nonconvex penalty. We will propose a new generalized path seeking method to solve the harmonic regularization.

We let

p(p) = lp} №)•

where Pjtfj) = 2/a(a-1))\pj\ + ((2-a)/(a-1))2 - (2-a)/(a- 1). Note that

dPj(Pj) 1

d\fij\ ^2a(a-1) (v)\+a2(a-2)'

which shows that each additive term Pj(fij) is a monotonically increasing function of absolute value of its argument. This implies that the net regularization penalty function we have suggested meets the validity of the general path seeking algorithm [11]. Let v measure length along the path and Av > 0 a small increment. Define

0i (v) = -

=p(v) i=1

pj (v) = -

dPj (Pj)

=«V) 1

and let

r, (v) =

0j (v) Pj (v)

Then, we give the harmonic regularization algorithm procedures for the Cox model as follows:

(1) initialize v =0,k = 0, [pk}(v = 0) = 0}pj=1;

(2) compute x and y based on (3) using the current value pk(v),j=1,...,p;

(3) loop {

(4) compute t}(v) = gj(v)/pj(v), j = 1,...,p;

(5) S={j\Tj(v)xpk(v)<0,j = 1,...,p};

(6) if (S = empty) j* = arg maxj\Tj(v)\;

(7) else j * = arg maxj£S\Tj(v)|;

(8) ft), (v + Av) = ft), (v) + Av x sign(r;-. (v));

(9) ftk(v + Av) = ftk(v), j = 1,..., n and j = j*;

(10) V ^ V + Av;

(11) } until Tj(v) = 0, j = 1,..., p;

(12) fc ^ 1, and go back to Step 2 until the convergence criterion is met.

In the above algorithm, after Step 2, at each step those coefficients ft^v) with sign opposite to that of the corresponding Tj(v) are identified. When the set S is empty, the coefficient corresponding to the largest component of |tj(v)}"=1, an absolute value is selected at Step 6. And when there are one or more elements in the set S, the coefficient with corresponding largest Tj(v) within this subset is instead selected. The selected coefficient is the increments by a small amount in the direction of the sign of its corresponding r* (v) while all other coefficients remain unchanged, yielding the solution for the next path point v+ Av. The iterations continue until all components of Tj(v) = 0 and the algorithm then reaches a regularized solution for the harmonic regularized Cox model.

4. Simulation

4.1. Selection of the Shrinkage Parameter a and the Tuning Parameter A. To select the shrinkage parameter a and the tuning parameter A, we use the maximization of the cross-validation partial likelihood (CVPL) method proposed by van Houwelingen et al. [27], which is defined as

CVPL (fl) A) = -- X {Z (ftH) («> A)) - /(_,) (ftH) («, A))},

where ft(—) (a, A) represents the estimation of ft based on the harmonic regularization procedure with the parameters a and A from the data without the (th subject. The terms /(ft) and l(_j)(ft) are the log partial likelihoods with all the subjects and without the (th subject, respectively. The optimal value of the parameters a and A are chosen to maximize the sum of the contributions of each subject to the log partial likelihood over a grid of (a, A). CVPL is the special case of a more general cross-validated likelihood approach for model selection and has been demonstrated to perform well in prediction in the context of the penalized Cox regression.

4.2. Model Validation Measures. The performance measures of censored survival data is more complicated: the measure can only be computed if the case is not right censoring. Thus, several specially designed measure method have been proposed in the literatures. In this paper, we employ the integrated brier score (IBS) [28] and the concordance index (CI)

[29] to evaluate the prediction ability of the regularization methods.

Integrated Brier Score (IBS). The brier score (BS) is defined as a function of time t > 0 by

BS (i) = -X

n L—'

S(f|X,)21(Í, <fAg, = 1)

+ (1-S(f |X,))2l(f, >Q + G(i)

where G(-) denotes the Kaplan-Meier estimation of the censoring distribution and S(- \ X;) stands to estimate survival for patient «. Note that the BS(i) is dependent on the point in time i, and its values are between 0 and 1. Good predictions at time t result in small values of BS. The integrated brier score (IBS) is given by

- rmax(tj)

I BS (í)dí.

t(í,-) Jo

max (i;)

The IBS is used to assess the goodness of the predicted survival functions of all observations at every time between 0 and max(ij).

Concordance Index (CI). The concordance index (CI) can be interpreted as the fraction of all pairs of subjects which predicted survival times are correctly ordered among all subjects that can actually be ordered. By the CI definition, we can determine > tj when ft > fj and Sj = 1, where /(•) is survival function. The pairs for which neither > tj nor < tj can be determined are excluded from the calculation of CI. Thus, the CI is defined as

L I; -(/, <fj A^, = 1)

I, Lj 1 (i, <ij A = l)'

Note that the values of CI are between 0 and 1, the perfect predictions of the building model would lead to 1 while have the CI value of 0.5 at random.

4.3. Analyses of the Simulated Data. In this section, we evaluate the performance of the harmonic regularization method for the Cox model in simulation study. We generate high-dimensional and low sample size data which contain many irrelevant features. Six methods are compared with our proposed harmonic regularization approach (HRA): the Lasso penalty (L1), the smoothly clipped absolute deviation penalty (SCAD), the minimax concave penalty (MCP), the adaptive Lasso (A-Lasso), the elastic net (Len), and the L1/2 penalty (L1/2).

We adopted the Cox model simulation scheme in Bender's work [30]. The data generation procedure is as follows.

Step 1. We generated the vectors yi0, y;1,..., yip (« = 1,..., n) independently from a standard normal distribution and the

Table 1: Average number of the variable selected and the recovery rate by the seven regularization methods on the simulated data in 500

Corr. Size L i SCAD Average of variable selected MCP A-Lasso Len L i/2 HRA

p = 0.1 a = 0.2 100 150 33.4 29.7 11.6 9.8 9.2 7.9 17.8 12.8 35.8 31.2 6.6 5.9 6.6 6.3

200 24.4 8.4 6.1 9.4 24.8 5.7 5.8

p = 0.1 a = 0.5 100 150 43.2 34.5 14.8 11.2 11.7 8.7 22.9 16.5 46.9 36.3 9.9 7.2 9.7 7.3

200 26.7 9.7 7.7 11.4 28.2 6.5 7

p = 0.5 a = 0.2 100 150 45.1 39.3 15.1 11.9 12.2 10.8 27.2 20.1 47.8 44.3 10.8 8.4 10.9 8.3

200 27.1 10.1 8.4 12.6 30.6 7.3 7.7

p = 0.5 a = 0.5 100 150 55.6 47.3 17.7 15.9 16 14.8 32.7 25.9 56.3 48.6 13.9 9.4 15.2 10

200 36.8 13.7 12.5 19.5 41.6 7.8 7.8

Corr. Size L i SCAD MCP Recovery rate A-Lasso Len L i/2 HRA

p = 0.1 a = 0.2 100 150 0.14 0.16 0.43 0.51 0.54 0.63 0.28 0.39 0.13 0.16 0.71 0.8 0.72 0.79

200 0.2 0.59 0.81 0.53 0.2 0.87 0.86

p = 0.1 a = 0.5 100 150 0.11 0.14 0.33 0.44 0.42 0.57 0.21 0.3 0.1 0.13 0.5 0.65 0.51 0.63

200 0.18 0.51 0.64 0.43 0.17 0.7 0.71

p = 0.5 a = 0.2 100 150 0.11 0.12 0.33 0.42 0.4 0.46 0.18 0.24 0.1 0.11 0.46 0.59 0.45 0.6

200 0.18 0.49 0.59 0.39 0.16 0.62 0.63

p = 0.5 a = 0.5 100 150 0.08 0.1 0.28 0.31 0.31 0.33 0.15 0.19 0.08 0.1 0.35 0.51 0.32 0.5

200 0.13 0.36 0.4 0.25 0.12 0.61 0.61

predictor vector xj is generated by xtj = ytj - p + Yio VP (J = 1'---'p)> where p is the correlation parameter of the predictor vectors.

Step 2. The survival time t[ (i = 1,...,n, n indicates the sample size) is constructed from a uniformly distributed variable U by t\ = (1/y) log(1 - (y x log(U))/(w exp(xjfi + a x e))), where y is the shape parameter, w is the scale parameter, p is the ground-true regression coefficients, e is the independent random error generated from N(0,1), and a is the parameter which controls the signal to noise.

Step 3. Censoring time point t'{ (i = 1,...,n) is obtained from an exponential distribution E(d), where d is determined by the specify censoring rate.

Step 4. Here we define tj = min(t'j,t'/) and Sj = I(t' < t''), the observed data represented as (tjt xj) for the Cox model (1) are generated.

In every simulation, the dimension p of the predictor vector xj is 1000, and the first five true coefficients are

nonzero: ft = 1, ft = 0.8, ft = -1, ft = -0.8, ft = 1, and ft = 0 (6 < j < 1000). About 25% of the data are right censored. We consider the cases with the training sample sizes n = 100, 150, 200, the correlation coefficient p = 0.1, 0.5, and the noise control parameter a = 0.2, 0.5, respectively. To assess the variability of the experiment, each method is evaluated on a test set including 100 random generated samples.

The estimation of the optimal tuning parameter A in the regularization models can be done in many ways and is often done by fc-fold cross-validation (CV). Note that the choice of k will depend on the size of the training set. In our experiments, we use 10-fold cross-validation (k = 10). The elastic net method has two tuning parameters; we need to cross-validate on a two-dimensional surface.

Table 1 shows the average number of variable selected and the recovery rate by each regularization method in 500 runs. The recovery rate is defined as the ratio of the average number of the selected relevant variables (x1-x5) to the average number of the selected variables [9]. As shown in Table 1, when the sample size n increases, the prediction

Table 2: Average IBS and CI results of by the seven regularization methods on the simulated data in 500 runs.

Corr. Size L i SCAD MCP Average IBS A-Lasso Len ^ i/2 HRA

p = 0.1 a = 0.2 100 150 0.084 0.081 0.094 0.087 0.086 0.084 0.084 0.08 0.082 0.08 0.091 0.084 0.086 0.079

200 0.078 0.086 0.079 0.083 0.076 0.078 0.076

p = 0.1 a = 0.5 100 150 0.096 0.092 0.092 0.091 0.097 0.094 0.096 0.094 0.094 0.087 0.098 0.089 0.093 0.09

200 0.088 0.088 0.086 0.087 0.085 0.086 0.086

p = 0.5 a = 0.2 100 150 0.105 0.098 0.098 0.096 0.101 0.099 0.102 0.102 0.097 0.092 0.101 0.098 0.094 0.091

200 0.091 0.092 0.096 0.096 0.089 0.095 0.09

p = 0.5 a = 0.5 100 150 0.108 0.101 0.103 0.097 0.108 0.1 0.106 0.096 0.099 0.093 0.01 0.094 0.098 0.097

200 0.084 0.094 0.086 0.084 0.082 0.091 0.086

Corr. Size L i SCAD MCP Average CI A-Lasso Len ^ i/2 HRA

p = 0.1 a = 0.2 100 150 0.749 0.832 0.788 0.853 0.822 0.869 0.757 0.838 0.851 0.868 0.838 0.865 0.845 0.87

200 0.85 0.847 0.857 0.859 0.864 0.859 0.862

p = 0.1 a = 0.5 100 150 0.728 0.82 0.758 0.841 0.767 0.833 0.716 0.831 0.727 0.853 0.761 0.847 0.763 0.837

200 0.847 0.857 0.862 0.846 0.869 0.862 0.866

p = 0.5 a = 0.2 100 150 0.726 0.781 0.758 0.818 0.752 0.821 0.745 0.793 0.752 0.819 0.758 0.813 0.748 0.821

200 0.786 0.835 0.826 0.792 0.839 0.824 0.828

p = 0.5 a = 0.5 100 150 0.699 0.766 0.712 0.777 0.701 0.817 0.685 0.788 0.719 0.814 0.716 0.818 0.714 0.814

200 0.776 0.801 0.82 0.808 0.821 0.819 0.819

performances of all the seven methods are improved. For example when p = 0.1 and a = 0.2, the average of the variables selected by the harmonic regularization method decreased from 6.6 to 5.8 and its recovery rate is improved from 0.72 to 0.86 with the sample sizes n increased from 100 to 200. When the correlation parameter p and the noise parameter a increase, the variable selection performances of all the seven methods are decreased. For example, when p = 0.1 and N = 200, the average of the recovery rate from the harmonic method decreased from 0.86 to 0.71, in which a increased from 0.2 to 0.5. When a = 0.5 and n = 150, the average of the recovery rate of the harmonic method decrease from 0.63 to 0.50, in which p increased from 0.1 to 0.5. Moreover, in our simulation, the influence of the noise may be slightly larger than that of the variable correlation for the prediction performance of all the seven methods. On the other hand, at the same parameter setting case, the recovery rates of the harmonic method and L 1/2 penalty are almost better than the results of the other five methods. For example, when p = 0.1, o = 0.2 and n = 100, the recovery rate of the harmonic method is 0.72 much better than 0.14, 0.43, 0.54, 0.28, and 0.13 got by the Lasso, SCAD, MCP, adaptive Lasso,

and elastic net, respectively, slight better than 0.71 got by L1/2 penalty method.

To evaluate prediction performance of the seven regu-larization methods for the Cox model, we presented their average IBC and CI values on the simulated datasets among 500 times in Table 2.

In terms of IBC and CI, for different parameters' settings, no methods almost performed better than others, but their prediction performances are only small differences. For example, when p = 0.1, a = 0.2, and n = 150, the average of IBS from the harmonic method is 0.079, better than 0.081, 0.087, 0.084, 0.08, 0.08, and 0.084 got by Lasso, SCAD, MCP, adaptive Lasso, and elastic net and L1/2 penalty. When p = 0.1, a = 0.2, and n = 100, the average of CI from the harmonic method is 0.845, better than 0.749, 0.788, 0.822, 0.757, and 0.838 got by Lasso, SCAD, MCP, adaptive Lasso, and L1/2, but slight worse than 0.851 got by elastic net penalty method.

Combined with the results reported in Table 1, we concluded that the harmonic penalized method showed better or equivalent predictive performance than the other regulariza-tion methods.

Table 3: The gene expression datasets are used in experiments.

Datasest Number of genes Number of samples Number of censored

DLBCL (2002) 7399 240 102

DLBCL (2003) 8810 92 28

Lung cancer 7129 86 62

AML 6283 116 49

Table 4: Results of the gene selected by the seven methods on the four public datasets.

Datasest L i SCAD MCP A-Lasso Len L i/2 HRA

DLBCL (2002) 174 129 129 146 180 76 71

DLBCL (2003) 138 106 95 168 142 32 37

Lung cancer 188 104 97 233 196 56 48

AML 161 120 110 176 166 65 70

In bold is the best performance.

Table 5: The IBS results obtained by the seven methods on the four public datasets.

Datasets L i SCAD MCP Average IBS A-Lasso Len L i/2 HRA

DLBCL (2002) 0.207 0.205 0.205 0.205 0.198 0.203 0.205

DLBCL (2003) 0.121 0.119 0.12 0.12 0.12 0.118 0.119

Lung cancer 0.169 0.161 0.167 0.164 0.169 0.163 0.161

AML 0.174 0.174 0.173 0.172 0.171 0.173 0.171

Datasets L i SCAD MCP Average CI A-Lasso Len L i/2 HRA

DLBCL (2002) 0.553 0.554 0.564 0.555 0.568 0.563 0.566

DLBCL (2003) 0.583 0.604 0.586 0.589 0.606 0.603 0.606

Lung cancer 0.628 0.634 0.666 0.646 0.675 0.673 0.674

AML 0.599 0.611 0.634 0.626 0.641 0.638 0.643

In bold-the best performance.

4.4. Analysis of the Real Microarray Datasets. In this section, we evaluated the performance of the harmonic regularization methods on the real survival gene expression datasets. Four publicly available datasets are used in this part. A brief description of these datasets is given below and summarized in Table 3.

Diffuse Large B-cell Lymphoma Dataset (DLBCL) 2002. This dataset published by Rosenwald et al. [31]. The dataset consists of 240 samples from patients. For each sample, 7399 gene expression measurements were obtained. The clinical outcome was survival time, either observed or censored.

Diffuse Large B-cell Lymphoma Dataset (DLBCL) 2003. This dataset is from Rosenwald et al. [32]. It consists of 92 lymphoma patients, and each patient has 8810 genes.

Lung Cancer Dataset. The lung cancer dataset is from Beer et al. [33]. It consists of gene expressions of 4966 genes for 83 patients. The survival time as well as the censoring status is available.

AML Dataset. The AML dataset is from Bullinger et al. [34]. It contains the expression profiles of 6283 genes for 116 patients, and the number of censored cases is 49.

We evaluated the prediction accuracies of the seven estimated regularization methods using random partition: a training set of about 2/3 of the patients used for estimation and a test set of about 1/3 of the patients used for testing of the prediction capability. For estimating A, we employed the five-fold cross-validation scheme using the training set. We repeated each procedure 200 times.

Table 4 reports the average number of genes selected by each method. The harmonic regularization method performs better than those of L1 type methods (Lasso, SCAD, MCP, adaptive Lasso, and elitist net), and slightly better than that of L1/2 penalty. As shown in Table 4, for DLBCL (2002) dataset, the harmonic penalized methods selected about 71 genes, compared to about 174, 129, 129, 146, and 180 about five Lasso, SCAD, MCP, adaptive Lasso and elitist net, slightly better than 76 got by L1/2 penalty. For DLBCL (2003) and AML datasets, the best one is L1/2 penalty and the second is harmonic methods.

To assess predictive performance, we summarize the results of IBS and CI obtained by the seven methods, respectively, in Table 5. Both the results of IBS and CI, the results of all regularization methods, were not much different and the elitist net and harmonic penalized method almost

outperforms than other five penalized methods. Combined with the results reported in Tables 4 and 5, we concluded that the harmonic penalized method selected the smaller subset of the key genes while give best or equivalent predictive performance.

5. Conclusion

Variable selection is a fundamental problem in statistics and machine learning, and the regularization method is one of the ways to solve this problem. Generally speaking, a regularization algorithm is always a combination of a loss function and a penalty function in the past research and applications. Particularly, in the procedure of variable selection, the harmonic regularization is like a net which can always catch the correct model. This demonstrates the stronger sparsity and better correctness of the harmonic regularization. We have provided a serous of simulations to demonstrate that L1 type regularization methods are inefficient; the harmonic regularization and L1/2 penalty methods proved are efficient and effective.

In the simulation part, we use four real datasets. There are the DLBCL (2002), the DLBCL (2003), the Lung cancer, and the AML. Results indicate that our harmonic regularization algorithm is very competitive in analyzing high dimensional survival data in terms of sparsity. Simulation results indicate that the harmonic penalized Cox model is very competitive in analyzing high dimensional survival data, because it was able to reduce the size of the predictor even further at moderate costs for the prediction accuracy [8]. The harmonic penalized Cox model will provide an efficient tool in building a prediction model for survival time based on high dimensional biological data.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.


This research was supported by Macau Science and Technology Development Funds (Grant no. 099/2013/A3) of Macau Special Administrative Region of the People's Republic of China.


[1] E. L. Leung, Z. W. Cao, Z. H. Jiang, H. Zhou, and L. Liu, "Network-based drug discovery by integrating systems biology and computational technologies," Briefings in Bioinformatics, vol. 14, no. 4, pp. 491-505, 2013.

[2] D. R. Cox, "Regression models and life-tables," Journal of the Royal Statistical Society B. Methodological, vol. 34, no. 2, pp. 187220,1972.

[3] D. R. Cox, "Partial likelihood," Biometrika, vol. 62, no. 2, pp. 269-276, 1975.

[4] J. Fan and R. Li, "Variable selection via nonconcave penalized likelihood and its oracle properties," Journal of the American Statistical Association, vol. 96, no. 456, pp. 1348-1360, 2001.

[5] R. Tibshirani, "Regression shrinkage and selection via the lasso," Journal oftheRoyal Statistical SocietyB. Methodological, vol. 58, no. 1, pp. 267-288,1996.

[6] H. H. Zhang and W. Lu, "Adaptive Lasso for Cox's proportional hazards model," Biometrika, vol. 94, no. 3, pp. 691-703, 2007.

[7] C. H. Zhang, "Nearly unbiased variable selection under minimax concave penalty," The Annals of Statistics, vol. 38, no. 2, pp. 894-942, 2010.

[8] Z. Liu, F. Jiang, G. Tian et al., "Sparse logistic regression with Lp penalty for biomarker identification," Statistical Applications in Genetics and Molecular Biology, vol. 6, no. 1, 2007.

[9] Z. B. Xu, H. Zhang, Y. Wang, X. Y. Chang, and Y. Liang, "L 1/2 regularization," Science in China Series F: Information Sciences, vol. 53, no. 6, pp. 1159-1169, 2010.

[10] Z. B. Xu, X. Y. Chang, and F. M. Xu, "L-1/2 regularization: a thresholding representation theory and a fast solver," IEEE Transaction on Neural Networks and Learning Systems, vol. 23, no. 7, pp. 1013-1027, 2012.

[11] R. Tibshirani, "The lasso method for variable selection in the Cox model," Statistics in Medicine, vol. 16, pp. 385-395,1997.

[12] J. Gui and H. Li, "Penalized Cox regression analysis in the high-dimensional and low-sample size settings, with applications to microarray gene expression data," Bioinformatics, vol. 21, no. 13, pp. 3001-3008, 2005.

[13] H. Zou and T. Hastie, "Regularization and variable selection via the elastic net," Journal of the Royal Statistical Society B: Statistical Methodology, vol. 67, no. 2, pp. 301-320, 2005.

[14] S. Rosset and J. Zhu, "Piecewise linear regularized solution paths," Annals of Statistics, vol. 35, no. 3, pp. 1012-1030, 2007.

[15] T. Zhang, "Analysis of multi-stage convex relaxation for sparse regularization," Journal of Machine Learning Research (JMLR), vol. 11, pp. 1081-1107, 2010.

[16] R. Mazumder, J. H. Friedman, and T. Hastie, "SparseNet : coordinate descent with non-convex penalties," Journal of the American Statistical Association, vol. 106, no. 495, pp. 1125-1138, 2011.

[17] Y. Liang, C. Liu, X.-Z. Luan et al., "Sparse logistic regression with a L1/2 penalty for gene selection in cancer classification," BMC Bioinformatics, vol. 14, no. 1, article 198, 2013.

[18] C. Liu, Y. Liang, X. Z. Luan et al., "The L1/2 regularization method for variable selection in the Cox model," Applied Soft Computing, vol. 14, pp. 498-503, 2013.

[19] W. J. Fu, "Penalized regressions: the bridge versus the lasso," Journal of Computational and Graphical Statistics, vol. 7, no. 3, pp. 397-416, 1998.

[20] T. Blumensath, M. Yaghoobi, and M. E. Davies, "Iterative hard thresholding and L0 regularisation," in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '07), vol. 3, pp. 877-880, April 2007.

[21] I. Sohn, J. Kim, S.-H. Jung, and C. Park, "Gradient lasso for Cox proportional hazards model," Bioinformatics, vol. 25, no. 14, pp. 1775-1781, 2009.

[22] P. Zhao and B. Yu, "Stagewise lasso," Journal of Machine Learning Research, vol. 8, pp. 2701-2726, 2007.

[23] D. L. Donoho and X. Huo, "Uncertainty principles and ideal atomic decomposition," IEEE Transactions on Information Theory, vol. 47, no. 7, pp. 2845-2862, 2001.

[24] D. L. Donoho and E. Elad, "Maximal sparsity representation via L1 minimization," Proceedings of the National Academy of Sciences of the United States of America, vol. 100, no. 5, pp. 21972202, 2003.

[25] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, "Least angle regression," The Annals of Statistics, vol. 32, no. 2, pp. 407-499, 2004.

[26] J. H. Friedman, "Fast sparse regression and classification," Tech. Rep., Stanford University, Deptartment of Statistics, 2008.

[27] H. C. van Houwelingen, T. Bruinsma, A. A. M. Hart, L. J. van't Veer, and L. Wessels, "Cross-validated Cox regression on microarray gene expression data," Statistics in Medicine, vol. 25, no. 18, pp. 3201-3216, 2006.

[28] E. Graf, C. Schmoor, W. Sauerbrei, and M. Schumacher, "Assessment and comparison of prognostic classification schemes for survival data," Statistics in Medicine, vol. 18, no. 17-18, pp. 25292545, 1999.

[29] F. E. Harrel, Regression Modeling Strategies, with Applications to Linear Models, Logistic Regression, and Survival Analysis, Springer, New York, NY, USA, 2001.

[30] R. Bender, T. Augustin, and M. Blettner, "Generating survival times to simulate Cox proportional hazards models," Statistics in Medicine, vol. 24, no. 11, pp. 1713-1723, 2005.

[31] A. Rosenwald, G. Wright, W. C. Chan et al., "The use of molecular profiling to predict survival after chemotherapy for diffuse large B-cell lymphoma," The New England Journal of Medicine, vol. 346, pp. 1937-1946, 2002.

[32] A. Rosenwald, G. Wright, A. Wiestner et al., "The proliferation gene expression signature is a quantitative integrator of oncogenic events that predicts survival in mantle cell lymphoma," Cancer Cell, vol. 3, no. 2, pp. 185-197, 2003.

[33] D. G. Beer, S. L. R. Kardia, C.-C. Huang et al., "Gene-expression profiles predict survival of patients with lung adenocarcinoma," Nature Medicine, vol. 8, no. 8, pp. 816-824, 2002.

[34] L. Bullinger, K. Dohner, E. Bair et al., "Use of gene-expression profiling to identify prognostic subclasses in adult acute myeloid leukemia," The New England Journal of Medicine, vol. 350, no. 16, pp. 1605-1616, 2004.

Copyright of Computational & Mathematical Methods in Medicine is the property of Hindawi Publishing Corporation and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use.