- Research
- Open Access

# Derivative component analysis for mass spectral serum proteomic profiles

- Henry Han
^{1, 2}Email author

**7 (Suppl 1)**:S5

https://doi.org/10.1186/1755-8794-7-S1-S5

© Han; licensee BioMed Central Ltd. 2014

**Published:**8 May 2014

## Abstract

### Background

As a promising way to transform medicine, mass spectrometry based proteomics technologies have seen a great progress in identifying disease biomarkers for clinical diagnosis and prognosis. However, there is a lack of effective feature selection methods that are able to capture essential data behaviors to achieve clinical level disease diagnosis. Moreover, it faces a challenge from data reproducibility, which means that no two independent studies have been found to produce same proteomic patterns. Such reproducibility issue causes the identified biomarker patterns to lose repeatability and prevents it from real clinical usage.

### Methods

In this work, we propose a novel machine-learning algorithm: derivative component analysis (DCA) for high-dimensional mass spectral proteomic profiles. As an implicit feature selection algorithm, derivative component analysis examines input proteomics data in a multi-resolution approach by seeking its derivatives to capture latent data characteristics and conduct de-noising. We further demonstrate DCA's advantages in disease diagnosis by viewing input proteomics data as a profile biomarker via integrating it with support vector machines to tackle the reproducibility issue, besides comparing it with state-of-the-art peers.

### Results

Our results show that high-dimensional proteomics data are actually linearly separable under proposed derivative component analysis (DCA). As a novel multi-resolution feature selection algorithm, DCA not only overcomes the weakness of the traditional methods in subtle data behavior discovery, but also suggests an effective resolution to overcoming proteomics data's reproducibility problem and provides new techniques and insights in translational bioinformatics and machine learning. The DCA-based profile biomarker diagnosis makes clinical level diagnostic performances reproducible across different proteomic data, which is more robust and systematic than the existing biomarker discovery based diagnosis.

### Conclusions

Our findings demonstrate the feasibility and power of the proposed DCA-based profile biomarker diagnosis in achieving high sensitivity and conquering the data reproducibility issue in serum proteomics. Furthermore, our proposed derivative component analysis suggests the subtle data characteristics gleaning and de-noising are essential in separating true signals from red herrings for high-dimensional proteomic profiles, which can be more important than the conventional feature selection or dimension reduction. In particular, our profile biomarker diagnosis can be generalized to other omics data for derivative component analysis (DCA)'s nature of generic data analysis.

## Keywords

- Feature Selection
- Partial Less Square
- Discrete Wavelet Transform
- Support Vector Machine Classifier
- Feature Selection Method

## Background

With the surge in serum proteomics, large volumes of mass spectral serum proteomic data are available to make molecular diagnosis of complex disease phenotypes possible. As a promising way to revolutionize medicine, serum proteomics demonstrates a great potential in identifying novel biomarker patterns from the serum proteome for diagnosis, prognosis, and early disease discovery [1–3]. However, high-performance disease phenotype discrimination remains a challenge in translational bioinformatics due to special characteristics of serum proteomics data, in addition to its well-known data reproducibility issue, which means that no two independent studies have been found to produce same proteomic patterns [3–5].

A serum proteomic data set can be represented as a matrix $X\in {\Re}^{n\times p}$ after preprocessing, where each row represents protein expression at a mass-to-charge (*m/z*) ratio of peptides or proteins and each column represents protein expression from a sample/observation (e.g., a control or cancer subject) across all *m/z* ratios in experiment. The number of rows is much greater than the number of columns, $p\text{<<}n,$ that is #variables (peptides/proteins) is much greater than #samples. Usually $n~O\left(1{0}^{4}\right),$ and $p~O\left(1{0}^{2}\right).$ Although there are a large amount of *m/z* ratios (peptides or proteins), only a few numbers of them (e.g., peaks) have meaningful contribution to disease diagnosis and data variations. Moreover, such data are not noise-free because normalization methods cannot remove built-in systems noise from mass spectrometry technology itself [6, 7]. In particular, the high-dimensionality directly prevents conventional classification algorithms from achieving clinical rivaling disease diagnosis, limits its generalization capability or even causes some regularity problem in classification [7].

Quite a lot feature selection methods have been employed in serum proteomic data classification to glean informative features, reduce dimension, or conduct de-noising in order to achieve high accuracy disease diagnosis [7–10]. It is noted that a feature refers to a row in a serum proteomic data set, which are biologically peptides or proteins. In this work, we categorize them into input-space and subspace methods respectively. The former seeks a feature subset ${X}^{\prime}\in {\Re}^{m\times p},$$m\text{<<}n$ in the same space ${\Re}^{n\times p}$ as input data $\phantom{\rule{0.1em}{0ex}}X$ by conducting a hypothesis test (e.g., *t-test*), or wrapping a classifier to features recursively; The latter conducts dimension reduction by transforming data $\phantom{\rule{0.1em}{0ex}}X$ into a subspace *S* induced by a linear or nonlinear transformation $f:X\to S$ where $S=span\left({s}_{1},{s}_{2}\dots {s}_{k}\right),$${s}_{k}\in {\Re}^{n\times p},$$k\le p\le n,$ and seeking meaningful linear combinations of features. For example, the subspace spanned by all principal components when the transformation is induced by principal component analysis (PCA) [11].

All subspace methods can be formulated as a matrix decomposition problem: $X~S{P}^{T},$$S\in {\Re}^{n\times k},$$P\in {\Re}^{p\times k}$ where different methods construct different basis matrices *S* and different feature matrices *P* according to different termination conditions. For instance, nonnegative matrix factorization (NMF) seeks nonnegative matrix decomposition such that $\left|\right|X~S{P}^{T}\left|\right|$ is minimized under an Euclidean distance or K-L divergence [12, 13]. In fact, almost all PCA, ICA, and NMF 's extensions such as nonnegative principal component analysis (NPCA), sparse NMF, and other methods such as random projection methods all fall into this category [8, 12–16].

However, these methods may not always contribute to improving diagnosis in serum proteomics robustly. Instead, it was reported that classifiers integrated with them may usually demonstrate large oscillations in performance for different data sets and some even got worse performance than the case without feature selection [7, 8, 10]. Moreover, there was no systematic work on addressing the limitations of those feature selection methods. In this work, we address these methods' limitations before introducing our novel derivative component analysis (DCA).

### Lack of de-noising schemes

The input-space methods usually lack de-noising schemes and assume input data is clean or nearly clean. Such an assumption can be true for the data that are by nature clean or with quite low-level noise (e.g., financial data). However, it appears to be inappropriate for serum proteomics data since they usually contain nonlinear noise from profiling systems, and technical/biological artifacts. The noise would enter feature selection as outliers and produce less informative or even ad-hoc feature sets (e.g., peaks with less biological meaning), which would lead to an inaccurate or even poor decision function in classification and affect the disease phenotype diagnosis, generalization, and biomarker discovery in translational bioinformatics.

### Latent data characteristics missing

Those subspace methods have difficulties in capturing subtle or latent data characteristics, because subspace methods transform data into another subspace to seek meaningful feature combination and original spatial coordinates are 'lost', which makes it almost impossible to track those features contributing to the behaviors. The latent data characteristics refer to subtle data behaviors interpreting transient data changes (we use words 'subtle' and 'latent' equivalently when describing data characteristics in our context). Quite different from global data characteristics that referring to the holistic data behaviors interpreting long-time interval data changes, subtle data characteristics have to be represented by the first or even high-level derivative of data mathematically [8, 10].

We use principal component analysis (PCA) as an example to address this issue. Given input data with zero mean $X\in {\Re}^{n\times p},$ the subspace is spanned by selected PCs, i.e. $S=span\left({u}_{1},{u}_{2},\dots {u}_{k}\right)$$1\le k\le p$. Since each subspace basis (PC) receives contributions from all features (peptides/proteins) in the linear combinations, changes in one feature will inevitably affect all bases globally. Although it is biologically important to identity which protein/peptide has more contributions to the data change, it is quite hard to achieve it because their coefficients in the linear combination are not usually comparable [6]. Moreover, subspace basis calculation does not involve the feature derivative information or its related approximation, which causes each PC not to be able to capture latent (subtle) data characteristics well. As such, only global data characteristics can be captured well and subtle data characteristics, which are essential in achieving high performance diagnosis, may be totally missed. For example, some malignant and benign tumors may have similar global data characteristics but different subtle data characteristics in serum profiling. As such, detecting subtle data characteristics is essential to achieve a clinical level diagnosis.

Although various subspace methods such as sparse-PCA, nonnegative-PCA, and sparse-NMF [12, 8, 14, 16], have been proposed to enhance subtle data characteristics capturing by imposing non-negativity or sparsity constraints in order to seek subspace bases through solving a nonlinear optimization problem, they are usually characterized by high complexities (e.g., nonnegative PCA [6, 8]) and none of them seems to be able to catch subtle data characteristics by examining the features 'beyond' their original data level.

In this work, we propose a *de novo* derivative component analysis (DCA), which evolves from author's previous work in gene and protein expression omics data analysis [8, 9], to overcome the current feature selection methods' weaknesses for the sake of clinical level disease diagnosis in serum proteomics. It is worthwhile to point out that our DCA is a novel machine learning algorithm based on our global and local feature selection theory proposed in [8], which is more complicated and powerful than the serum proteomics data analysis methods that straight-forwardly apply wavelet transforms to a proteomic sample and conduct classic statistical tests to following wavelet coefficients [17]. Our DCA employs discrete wavelet transforms (DWT) [18] to look at serum proteomics data in *'multiple windows*' to extract latent data characteristics and achieve de-noising by retrieving 'data derivatives'.

Furthermore, we employ benchmark serum proteomic data to demonstrate DCA's superiority in disease diagnosis by proposing a novel diagnosis algorithm DCA-SVM and comparing it with the other state-of-the-art peers. The exceptional performance of our DCA-SVM suggests it can be a potential way to overcome the serum proteomics' reproducibility by viewing input data as a profile biomarker. As a key result in this work, we present DCA-MARK, a DCA-based biomarker discovery algorithm that strongly demonstrates high-dimensional serum proteomics data's linear separability, which not only has an important meaning in machine learning, but also has practical impacts on translational bioinformatics for its novelty. To the best of our knowledge, it is the first work that is able to linearly separate high-dimensional serum proteomic data with few biomarkers.

## Derivative Component Analysis (DCA)

Different from its conventional definition, a feature is no longer viewed as an indecomposable information unit in DCA. Instead, all features are hierarchically decomposed into different components to discover data derivatives to capture subtle data characteristics and conduct de-noising. The proposed derivative component analysis (DCA) consists of the following three steps.

First, a discrete wavelet transform (DWT) is applied to all features to decompose it hierarchically as a set of detail coefficient matrices $c{D}_{1},c{D}_{2}\dots c{D}_{J}$ and an approximation matrix $c{A}_{J}$ under a transform level *J*. Since DWT is done on a set of dyadic grid points hierarchically, the dimensionalities of the approximation and detail coefficient matrices shrink dyadically from level 1 to level *J* [17]. For example, given a proteomic data set with 10 samples across 1024 *m/z* ratios under a DWT with a transform level *J = 5, cD*_{
1
} is a 10 × 512 matrix and *cD*_{
2
} is 10 × 256 matrix. Similarly, *cD*_{
5
} and *cA*_{
5
} both are 10 × 32 matrices.

The approximation matrix and coarse level detail coefficient matrices (e.g., *cDJ*) capture the global data characteristics, because they contain contributions from the features disclose slow changes in 'long-time windows', if we view each *m/z* ratio as a corresponding time point in our context. Similarly, the fine level detail coefficient matrices (e.g., *cD*_{
1
}*, cD*_{
2
}), capture subtle data characteristics, because they contain contributions from the features that disclose quick changes in 'short-time windows'. In fact, the fine level detail matrices are components to reflect data derivatives in different time windows. Furthermore, most system noises are hidden in these components for its heterogeneity with respect to true signals. In summary, the first step separates global characteristics, subtle data characteristics, and noise in different resolutions.

Second, retrieve the most important subtle data behaviors and remove noise by reconstructing the fine level detail coefficient matrices before or at a presetting cutoff level *τ* (e.g*.,τ = 3*). Such construction consist of two steps: 1) Conduct principal component analysis (PCA) for the detail matrices $c{D}_{1},c{D}_{2}\dots c{D}_{\tau}$ 2) Reconstruct each detail coefficient matrix by using its first *m* leading loading vectors, i.e., principal components, in its each principal component (PC) matrix. Usually, we set *m* = 1, i.e., we employ the first principal component to reconstruct each detail coefficient matrix, which means we only retrieve the most important subtle data characteristics in detail coefficient matrix reconstruction. In fact, the first PC based reconstruction also achieves de-noising by suppressing noise's contribution in the detail coefficient matrix reconstruction because noise has is usually unlikely to appear in the 1^{st} PC.

On the other hand, the coarse level detail coefficient matrices after the cutoff *τ* : $c{D}_{\tau +1},\phantom{\rule{2.77695pt}{0ex}}c{D}_{\tau +2}\dots c{D}_{J}$ and approximation coefficient matrix $c{A}_{J}$ are kept intact to retrieve global data characteristics. In fact, parameter *m* can be also determined by using a variability explanation ratio ${\rho}_{m}$ defined as follows, such that it is greater than a threshold $\phantom{\rule{0.1em}{0ex}}\rho $ (e.g., $\rho =60\%$), which is the variability explanation ratio by the first principal component of those detail coefficient matrices before or equal the cutoff.

### Variability explanation ratio

Given a data set with *n* variables and *p* observations, usually, *p < n*, the variability explanation ratio is the ratio between the variance explained by the first *m* PCs and the total data variances: ${\rho}_{m}={\displaystyle \sum _{i=1}^{m}}{\sigma}_{i}/{\displaystyle \sum _{i=1}^{p}}{\sigma}_{i}$, where ${\sigma}_{j}$ is the variance explained by the ${j}^{th}$ PC, which is actually the ${j}^{th}$ eigenvalue of the covariance matrix of the input proteomic data.

It is noted that such a selective reconstruction process in the second step extracts the most important subtle data characteristics and conduct de-noising by suppressing the contribution from system noise. This is because only one or few principal components are employed in reconstructing each targeted fine level coefficient matrix $c{D}_{j}$ and those less important and noise-contained principal components are dropped in reconstruction.

Third, conduct the corresponding inverse DWT by using the current detail and approximation coefficient matrices to obtain an meta-data ${X}_{*}$ that is the corresponding de-noised data set with subtle data characteristics extraction and system noise removal, because of the highlight of the most significant subtle data behaviors in the "derivative components" based reconstructions. The meta-data are just 'true signals' separated from red herrings that share the same dimensionality with the original data but with less memory storage because less important PCs are dropped in our reconstruction.

It is noted that, unlike traditional feature selection methods, DCA is an implicit feature selection method, where useful characteristics are selected implicitly without an obvious variable removal or dimension reduction. Algorithm 1 gives the details about DCA as follows, where we use ${X}^{T}$ instead of *X* to represent input proteomic data for the convenience of description, i.e. each row is a sample and each column is a feature in the current context.

**Algorithm Derivative Component Analysis (DCA)**

1. **Input:** ${X}^{T}=\left[{x}_{1},{x}_{2},\dots {x}_{n}\right]$${x}_{i}\in {\Re}^{p},$*DWT level J; cutoff τ; wavelet* $\psi ,$*thereshold* $\rho ,$

2. **Output:** Meta-data ${X}_{*}^{T}$

3. **Step 1**. *Column-wise discrete wavelet transforms (DWT)*

4. Conduct J-level DWT with wavelet $\phantom{\rule{0.1em}{0ex}}\psi $ for each column of ${X}^{T}$ to obtain

$\left[c{D}_{1},\phantom{\rule{2.77695pt}{0ex}}c{D}_{2}\dots c{D}_{J};c{A}_{J}\right],$$cDj\in {\Re}^{{p}_{j}\times n},$$c{A}_{J}\in {\Re}^{{p}_{J}\times n},$ and ${p}_{j}=\left[p/{2}^{j}\right],$$j=1,2,\dots J$.

5. **Step 2.** *Derivative component analysis for latent data characteristics extraction and de-noising*

6. *for j* = 1 to *J*

7. if *j* ≥ *τ*

8. a) Do principal component analysis for each detail matrix $c{D}_{j}$ to obtain its PC and score matrix,

9. $U=\left[{u}_{1},{u}_{2},\dots {u}_{p}\right],$${u}_{i}\in {\Re}^{n}$ and $S=\left[{s}_{1},{s}_{2}\dots {s}_{{p}_{J}}\right],$${s}_{i}\in {\Re}^{{p}_{j}},$$i=1,2,\cdots {p}_{j}$

10. b) Reconstruct matrix $c{D}_{j}$ by employing first *m* principal components ${u}_{1},{u}_{2},\dots {u}_{m},$*s.t.* ${\rho}_{m}\ge p$

11. $c{D}_{j}\leftarrow c{D}_{j}\times \left(I\times {I}^{T}\right)/{p}_{{}_{j}}+{\displaystyle \sum _{i=1}^{m}}{u}_{i}\times {s}_{i}^{T},$$I={\left[1,1,\cdots 1\right]}^{T}\in {\Re}^{{p}_{j}}$

12. *end if*

13. *end for*

14. **Step 3**. *Approximate the original data by the inverse discrete wavelet transform*

15. ${X}_{*}^{T}\leftarrow inverseDWT\left(\left[c{D}_{1},\phantom{\rule{2.77695pt}{0ex}}c{D}_{2}\cdots c{D}_{J};c{A}_{J}\right]\right)$ with the wavelet $\phantom{\rule{0.1em}{0ex}}\psi $

### Tuning parameters in derivative component analysis

Although an optimal DWT level can be obtained theoretically by following the maximum entropy principle [19], it is reasonable to adaptively select the DWT level *J* according to the *'nature'* of input data, where large #samples corresponds to a relatively large *J* value, for the convenience of computation. Although the convolution in the DWT always introduces a few extra entries into each feature's corresponding detail coefficient vector in *cDj+1* such that its length is slightly more than the half of that of in *cD*_{
j
} [18], we have found that a large transform level does not show advantages compared with the a small transform level in feature selection. However, a small transform level (e.g., *J = 3)* may bring some hard time in separating subtle and global data characteristics because of the limited choice for the cutoff *τ*. As such, we select the DWT level as $4\le J\le \lceil {\text{log}}_{2}p\rceil $ considering the magnitude level of the #samples, i.e. $p~O\left(1{0}^{2}\right)$ for a proteomics data set. Correspondingly, we empirically set the cutoff as $1<\tau \le J/2$ to separate the fine and coarse level detail coefficient matrices for its robust performance.

Furthermore, we require the wavelet $\phantom{\rule{0.1em}{0ex}}\psi $ in the DWT to be orthogonal and have compact supports such as *Daubechies* wavelets (e.g., '*db8*'), for the sake of the subtle data behavior capturing. The variability explanation ratio threshold is usually set as *ρ ≥ 60%*, which means the reconstructed fine level detail coefficient matrix *cD*_{
j
} (*1 ≤ j ≤ τ*) contains at least 60% variances of the original one, to retrieve the most important subtle data behaviors interpreted by *cD*_{
j
}. Interestingly, we have found that the first PC of each fine-level detail coefficient matrix usually count quite a high variability explanation ratio (e.g. >60%) for each fine-level detail coefficient matrix *cD*_{
j
} (*1 ≤ j ≤ τ* ). Thus, we relax the variability explanation ratio threshold ρ by only using the first PC to reconstruct each *cD*_{
j
} matrix to catch the subtle data characteristics along the maximum variance direction. In fact, we have found that using more PCs in the fine-level detail coefficient matrix reconstruction does not demonstrate advantages in subtle data characteristics extraction and de-noising than using the first PC.

*Ovarian-qaqc*data with 95 controls and 121 ovarian cancer samples across 15,000

*m/z*ratios [20], and its two level detail coefficient reconstructions under DCA with

*τ*=

*2, J*=

*7*, and wavelet 'db8'. Interestingly, the meta-data are

*smoother*and have

*values in a smaller range*than the original feature for its subtle data characteristics capturing and de-noising, which reflect the true expression of the peptides/proteins at the

*m/z*ratio better. In other words, DCA provides a 'zooming' mechanism to capture the original data's subtle behaviors that are usually latent in general feature selection methods. It is noted that similar results can be obtained for other mass spectral proteomic profiles also.

*Colorectal*data [17] with total 48 controls and 64 cancer samples across 16,331

*m/z*ratios, extracted by our DCA under the cutoff

*τ*=

*2*, transform-level

*J*=

*7*, and wavelet

*'db8'*. For the convenience of description, true signals are highlighted between 1,400 Da and 1,500 Da. Interestingly, the each type of samples in the extracted true signals appear to be smoother and more proximal to each other besides demonstrating less variations, because of major subtle data characteristics extraction and system noise removal. Obviously, from a classification viewpoint, these true signals will contribute to high accuracy diagnoses than the original proteomic data, because the built-in noises and redundant global data characteristics would have a much lower chance to get involved in classification due to derivative component analysis. Instead, subtle data characteristics would have a greater chance of participating in the decision rule inference.

### Disease diagnosis with Derivative Component Analysis

Since DCA can separate true signals from red herrings by extracting subtle data characteristics and removing built-in noises, it is natural to combine DCA with the start-of-the-art classifiers to demonstrate its effectiveness in serum proteomic disease diagnosis. We choose support vector machines (SVM) for its efficiency and popularity in translational bioinformatics [21]. As such, we propose novel derivative component analysis based support vector machines (DCA-SVM) to handle serum proteomic disease diagnosis, which is equivalent to a binary or multi-class classification problem. Thus, we briefly describe the corresponding binary and multiclass DCA-SVM as follows.

*Y*, which is equivalent to solving the following quadratic programming problem (standard SVM, i.e., C-SVM):

The normal of the maximum-margin hyperplane can be calculated by the equation $w={\sum}_{i=1}^{p}{\alpha}_{i}{c}_{i}{y}_{i},$ where the sparsity of variables ${\alpha}_{1}$*i = 1,2,...p*, makes classification only dependent on few training points, which are few cancerous patients or healthy subjects in the proteomics data used for training. The decision function $f\left({x}^{\prime}\right)=sign({\sum}_{i=1}^{p}{\alpha}_{i}k\left({y}_{i}\u2022{y}^{\prime}\right)+b$ is used to determine the class type of a testing sample ${x}^{\prime},$ where ${y}^{\prime}$ is its corresponding meta-sample computed from DCA. The function $k\left({y}_{i}\u2022{y}^{\prime}\right)$ is a kernel function mapping $\phantom{\rule{0.1em}{0ex}}y$ and ${y}^{\prime}$ into a same-dimensional or high-dimensional feature space. In this work, we employ the *'linear'* kernel $k\left(x\u2022y\right)=\left(x\u2022y\right)$ for its simplicity and efficiency (more detailed reason for such a kernel selection can be found in the following section). Such a decision function answers the query: 'is this proteomic sample is from a patient with a specified disease or a normal individual?'

Our multiclass DCA-SVM algorithm employs the *'one-against-one'* for its proved advantage over the *'one-against-all'* and *'directed acyclic SVM'* methods [21, 22]. The *'one-against-one'* method builds *k(k-1)/2* binary SVM classifiers for a data set with *k* classes *{1,2,...k}*, each of which correspond to a pathological state. Each classifier is trained on data from two classes, i.e. training samples are from the *i-th* and *j-th* classes, *i, j = 1* *, 2 ... k*. After building all *k(k-1)/2* classifiers, we employ the *'Max-wins'* voting approach to infer its final class type: if the local decision function says ${x}^{\prime}$ is in the class *i*, then the class *i* wins one vote; Otherwise, the class *j* wins one vote. Finally, sample ${x}^{\prime}$ will belong to the class with the largest vote.

### The DCA-SVM 's advantages over SVM in disease diagnosis

It is worthwhile to point out that, compared with the standard SVM, our DCA-SVM has a different feature space due to the true-signals extraction from DCA, which leads to a more robust decision rule than the standard SVM (C-SVM) for inviting the de-noised data with the subtle data characteristics in the optimal hyperplane construction. Obviously, the decision rule inferred from our DCA-SVM would avoid the traditional bias from that of the standard SVM. On the other hand, the standard SVM's feature space usually contains noises from input proteomic data, and misses the subtle data characteristics, which limit the classifier's performance and lead to a biased, global data characteristics favored decision rule.

Alternatively, the DCA-SVM 's feature space contains 'de-noised' true signals with the subtle data characteristics, which avoids the global data characteristics favored decision rule inference because the subtle data characteristics are also invited in SVM hyperplane construction besides the global data characteristics. As such, the DCA-SVM can efficiently detect those samples with similar global characteristics but different subtle characteristics in disease diagnosis than the standard SVM, which contributes to the high accuracy diagnosis.

## Results

We demonstrate our DCA-SVM can achieve rivaling-clinical diagnosis by using five benchmark high-dimensional serum proteomic data sets [17, 20, 23–25] and compare it with state-of-the-art peers on these data. We introduce details about the data sets as follows.

### Data sets

The benchmark data sets used in the experiment are heterogeneous data generated from different experiments via different high-resolution serum profiling technologies such as MALDI (matrix-assisted laser desorption)-TOF (time-of-flight), SELDI (surface enhanced laser desorption and ionization)-TOF (time-of-flight), and SELDI-QqTOF (quadrupole time-of-flight). The details of the data sets are as follows.

*Cirrhosis* data set is a three-class MALDI-TOF serum proteomic data with total 201 spectra that consisting of 72 samples from healthy individuals, 78 samples from patients with hepatocellular carcinoma (HCC), the most common liver cancer, and 51 samples form cirrhosis patients, across 23,846 *m/z* ratios [24]. As the major cause of hepatocellular carcinoma, cirrhosis can be viewed as a key intermediate stage pathologically between a normal state and a state with hepatocellular carcinoma.

*Colorectal* (CRC) data set consists of 48 control and 64 cancer spectra across 16,331 *m/z* values [17], which are selected from the raw data with 65, 400 *m/z* values profiled by MALDI-TOF technologies to cover a range from 0.96 to 11.16 kDa; *HCC* data set is a binary SELDI-QqTOF proteomic data with total 358 spectra that consisting of 181 controls and 176 cancers across 6,107 *m/z* ratios, which are selected from about 340,000 *m/z* values through a binning procedure for original mass spectra [23]. As a well-known benchmark data, *Ovarian-qaqc* data consist of 95 controls and 121 ovarian cancers across 15,000 *m/z* values, which is a high-resolution serum proteomics data produced by SELDI-TOF profiling [20]. *Toxpath* data were generated from a toxicoproteomics experiment to conduct serum proteomic diagnosis for doxorubicin-induced cardiotoxicity by Petricoin et al [25]. This data set has 115 mass spectra consisting of 28 normal, 43 potential normal, 34 cardiotoxicities, and 10 potential cardiotoxicities, across 7,105 *m/z* values, which were obtained by a binning procedure from ~350,000 *m/z* values in the raw data.

*Ovarian-qaqc*data. The baseline for each profile was estimated within multiple shifted windows of widths 200

*m/z*, and the spline approximation was employed to predict the baseline. The mass spectra were further smoothed using the '

*lowess*' method, and normalized by standardizing the area under the curve (AUC) to the group median [26]. Alternatively, we only conducted the baseline correction, normalization and smoothing for the

*HCC*and

*Cirrhosis, HCC*and

*ToxPath*data (The smoothing method is selected as a different '

*least-square polynomial'*algorithm) [25, 26]. We did not conduct our own preprocessing for the

*Colorectal*data because it was preprocessed data [17]. Table 1 sketches the basic information about the five mass spectra data.

Benchmark proteomic data

Data | #Feature | #Sample | Platform |
---|---|---|---|

| 23846 | 72 controls + 78 HCCs + 51 cirrhosis | MALDI-TOF |

| 16331 | 48 controls + 64 cancers | MALDI-TOF |

| 6107 | 181 controls +176 cancers | SELDI-QqTOF |

| 15000 | 95 controls + 121 cancers | SELDI-TOF |

| 7105 | 28 normals + 43 potential normals + 34 cardiotoxicities + 10 potential cardiotoxicities | SELDI-QqTOF |

### The state-of-the-art comparison algorithms in proteomic diagnosis

We compare our DCA-SVM based profile biomarker diagnosis with following state-of-the-arts in this work. They include a partial least square (PLS) based linear logistic discriminant analysis (PLS-LLD) [27, 28], standard SVM [21], a SVM combining with principal component analysis: PCA-SVM [8], and a SVM with input-space feature selection: *fs*-SVM.

These comparison classifiers can be categorized into three groups, i.e., The group 1 only consists of standard SVM itself; The group 2 consists of those classifiers integrating SVM with input space and subspace feature selection methods respectively, i.e., PCA-SVM and *fs*-SVM; The group 3 consists of a non-SVM classifier, which employs partial least square (PLS) to conduct dimension reduction for linear logistic discriminant analysis [27, 28]. The reason we select PLS-LLD classifier is that it generally outperforms the other similar non-SVM (e.g., PCA-LDA) methods according to our implementations and Sampson et al 's work [29].

It is noted that we employ two different input-space methods: *t-test* and *anona1 (one-way ANOVA)* in *fs*-SVM to conduct feature selection for binary and multi-class data respectively [30]. Since serum proteomics data usually follow or approximately follow a normal distribution after normalization, it is reasonable to use a two-sample *t-test* to rank each feature under a binary case. For multi-class data such as *Cirrhosis* and *Toxpath*, we use one-way *ANOVA (anova1)* to identify its statistically significant features [30]. As such, we select a feature set including all features with *p-values < 0.05* under *the t-test* and *anova1* for each data. Moreover, since the PLS-LLD classifier involved matrix inverse calculation, which is notorious for its high computing demand for a large matrix (e.g., a 5,000 × 5,000 matrix), we only pick 2000 top-ranked features from for this method to avoid large computing overhead.

#### Kernel selection, cross validation, and parameter setting

It is noted that we employ the *'linear'* kernel $k\left(x,y\right)=\left(x\u2022y\right)$ in all SVM-related classifiers for its efficiency in omics data classification, rather than nonlinear kernels (e.g., Gaussian kernels). In our previous work, we actually have pointed out that nonlinear kernels (e.g., Gaussian kernels) would lead to overfitting for gene expression and proteomics data [6, 8]. Although Gaussian kernels are quite popular in serum proteomics diagnosis, it would give deceptive diagnosis due to overfitting [6]. In fact, we will show serum proteomics data diagnosis is a linear separable problem, for which a linear kernel should be the optimal kernel selection in next section.

To avoid potential biases from presetting training/test data partition on classification, we employ the *k*-fold (k = 5) cross-validation in our experiments to evaluate the five classifiers' performance for all data sets instead of the independent test set approach. In the 5-fold cross-validation, proteomic samples are randomly partitioned into k = 5 folds equally, k = 4 folds are used as training data each time, the fold left is used for evaluation. Such a process is repeated k = 5 times. In addition to choosing the first ten PLS components in the PLS-LLD classifier, we uniformly set the transform level *J = 7*; cutoff *τ = 2*; and apply the first loading vector based detail coefficient matrix reconstruction in DCA for all data sets for the convenience of comparison, though these parameter setting may not be optimal.

#### Diagnostic performance measures

Before we demonstrate our profile biomarker approach's advantages. We introduce several key diagnosis performance measures, which are diagnostic accuracy, sensitivity, specificity and positive predication ratios, as follows. The diagnostic accuracy is the ratio of the correctly classified test samples over total test samples. The sensitivity, specificity, and positive predication ratio are defined as the rates *TP/(TP+FN), TN/(TN+FP)*, and *TP/(FP+TP)* respectively, where *TP (TN)* is the number of positive (negative) targets (a positive (negative) target is a proteomic sample with '+1' ('-1') label) correctly diagnosed and *FP (FN)* is the number of negative (positive) targets incorrectly diagnosed by the classifier (e.g., SVM). It is noted that the sensitivity, specificity, and positive predication ratio for multiclass data *Cirrhosis* and *Toxpath* are obtained by treating them as a corresponding binary data. For instance, we group 78 *HCC* and 51 c*irrhosis* samples into a same class type.

For example, DCA-SVM achieves 99.52% (sensitivity: 100%, specificity: 99.17%), 100% (sensitivity: 100%, specificity: 100%), and 99.44% (sensitivity: 98.00%, specificity: 100%) diagnostic accuracies on the *Ovarian-qaqc, Colorectal* and *HCC* data respectively. However, the SVM classifier only attains corresponding 97.68% (sensitivity: 96.78%, specificity: 98.40%), 96.48% (sensitivity: 96.92%, specificity: 95.78%), 87.93% (sensitivity: 90.32%, specificity: 85.62%) diagnostic accuracies respectively for these three data sets.

Such a consistently leading performance is highlighted further in multiclass phenotype diagnosis. Our DCA-SVM algorithm reaches 97.50%, 99.01% diagnostic rates for *Toxpath* and *Cirrhosis* data respectively. However, the SVM classifier can only achieve 75.80% and 88.06% diagnosis for the same data sets respectively.

Although the input-space or subspace methods may boost diagnosis sometimes for binary-type data set (e.g., for *HCC* data PCA-SVM, *fs*-SVM attains 93.56% and 90.18% diagnosis which are higher than the 87.93% diagnostic ration from the SVM classifier), they seem not be able to increase a SVM classifier's diagnosis and generation abilities significantly, especially for multiclass data. For instance, the *fs*-SVM and PCA-SVM both have lower or the same level diagnosis than the original SVM without feature selection on *Toxpath* and *Cirrhosis* data. This may suggest the selected features' unpredictable impacts on serum proteomics diagnosis due to the input and subspace feature selection methods' limitations in de-noising and latent data characteristics capturing.

In contrast to the proposed DCA-SVM algorithm, all the comparison algorithms including PLS-LLD, which achieves slightly better diagnosis than SVM, PCA-SVM, and *fs*-SVM, shows high-level oscillations in diagnosis like the others, across different data. It is noteworthy that the high-level oscillations in diagnosis is further highlighted by corresponding large standard deviation values in diagnosis from those classifiers in Figure 3, where DCA-SVM demonstrates its good stability and generalization for its smallest standard deviation values across all the data sets.

We have to point out that such an excellent performance is because DCA forces the SVM hyperplane construction to rely on the both latent and global data characteristics in a de-noised feature space under a linear kernel, which contributes to a robust and consistent high-accuracy diagnosis. Such consistent performance applies all five data sets, which prevents from any possible overfitting possibility. On the other hand, just as we pointed out in our previous work, overfitting always happens on nonlinear kernels (e.g., Gaussian kernels) in omics data classification [6, 8].

#### A potential solution to overcome the data reproducibility

*fs*-SVM achieves 96.48% diagnosis for the

*Colorectal*data but only 70.47% for the

*Toxpath*data. To further demonstrate DCA'superiority in serum proteomics data diagnosis, we compare DCA-SVM results with those previous results obtained for these data sets in the literature as follows.

For *Colorectal* data, a 97.5% diagnosis accuracy with sensitivity 98.4% and specificity 95.8% were attained under 5-fold cross-validation in [17], where a wavelet transform is directly applied to each proteomic sample by applying Kolmogorov-Smirnov (KS) and Mann-Whitney (MW) tests to the wavelet coefficients before calling a standard SVM classifier [30]. However, our DCA-SVM achieves 100% diagnosis accuracy with sensitivity 100% and specificity 100%. It is worthwhile to point out that our comparison algorithms: *fs*-SVM and PLS-LLD have attained 96.48% (sensitivity: 96.92%, specificity: 95.78%), and 97.31% (sensitivity: 96.00%, specificity: 98.46%) diagnosis accuracies with very general feature selection under 5-fold CV ([14] uses a double CV consisting of 5-fold CV and leave-one-out CV).

For *HCC* data, a ~90%+ diagnosis accuracy with sensitivity 91% and specificity 92% is achieved by a particle swarm optimization based support vector machines (PSO-SVM) with baseline selection under a 10-fold cross-validation [23]. Instead, our DCA-SVM achieves 99.44% diagnosis accuracy (sensitivity: 99.44%, specificity: 99.44%) under 5-fold CV. In fact, all comparison algorithms expect SVM achieves same or high level performance than the previous PSO-SVM approach.

For *Ovarian-qaqc* data, our DCA-SVM achieves a 99.53% clinical-level diagnosis accuracy with sensitivity 98.95% and specificity 100%, which is better than the original diagnosis level obtained in [23] and all the other peers; For *Cirrhosis* data, Ressom *et al* partitioned this three-class data into two binary data sets and proposed a novel hybrid ant colony optimization based support vector machines (ACO-SVM) to achieve 94% and 100% specificity to distinguish hepatocellular carcinoma (HCC) from *Cirrhosis* [24]. There was no result available to distinguish normal, HCC, and cirrhosis in a multiclass diagnostic way. However, our proposed DCA-SVM has achieved 99.01% diagnosis accuracy for this multi-class data sets; The DCA-SVM achieves a rivaling clinical diagnosis accuracy 97.5% for the *Toxpath* data, which is a subset of the original data with 203 samples in [25] (we remove the 88 samples whose class-type is 'unknown' to avoid ambiguity in diagnosis).

It is noted that those algorithms applied to these data sets are generally individualized methods designed for a specific proteomics data. However, our proposed derivative component analysis based classifier (DCA-SVM) can apply to all data sets generated from different experiments and profiling technologies with rival-clinical diagnosis. Moreover, since DCA outputs a same-dimensional meta-data for each input proteomics data, it seems to be able to provide a potential profile-biomarker approach to overcome the data reproducibility issue by viewing the meta data as a uniform profile-biomarker by employing DCA-SVM to achieve rivaling-clinical diagnosis. To some degree, DCA and DCA-SVM show some promising to use a profile-biomarker way to resolve such a problem for its latent data characteristics extraction and exceptional diagnosis.

### Serum proteomics data are linearly separable

Our DCA-SVM algorithm's rivaling clinical level performance may suggest that serum proteomic data classification can be a linearly separable problem under appropriate feature selection. Such a proposition would provide a direct theoretical support to clarify some doubts about the nonlinearity in serum proteomics data may prevent it from complex disease diagnosis clinical routine [3, 5, 17], and suggest feasibility to conduct disease phenotype discrimination by using few biomarkers. In other words, if serum proteomics data are linearly separable, then, using biomarker patterns can guarantee disease phenotype discrimination, which is a key in early cancer discovery. Otherwise, seeking biomarker patterns only have a partial meaning if serum proteomics data are linearly non-separable or nonlinear because these biomarkers cannot attain 100% or rival clinical (e.g., 99%) disease phenotype separation. Moreover, serum proteomics data are linearly separable indicates '*linear* ' kernels rather than nonlinear ones would be optimal one for SVM in disease diagnosis. We sketch the definition of a linear separable problem as follows.

#### Linearly separable problem

A linearly separable problem can be simply described as follows. Given $P={\left[{x}_{1},{x}_{2},\cdots {x}_{N}\right]}^{T},$$Q={\left[{y}_{1},{y}_{2},\cdots {y}_{M}\right]}^{T},$${x}_{i}{y}_{j}\in {\Re}^{n},$$i=1,2,\cdots N,$$j=1,2,\cdots M,$ if there exists a hyperplane *H:* ${w}^{T}v+b=0,$$w,v,\in {\Re}^{n},$$b\in \Re ,$ such that $\forall x\in P,$$\forall y\in Q,$${w}^{T}x+b>0$ and ${w}^{T}y+b<0,$ then *P* and *Q* are linearly separable data, i.e. classifying P and Q is a linearly separable problem. In other words, it is equivalent to mapping entries in *P* and *Q* to two different types of labels (e.g., +1 and -1) respectively. Such a definition can be extended similarly to more than two sets, e.g., *P*_{
1
}*; P*_{
2
}...*P*_{
m
}, *m ≥ 2*, which is equivalent to mapping the *m* sets to the labels *1,2,...m* respectively.

It's clear to see that binary and multiclass SVMs by nature are linear separable test methods for its optimal hyperplane construction. However, due to the fact that serum proteomic profiles are noisy data with redundant information, it is rather difficult to draw a conclusion that they are linearly separable data because of its relatively low classification accuracies from most SVM classifier.

However, the DCA-SVM's exceptional performance reaches 99.53% 99.44%, 100%, for *Ovarian-qaqc, HCC*, and *Colorectal*, respectively, which strongly demonstrates they are linearly separable data. Although DCA-SVM only achieves 97.50% and 99.01% for *Toxpath* and *Cirrhosis* respectively, which are much better than those of the state-of-the-arts, we still believe the performances indicate these serum proteomic data are linearly separable, considering possible factors to lead to small misclassifications such as complexities of multi-class SVM hyperplane construction, possible numerical artifacts in SVM algorithm implementations, and small likelihoods that the SVM decision function may not provide a deterministic answer [21]. Thus, DCA-SVM disease classification results demonstrate that these high-dimensional data are actually linearly separable in a de-noised feature space when their latent data characteristics are extracted by DCA. Alternatively, it means the linear kernel is the optimal kernel for SVM.

### DCA-MRAK: a DCA-induced biomarker discovery

Motivated by DCA-SVM's exceptional performance, we present a DCA-induced biomarker discovery algorithm: DCA-MARK to further validate the linear separability of serum proteomics data, where each biomarker can be viewed as a statistically significant feature with respect to the others [30]. That is, we demonstrate a serum proteomic data set 's linear separability by employing the few biomarkers discovered from its meta data obtained from DCA. We will demonstrate that these biomarkers from DCA-MARK can easily separate disease phenotype completely for high-dimensional proteomics data. To the best of our knowledge, there is no similar result available in the previous research. The *DCA-MARK* can be sketched as follows.

1). Given an input dataset $X\in {\Re}^{n\times p},$ we seek the biomarkers by looking at its meta data ${X}^{*}$ from DCA through scoring and ranking each feature in ${X}^{*}$ by using the *t*-statistic for the binary data and *F*-statistic for the multiclass data [30].

2). Given a feature in a binary-class dataset $x={x}_{1}\cdots {x}_{{n}_{1}+1}\cdots {y}_{{n}_{1}+{n}_{2}}$ in ${X}^{*},$ the *t-statistic* is calculated as $t=|\overline{x}-\overline{y}|/\sqrt{{s}_{x}^{2}/{n}_{1}+{s}_{y}^{2}/{n}_{2}},$ where $\overline{x},\overline{y},$${s}_{x}^{2},{s}_{y}^{2}$ are the mean and variance values of the two classes of entries in the feature *x*. In practice, we can employ the pooled variance estimation to calculate a same variance for two types of entries as ${s}_{p}^{2}=\left(\left({n}_{1}-1\right){s}_{x}^{2}+\left({n}_{2}-1\right){s}_{y}^{2}\right)/\left({n}_{1}+{n}_{2}-2\right).$

3). Given a feature in a multi-class dataset with *k > 2* classes, the *F-statistic* is calculated as $F={\sum}_{j=1}^{k}{{n}_{j}/\left(\overline{{x}_{j}^{*}}-\overline{\overline{{x}^{*}}}\right)}^{2}/(k-1))/{\sum}_{j=1}^{k}\left({n}_{j}-1\right){s}_{j}^{2}/\left({n}_{T}-k\right)),$ where ${n}_{j}$ is the sample size, parameters ${x}_{j}^{*}$ and ${s}_{j}^{2}$ are the sample mean and sample variance for the *j-th* class. $\overline{\overline{{x}^{*}}}={\sum}_{j=1}^{k}{\sum}_{i=1}^{{n}_{j}}{x}_{ij}^{*}/{n}_{T}$is the overall sample mean where ${x}_{ij}^{*}$ is the expression value of *i-th* observation for the class *j* and ${n}_{T}={\sum}_{j=1}^{T}{n}_{j}$ is the total sample size for the *k* groups.

4). The biomarkers are the top-ranked features with the largest statistic values or the smallest *p-values*, i.e. we pick the three top-scored biomarkers for the sake of 3-dimensional visualization convenience.

*HCC*data, which is generated by high resolution mass spectral SELDI-Qq-TOF platform, by the top-ranked biomarkers (peaks) at 2534.2, 2584.3, and 6486.2

*m/z*ratios, where each dot represents a sample (a patient with HCC or a healthy subject). It is clear that we achieve linear separability for this data by using only three biomarkers. It is also interesting to see that two biomarkers are from downstream

*m/z*ratios, which were believed to be more sensitive to detect phenotype information than those from upstream

*m/z*ratios [24].

Such a separation actual fits to the linearly separable case for an SVM classifier. Thus, it is quite easy to identify a hyperplane to separate two classes phenotypes completely. For example, we run SVM for the three biomarkers for the total 357 samples and achieve 100% classification accuracy (sensitivity: 100%, specificity: 100%). Such a result demonstrates a strong advantages in phenotype discrimination over the previous work [17, 23, 24], just as we pointed out before, which employed quite complicate evolutionary algorithm (PSO-SVM) to collect a set of informative peaks and achieved 90%+ diagnosis accuracy under a 5-fold cross validation [23].

Moreover, we select three top-ranked biomarkers at 1668.99, 5907.73, 5907.13 m/z ratios for the *Cirrhosis* dataset, which is a three-class high-resolution MALDI-TOF proteomic profile with 23,846 features [24]. In addition to demonstrating the linear separability, the phenotype separations provided by the three biomarkers give very meaningful biological information. The SE plot in Figure 5 shows the three clearly separable clusters, where Cirrhosis cluster with 51 samples (blue) have closer spatial distances to the HCC cluster 78 samples (red) than the normal cluster with 72 samples (yellow). Such spatial distances demonstrated by our biomarkers are actually consistent to their pathological distances: Cirrhosis is the middle stage to hepatocellular carcinoma (HCC) for a healthy subject [31]. To the best of our knowledge, no previous work achieved the similar results.

## Discussion

In this study, we propose a novel feature selection algorithm: derivative component analysis (DCA) to overcome the weakness of the traditional feature selection methods. Unlike the traditional methods, the DCA focuses on latent data characteristics gleaning and de-noising by analyzing derivative data components for input data to calculate a same dimensional meta-data.

We further embed derivative component analysis into support vector machines to achieve rivaling clinical level phenotype discrimination for five benchmark serum proteomics data by comparing it with the other state-of-the-arts. The DCA-SVM 's exceptional classification accuracies suggest the serum proteomics data's linear separability and further inspire DCA-MARK, a DCA-induced biomarker discovery approach, which in turn demonstrate high-dimensional proteomics data 's linear separability with few biomarkers. Moreover, derivative component analysis (DCA) demonstrate a potential to resolve data reproducibility problem of serum proteomics by viewing each input data's meta-data as a profile biomarker by employing DCA-SVM to achieve clinical level disease diagnosis, because of DCA's true signal extraction for input proteomics data.

Such profile biomarker diagnosis approach actually demonstrates strong advantages over the existing biomarker discovery oriented diagnosis by treating input proteomic data as a profile biomarker. The systems approach seems to fit the "personalized diagnostics" better [32], because it can be difficult both biologically and computationally to achieve a clinical level diagnostics for those complex diseases like cancer, in which thousands genes can be involved, based on several differentially expressed proteins, especially when the source data suffer from the reproducibility issue.

Our experimental results demonstrated that the DCA's parametric tuning works efficiently though they may not be the optimal ones theoretically. It is possible to seek optimally parametric settings in derivative component analysis for each proteomic data from an information entropy analysis or Monte Carlo simulation standing point [18]. However, we are not sure such computing demand way is practically worthwhile because the clinical level diagnostics are already attained under our current parametric tuning.

## Conclusions

Our DCA provides an alternative feature selection by implicitly extracting useful data characteristics whiling maintaining the data 's original dimensionality. It suggests that subtle data characteristics gleaning and de-noising may be more important in proteomics data feature selection and following phenotype discrimination. It is worthwhile to point out that DCA-related techniques developed can be also applied to gene expression data smoothly. Although we are quite optimistic to see that our DCA-MARK can capture meaningful peaks from low-weight sera from different data sets, there is still an urgent need to verify and compare these biomarkers with the previous ones to seek potential pathological meaning and clinical application. Although derivative component analysis does show a potential to conquer the reproducibility problem of serum proteomics, a future concrete proteomics clinical test is still needed to explore such a potential. Although we are quite optimistic to see that our DCA-SVM based diagnosis will be a potential candidate to achieve a clinical disease diagnosis in proteomics by conquering the reproducibility problem, rigorous proteomics clinical tests are needed urgently to explore such a potential and validate its clinical effectiveness. In our ongoing work, we are working with pathologists to investigate extending the profile-biomarker diagnosis approach to TCGA and RNA-Seq data besides genes expression array analysis [33, 34].

## Declarations

### Acknowledgements

Special thanks are also due to anonymous reviewers for their valuable comments and insightful suggestions.

**Declarations**

Publication for this article has been funded by Fordham University. The author would like to thank Fordham University for awarding him faculty research fellowship to support this project.

This article has been published as part of *BMC Medical Genomics* Volume 7 Supplement 1, 2014: Selected articles from the 3rd Translational Bioinformatics Conference (TBC/ISCB-Asia 2013). The full contents of the supplement are available online at http://www.biomedcentral.com/bmcmedgenomics/supplements/7/S1.

## Authors’ Affiliations

## References

- Rath T, Hage L, Kügler M, Menendez Menendez K, Zachoval R, Naehrlich L, Schulz R, Roderfeld M, Roeb E: Serum Proteome Profiling Identifies Novel and Powerful Markers of Cystic Fibrosis Liver Disease. PLoS ONE. 2013, 8 (3): e58955-10.1371/journal.pone.0058955.PubMedPubMed CentralView ArticleGoogle Scholar
- Coombes KR, Morris JS, Hu J, Edmonson SR, Baggerly KA: Serum proteomics profiling - a young technology begins to mature. Nature biotechnology. 2005, 23 (3): 291-2. 10.1038/nbt0305-291.PubMedView ArticleGoogle Scholar
- Ioannidis JP, Khoury MJ: Improving Validation Practices in ''Omics'' Research. Science. 2011, 334: 1230-2. 10.1126/science.1211811.PubMedView ArticleGoogle Scholar
- Hüttenhain R, Soste M, Selevsek N, Röst H, Sethi A, Carapito C, Farrah T, Deutsch EW, Kusebauch U, Moritz RL, Niméus-Malmström E, Rinner O, Aebersold R: Reproducible Quantification of Cancer-Associated Proteins in Body Fluids Using Targeted Proteomics. Sci Transl Med 4. 2012, 142ra94-Google Scholar
- Levin VA, Panchabhai SC, Shen L, Kornblau SM, Qiu Y, Baggerly KA: Different Changes in Protein and Phosphoprotein Levels Result from Serum Starvation of High-Grade Glioma and Adenocarcinoma Cell Lines. J Proteome Res. 2010, 9 (1): 179-91. 10.1021/pr900392b.PubMedPubMed CentralView ArticleGoogle Scholar
- Han H: Nonnegative principal component analysis for mass spectral serum profiles and biomarker discovery. BMC Bioinformatics. 2010, 11 (Suppl 1): S1-10.1186/1471-2105-11-S1-S1.PubMedPubMed CentralView ArticleGoogle Scholar
- Han H: A high performance profile-biomarker diagnosis for mass spectral profiles. BMC Syst Biol. 2011, 5 (Suppl 2): S5-10.1186/1752-0509-5-S2-S5.PubMedPubMed CentralView ArticleGoogle Scholar
- Han X: Nonnegative Principal component Analysis for Cancer Molecular Pattern Discovery. IEEE/ACM Transaction of Computational Biology and Bioinformatics. 2010, 7 (3): 537-549.View ArticleGoogle Scholar
- Hilario M, and Kalousis A: Approaches to dimensionality reduction in proteomic biomarker studies. briefings in bioinformatics. 2008, 2 (9): 102-118.Google Scholar
- Han H, and Li X: Multi-resolution independent component analysis for high-performance tumor classification and biomarker discovery. BMC Bioinformatics. 2011, 12 (S1): S7-View ArticleGoogle Scholar
- Jolliffe I: Principal component analysis. 2002, Springer, New YorkGoogle Scholar
- Hyvärinen A: Fast and robust fixed-point algorithms for independent component analysis. IEEE Transactions on Neural Networks. 1999, 10 (3): 626-634. 10.1109/72.761722.PubMedView ArticleGoogle Scholar
- Brunet JP, Tamayo P, Golub TR, Mesirov JP: Molecular pattern discovery using matrix factorization. Proc Natl Acad Sci USA. 2004, 12 (101): 4164-4169.View ArticleGoogle Scholar
- Hoyer P: Non-negativematrix factorization with sparseness constraints. Journal of Machine Learning Research. 2004, 5: 1457-1469.Google Scholar
- Li P, Church KW: Very sparse random projections. Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining. 2006, 287-296.View ArticleGoogle Scholar
- d'Aspremont A, Chaoui L, Jordan M, Lanckriet G: A direct formulation for sparse PCA using semidefinite programming. SIAM Review. 2007, 49 (3): 434-448. 10.1137/050645506.View ArticleGoogle Scholar
- Alexandrov T, Decker J, Mertens B, Deelder AM, Tollenaar RA, Maass P, Thiele H: Biomarker discovery in MALDI-TOF serum protein profiles using discrete wavelet transformation. Bioinformatics. 2009, 25 (5): 643-649. 10.1093/bioinformatics/btn662.PubMedPubMed CentralView ArticleGoogle Scholar
- Mallat S: A wavelet tour of signal processing. 1999, Acad. Press, CA, USAGoogle Scholar
- Kapur JN, Kesevan HK: Entropy optimization principles with applications. 1992, Toronto: Academic PressView ArticleGoogle Scholar
- Conrads TP, Fusaro VA, Ross S, Johann D, Rajapakse V, Hitt BA, Steinberg SM, Kohn EC, Fishman DA, Whitely G, Barrett JC, Liotta LA, Petricoin EF, Veenstra TD: High-resolution serum proteomic features for ovarian detection. Endocrine-Related Cancer. 2004, 11 (2): 163-178. 10.1677/erc.0.0110163.PubMedView ArticleGoogle Scholar
- Vapnik V: Statistical Learning Theory. 1998, John Wiley, New YorkGoogle Scholar
- Hus C, Lin C: A Comparison of Methods for Multi-class Support Vector Machines. IEEE Transactions on Neural Networks. 2002, 13 (2): 415-425. 10.1109/72.991427.View ArticleGoogle Scholar
- Ressom HW, Varghese RS, Abdel-Hamid M, Eissa SA, Saha D, Goldman L, Petricoin EF, Conrads TP, Veenstra TD, Loffredo CA, Goldman R: Analysis of mass spectral serum profiles for biomarker selection. Bioinformatics. 2005, 21 (21): 4039-4045. 10.1093/bioinformatics/bti670.PubMedView ArticleGoogle Scholar
- Ressom HW, Varghese RS, Drake SK, Hortin GL, Abdel-Hamid M, Loffredo CA, Goldman R: Peak selection from MALDI-TOF mass spectra using ant colony optimization. Bioinformatics. 2007, 23 (5): 619-626. 10.1093/bioinformatics/btl678.PubMedView ArticleGoogle Scholar
- Petricoin EF, Rajapaske V, Herman EH, Arekani AM, Ross S, Johann D, Knapton A, Zhang J, Hitt BA, Conrads TP, Veenstra TD, Liotta LA, Sistare FD: Toxicoproteomics: serum proteomic pattern diagnostics for early detection of drug induced. Toxicologic Pathology. 2004, 32 (Suppl 1): 1-9.Google Scholar
- Callister SJ, Barry RC, Adkins JN, Johnson ET, Qian WJ, Webb-Robertson BJ, Smith RD, Lipton MS: Normalization approaches for removing systematic biases associated with mass spectrometry and label-free proteomics. Journal Proteome Res. 2006, 5 (2): 277-86. 10.1021/pr050300l.View ArticleGoogle Scholar
- Nguyen D, and Rocke D: Tumor classification by partial least squares using microarray gene expression data. Bioinformatics. 2002, 18: 39-50. 10.1093/bioinformatics/18.1.39.PubMedView ArticleGoogle Scholar
- Fort G, and Lambert-Lacroix S: Classification using partial least squares with penalized logistic regression. Bioinformatics. 2005, 21 (7): 1104-1111. 10.1093/bioinformatics/bti114.PubMedView ArticleGoogle Scholar
- Sampson DL, Parker TJ, Upton Z, Hurst CP: A Comparison of Methods for Classifying Clinical Samples Based on Proteomics Data: A Case Study for Statistical and Machine Learning Approaches. PLoS ONE. 2011, 6 (9): e24973-10.1371/journal.pone.0024973.PubMedPubMed CentralView ArticleGoogle Scholar
- Jung K: Statistical methods for proteomics. Methods Mol Biol. 2010, 620: 497-507. 10.1007/978-1-60761-580-4_18.PubMedView ArticleGoogle Scholar
- Zhang L, Guo Y, Li B, Qu J, Zang C, Li F, Wang Y, Pang H, Li S, Liu Q: Identification of biomarkers for hepatocellular carcinoma using network-based bioinformatics methods. Eur J Med Res. 2013, 18: 35-10.1186/2047-783X-18-35.PubMedPubMed CentralView ArticleGoogle Scholar
- Richmond TD: The current status and future potential of personalized diagnostics: Streamlining a customized process. Biotechnol Annu Rev. 2008, 14: 411-22.PubMedView ArticleGoogle Scholar
- The Cancer Genome Atlas Network: Comprehensive molecular portraits of human breast tumours. Nature. 2012, 490 (4): 61-70.PubMed CentralGoogle Scholar
- Wang Z, Gerstein M, Snyder M: RNA-Seq: a revolutionary tool for transcriptomics. Nat Rev Genet. 2009, 10 (1): 57-63. 10.1038/nrg2484.PubMedPubMed CentralView ArticleGoogle Scholar

## Copyright

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.