- Open Access
Differentially private distributed logistic regression using private and public data
© Ji et al.; licensee BioMed Central Ltd. 2014
- Published: 8 May 2014
Privacy protecting is an important issue in medical informatics and differential privacy is a state-of-the-art framework for data privacy research. Differential privacy offers provable privacy against attackers who have auxiliary information, and can be applied to data mining models (for example, logistic regression). However, differentially private methods sometimes introduce too much noise and make outputs less useful. Given available public data in medical research (e.g. from patients who sign open-consent agreements), we can design algorithms that use both public and private data sets to decrease the amount of noise that is introduced.
In this paper, we modify the update step in Newton-Raphson method to propose a differentially private distributed logistic regression model based on both public and private data.
Experiments and results
We try our algorithm on three different data sets, and show its advantage over: (1) a logistic regression model based solely on public data, and (2) a differentially private distributed logistic regression model based on private data under various scenarios.
Logistic regression models built with our new algorithm based on both private and public datasets demonstrate better utility than models that trained on private or public datasets alone without sacrificing the rigorous privacy guarantee.
- Logistic Regression Model
- Newton Method
- Public Data
- Hessian Matrix
- Differential Privacy
Data about individuals are being collected at an unprecedented speed, which brings new opportunities for scientific discovery and healthcare quality improvement. In the meantime, there is increasing concern about people's privacy and inappropriate disclosure of sensitive information . This problem is especially challenging in biomedicine , where information sharing is one of the biggest pillars to facilitate meaningful analysis of complex medical data. For example, classifying complex or rare patterns in clinical and genomic data requires the availability of a large, labeled patient set, which needs to be obtained from multiple institutions .
Any data access mechanism involves a tradeoff between the privacy risk and the data utility. In biomedicine, data custodians can change the content of the data to make more difficult for attackers to re-identify individuals (k-anonymity , l-divergence , t-closeness , etc.) or can perturb the outputs of a query result to ensure "indistinguishability" of individuals (i.e., count queries satisfying differential privacy ). Because differential privacy  provides a provable guarantee and is immune to attacks with auxiliary information, it is acknowledged as a state-of-the-art privacy definition . The performance (i.e., privacy and utility) of a differentially private method is highly dependent on the nature of the application and the capability of the protection mechanism. To meet the need of different applications, many customized differentially private methods, including decision trees , logistic regression , principal components analysis , multi-class Gaussian classifiers , have been developed. There are several recent efforts in integrating the differential privacy framework into the system design and case studies for statistical health information release , . However, due to an inherited challenge of differential privacy in considering the entire sample space, the level of data perturbation often increases too quickly when the privacy assurance becomes stronger, which ends up adding too much noise , i.e., producing useless, albeit protected, data.
We believe the situation can be alleviated in an environment where both public and private data sets for the same study are available for analysis. This is useful in biomedical research (e.g., randomized clinical trial), where some patients are willing to sign an open-consent agreement to make their data (publicly) available for research, while other patients prefer to limit disclosure to a single institution. Our idea is to develop hybrid data mining models using both public and private data sets in a differentially private and distributed manner to achieve improved utility of the disclosed data. We will focus on the logistic regression model, which is one of the most popular approaches in biomedicine, to develop a distributed and privacy preserving solution in the healthcare context.
Our model is closely related to Grid Logistic Regression, a model developed by Wu . Their model is based on a distributed Newton-Raphson algorithm, however, it does not consider privacy risk during the exchange of aggregated statistics among participants. A recent work by Wu et al.  discusses institutional privacy of distributed logistic regression and introduces a secure-sum based approach to protect aggregated statistics using a trusted server, but it does not meet the differential privacy criteria. The underlying intuition of our model is close to Elkan's work  to represent a confidential database via importance weighting elements of a public database for general data mining purpose, but his approach is also not differentially private.
There are also some previous works on differentially private distributed learning. For example, Pathak, Rane, Raj et al.  suggested running local models on each private data set and aggregating estimated parameters. This approach is different from ours in two aspects. First, their model does not take public data into consideration. Second, they only prove that the final outputs (the aggregated parameters) are differentially private, while there is no guarantee that the intermediary outputs from individual private data sets (which need to be shared during the process) are differentially private. In contrast, our model ensures differential privacy for all steps. Rajkumar and Agarwal  recently proposed a distributed differentially private stochastic gradient descent algorithm, which also differs from ours as follows: (1) their approach used only private data; (2) their approach is -differentially private (weaker) while ours is -differentially private (stronger).
In this paper, we introduce a new distributed logistic regression model that runs on many data sets, e.g., both public and private ones. It treats these two kinds of data sets differently: it leverages public data sets to improve utility while protecting the private data sets. Background section introduces some background knowledge and methodology section elaborates on details of our method. In experiments and results section, we compare our model to other approaches, and explore the impact of different settings (i.e., the fraction of data that are public, the number of distributed private data sets, and impact of the regularization parameter) on the final model. Finally, discussions and conclusions are drawn.
In this Section, we will briefly review techniques related to this article.
Differential privacy (DP) is a privacy definition proposed by Dwork, Kenthapadi, McSherry et al , which states that any answer to a query based on a private data set should not be altered dramatically with the change of a single record in the data set.
In the following part of this paper, and always differ on at most one sample (i.e., and are "neighbors" with only one sample replaced).
Definition 1: Query function Function is a query function, if it is a projection from a data set (not a single sample, but the whole data set) to .
Many data mining models can be viewed as a query function, for example, the coefficients of the logistic regression can be seen as the projection of a data set to a real-valued vector.
Definition 2: -Differential Privacy A randomized algorithm (or mechanism) is -differentially private if for any neighbors and and for any when the following probabilities are well-defined,
where the probabilities reflect the randomness of the algorithm . Note the parameter is called privacy budget. The smaller is, the better privacy is preserved, and vice versa.
Definition 4: Laplacian mechanism  For any query function , the algorithm returns , where , satisfies differential privacy.
If two independent mechanisms are and differentially private, running them iteratively on the same data set will consume a privacy budget , which is known as the sequential composition property of differential privacy .
Typically, norm is used in calculating sensitivity and applying Laplacian mechanism, in which case the noise on different dimensions are independent. In this paper, we consider the norm instead, such that the sensitivity of penalized logistic regression parameter (i.e., the outputs of a query function) can be bounded, as proved in Corollary 2 by Chaudhuri et al . The norm has been used in previous differential private algorithms, e.g., the work of Chaudhuri  and Rajkumar .
Newton method for logistic regression
The Newton method (also known as the Newton-Raphson method ) is an iterative approach that uses gradient to find roots of a real-valued differentiable function. Since a function's extrema are also the roots of its gradient, the Newton method can also be used to find twice differentiable function's extrema. Due to its efficiency in handling convex functions, (i.e., usually only a few iterations (five or six) are needed to reach a very high precision ), the Newton method is a popular numerical approach for building a logistic regression model . Given the log-likelihood function of a logistic regression model , the Newton method approaches the maximum likelihood coefficients estimate β with the following steps,
Update and repeat the second step until converges.
Our goal is to develop a distributed logistic regression model that effectively synthesizes data (public and private) across different sites (institutions) in a differentially private manner.
Assumptions about the data set
To develop the model in the biomedical context, we will make some assumptions about the data sets. First, the number of samples in each data set is not very large. This assumption is reasonable as otherwise one data set is enough to build the model and many data sets will not bring much benefit. Second, the size of the public data set is significantly smaller than that of the private data sets. The reason is that by default biomedical data should be kept secure and private unless patients are willing to sign open-consent agreements to make their medical data available for research, which only applies to a small percentage of the total data. Third, we assume samples in different data sets follow the same predictive rule, which means must be the same across all data sets. This is necessary for constructing a distributed logistic regression model that can provide useful information in biomedical research. In practice, such assumption can be verified by checking the goodness-of-fit (e.g., Hosmer-Lemeshow test ) of the fitted local and global models without sacrificing individual privacy. Last, we assume that the distribution of samples in different data sets () are similar, although this assumption will be relaxed, as we will elaborate in the discussion section.
and the β that maximizes this log-likelihood function is the estimated parameter for the model.
Note that for the bounded sensitivity, if an intercept is introduced into the predictors, the corresponding parameter should be also regularized.
Because the logistic regression model has no analytic solutions, we need to solve it using numerical methods like the Newton algorithm (see the Background section), which involves several iterations of optimization. As the distributed logistic regression model is supposed to be trained on multiple data sets, we need to implement a privacy-preserving information exchange mechanism to transmit intermediary results across private data sets. In addition, the privacy budget is limited and it must be split across iterations. To maintain data utility, we must balance the number of iterations and the privacy budget spent on each iteration in the training process.
To use the Newton-Raphson algorithm, we need to compute the gradient and the Hessian matrix from the data sets. The simplest way to ensure differential privacy is to add Laplacian noise to the gradient and the Hessian matrix, and use the noisy version of these intermediary results to update parameters. Theoretically, the impacts of additive noise in Laplacian mechanism (see Definition 4) tend to be much smaller when the number of samples approaches infinity, as the sensitivity of the gradient and the Hessian matrix is irrelevant to the size of a data set. In reality, however, the effects of noise on the gradient and the Hessian matrix (for parameter estimation) are quite different. For example, the gradient is usually affected by the additive noise than is the Hessian matrix. This is because the gradient has linear impact on the parameter updates. However, this is not the case for the Hessian matrix. Since the inverse of the Hessian matrix is used in the update step, even a little noise in the Hessian matrix can lead to large changes on the parameters being updated. Such change can become very large when the noise destroys the Hessian matrix's positive definiteness, which implies that a global optimal solution (like the one for the log-likelihood function of a traditional logistic regression model) may not be attained. Although we can threshold the eigenvalues of the Hessian matrix to ensure positive definiteness, this method might generate useless coefficients. Therefore, the key to build a useful differentially private distributed logistic regression using the Newton-Ralphson algorithm is to reduce the noise in the Hessian matrix, especially reduce the chance of a non-positive definite Hessian matrix.
Our approach is to leverage public data sets (i.e., contributed by patients who signed the open-consent agreement). In our hybrid framework, the Hessian matrix is estimated solely using the public data set, and we use public and private data sets to compute the gradient. To estimate the Hessian matrix using only public data, we leverage the following advantages: (1) The Hessian matrix from the public data set is positive definite. Therefore, the worst case discussed above is avoided even though the absolute sample error on the Hessian matrix might be larger than the noisy Hessian matrix from individual data sets. (2) The sensitivity of Hessian matrix is when there are p features in each sample, while the sensitivity of the gradient is only . Therefore, if we use the public data to calculate the Hessian matrix, the total sensitivity can be reduced from to , which increases the accuracy of update steps.
Unlike the traditional Newton-Raphson algorithm, which iterates until convergence, our method uses a fixed number of iterations for the following reasons. First, the original Newton method stops after parameters converge but our algorithm will never terminate due to the noise added in each step. Second, the stop decision in our case cannot be accurately determined by comparing the likelihood associated with and because it is possible that the noisy performs worse than , which will never happen in the standard maximum likelihood estimation. Finally, a fixed number of iterations allows us to allocate privacy budgets easily, e.g., evenly split across iterations as in this work.
The details of our update step are illustrated in Algorithm 1, and the full model is described in Algorithm 2.
There are two some modifications that can further improve the performance. Please refer to Additional file 1, for these modifications and for the proof of differential privacy.
Modified update step in distributed logistic regression
Private dataset , public dataset , privacy budget for this iteration , coefficient of penalty , the upper bound of norm of samples , and old parameter obtained from the previous iteration.
Logistic regression parameter .
The coefficient is an adjustment, as the number of samples to obtain here is different from number of samples to get gradients below.
where are iid vectors with density
Distributed logistic regression
Private dataset , public dataset , privacy budget for this iteration , coefficient of penalty , the upper bound of norm of samples , and iteration times .
Logistic regression parameter .
2: Initialize logistic regression parameter , a vector with the same length as .
3: Given the data sets, and , use as , update with in Algorithm 1. Repeat for times
4: Output β
We will compare our algorithm with two baselines. The first is the meta analysis method. It first adds noise from distribution to logistic regression parameters learned from a private data set, where is the upper bound of predictors' norm. Then, it outputs weighted average (by number of samples in each data set) of these noisy parameters learned from private data sets. By weighting locally learned differentially private parameters, this method is similar to the method in Pathak, Rane, Raj et al. The only difference is that we make outputs from each private data be shared differentially privately (rather than transmitting encrypted partial local outputs). Our second baseline is to train a logistic regression model with only public data and neglect the private data sets. Our evaluation metric is model discrimination, i.e., the Area Under the ROC curve (AUC). We will explore the performance of all three models using various parameter values. Unless explicitly illustrated in the Figures, default values for the following parameters are set as follows: number of private data sets (3), fraction of public data in the training set (2%), privacy budget (1), number of iterations (2). The regularization strength for all three models is selected from a range [10−2,106] to maximize the expected AUC based on a 10-fold cross validation.
We used clinical data to conduct experiments, where the public and private data sets are split randomly. Each private data set contains roughly the same number of observations. We used a 60%/40% split for training and testing in all experiments.
Summary of data sets used in our experiments
Data set description
Class distribution (negative/positive)
German breast cancer
43.6% / 56.4%
4.4% / 95.6%
SEER breast Cancer
21.0% / 79.0%
Attribute description for each data set, where numerical attributes are indicated with "∗", non-binary categorical attributes were converted into binary representations through dummy coding and classification labels are shown in the last row.
Data set 1
Data set 2
Data set 3
1. Yes, 2. No.
1. Blood, 2. Urine, 3. sputum, 4. CSF
Race (25 categories)
1.Premenopausal, 2. Postmenopausal
Day of the week for collection
1. Weekday, 2. Weekend
Marital status (6 categories)
(Levels I, II, III)
Day of the week for the final result
1. Weekday, 2. Weekend
Number of nodes examined*
Number of positive nodes*
Number of positive nodes*
Recurrence free Survival time*
1.Medicare, 2. Medicaid,
3. Commercial, 4. Other
1. White, 2. Black, 3. Asian, 4. Hispanic,
Vital status recode
Pos: Alive, Neg: Died
Pos: Not a potential follow-up error, Neg: A potential follow-up error
Pos: Alive, Neg: Died
As the magnitudes of attributes have large impact on the overall sensitivity of logistic regression parameters and the gradients of the log-likelihood function, we normalized all attributes and truncated their values to [−2,2] in order to bound the impact. Note that we used the mean and the variance of public data to conduct normalization, which does not incur a privacy cost.
1) Model comparison using different parameters: In Figure 1a, we illustrated the effect of different number of private data sets on discrimination. The AUCs for our method and for the meta-analysis method drop at the same speed as the number of data sets increases, and our method results in higher AUCs in all scenarios. This is because both methods have to add noise to outputs from each private data set. Given a fixed number of observations (8,668), more noise is added when there are more private data sets (evenly split). The public data based method is not affected by this setting. In the experiments of this section, AUCs have small standard deviations (around 0.05) and the mean of results from 100 independent experiments is stable enough to represent performance of the methods (the standard deviation of the mean is around 0.005). Therefore, we only plot the mean AUCs).
Figure 1b shows the AUCs of the models given different fractions of public data. Our algorithm's performance is stable, but the AUCs of the public data based model grow quickly with increased sample size. When the fraction of public data is between 1% and 5%, our algorithm is the best.
Figure 1c shows how AUCs change at different regularization strengths ( in Algorithm 2). Our method has the most stable performance even when the regularization strength increases 100 times, or when it is 10% of the optimal value. This is very important in practice, as it is expensive to calculate the best regularization strength (i.e., have to reserve additional test data for tuning regularization strength, which will also spend some privacy budget). With the stable performance, we can hypothesize that even if the guess is away from the optimal regularization strength, our method can perform well in terms of discrimination.
Figure 1d shows the effect of the privacy budget ( in Algorithm 2) on AUCs. The method based on public data is not affected, but the other two have better performance with more budgets. Our algorithm is better than the meta analysis model in general, and outperforms the public data based method when the privacy budget is larger than 1.
Figure 1e shows how the number of iterations in distributed logistic regression algorithm ( in Algorithm 2) will affect our algorithm's performance. When there are no iterations (i.e., only locally calculated parameters are used), our method degrades to the public data based algorithm (the same starting point). Our performance is best with less than 3 iterations, and it gets worse with more iterations. There are several reasons for the degraded performance. First, our algorithm uses parameters trained on public data set as the starting point, which is expected to be close to the real one, and therefore, the necessary iterations are smaller than the traditional Newton method (starting from all zeros). Second, as noise is added to each iteration, the gain from later iterations can be masked by the increased amount of noise. Yet another reason is that our privacy budget is evenly split into a pre-determined number of iterations, and more iterations imply a large amount of total noise.
A natural question is when our method would be most useful. This is a hard question to answer in theory because our method has no guarantee of convergence (due to the added noise). However, we can answer the question empirically.
We decompose parameters into "external factors" and "controllable factors". The former corresponds to: the number of private data sets, the fraction of data that are public, and the private budget (set by the data owner), which researchers cannot control. The regularization strength and iteration numbers, however, are not external factors, as researchers can choose them.
2) Model comparison using different datasets: Our last study evaluated model performance using data of different sizes. Three biomedical datasets used in this experiment differ in the number of observations. All input parameters are set to default values. Each experiment is repeated 100 times to generate a boxplot.
We demonstrated a novel approach that combines public and private data sets to build a logistic regression model in a distributed manner. Our approach shows performance advantage over two other approaches under various conditions. There are still challenges in using our approach in practice. For example, categorical attribute values in private data sets may not appear in public data. The simplest solution is to pre-process private data sets by dropping values that rarely appear (or do not appear) in public data and that seem uncorrelated to the labels. This will not spend privacy budget as we only use public data to guide the process. It may improve a model's utility by removing some values and reducing the number of attributes. The upper bound of norm for the predictors therefore gets reduced, and so does the scale of noise.
Another limitation of our method is that we assume that all data sets follow the same (joint) distribution. However, it is possible that some data sets have sample bias, but may still follow the same learning rule. A solution for this is to use gradient descent instead of the Newton method in Algorithm 1. However, as gradient descent algorithm usually needs more iterations to get an accurate solution, and consequently it may add more noise and therefore generate less satisfactory outputs. In conclusion, we propose a new algorithm to extend the differential private framework to real world scenarios in biomedical research, where public and private data sets are available for analysis. Hybrid approaches that rigorously protect private data while leveraging public data to improve the utility show great promise to achieve "the best of both worlds" (i.e., data privacy and usefulness).
Funding for the publication of this article has partly come from NIH grants R00LM011392, U54HL108460, UL1TR0001000, AHRQ grant R01HS019913 and PCORI contracts.
This article has been published as part of BMC Medical Genomics Volume 7 Supplement 1, 2014: Selected articles from the 3rd Translational Bioinformatics Conference (TBC/ISCB-Asia 2013). The full contents of the supplement are available online at http://www.biomedcentral.com/bmcmedgenomics/supplements/7/S1.
- P Ohm: Broken promises of privacy: Responding to the surprising failure of anonymization. UCLA Law Review. 2010, 57: 1701-1776.Google Scholar
- McGraw D: Building public trust in uses of Health Insurance Portability and Accountability Act de-identified data. Journal of the American Medical Informatics Association. 2013, 20 (1): 29-34. 10.1136/amiajnl-2012-000936. JanPubMedPubMed CentralView ArticleGoogle Scholar
- Wu Y, Jiang X, Kim J, Ohno-Machado L: Grid Binary LOgistic REgression (GLORE): building shared models without sharing data. Journal of the American Medical Informatics Association. 2012, 19 (5): 758-64. 10.1136/amiajnl-2012-000862. SeptPubMedPubMed CentralView ArticleGoogle Scholar
- Sweeney L: k-anonymity: A model for protecting privacy. Int J Uncertainty Fuzziness Knowledge-based Systems. 2002, 10 (5): 557-570. 10.1142/S0218488502001648.View ArticleGoogle Scholar
- Machanavajjhala A, Kifer D, Gehrke J, Venkitasubramaniam M: l-diversity: privacy beyond k-anonymity. ACM Transactions on Knowledge Discovery from Data. 2007, 1 (1): 3-es-10.1145/1217299.1217302. MarView ArticleGoogle Scholar
- Li N, Li T, Venkatasubramanian S: t Closeness : Privacy Beyond k-Anonymity and -Diversity. ICDE 2007. 2007, 2: 106-115.Google Scholar
- Vinterbo SA, Sarwate AD, Boxwala AA: Protecting count queries in study design. Journal of the American Medical Informatics Association. 2012, 19 (5): 750-757. 10.1136/amiajnl-2011-000459.PubMedPubMed CentralView ArticleGoogle Scholar
- Dwork C, Kenthapadi K, McSherry F, Mironov I, Naor M: Our data, ourselves: Privacy via distributed noise generation. Adv In Cryptology - Eurocrypt 2006, Proc. 2006, 4004: 486-503. 10.1007/11761679_29.View ArticleGoogle Scholar
- Dwork C, Pottenger R: Toward practicing privacy. Journal of the American Medical Informatics Association. 2013, 20 (1): 102-108. 10.1136/amiajnl-2012-001047.PubMedPubMed CentralView ArticleGoogle Scholar
- Friedman A, Schuster A: Data mining with differential privacy. KDD. 2010, 493-502.Google Scholar
- Chaudhuri K, Monteleoni C: Privacy-preserving logistic regression. NIPS. 2008, 289-296.Google Scholar
- Jiang X, Ji Z, Wang S, Mohammed N, Cheng S, Ohno-Machado L: Privacy preserving data publishing through component analysis. Transactions on Data Privacy. 2013, 6 (1): 19-34.PubMedPubMed CentralGoogle Scholar
- Pathak MA, Raj B: Large margin multiclass Gaussian classification with differential privacy. PSDML. 2010, 99-112.Google Scholar
- Gardner J, Xiong L, Xiao Y, Gao J, Post AR, Jiang X, Ohno-Machado L: Share: system design and case studies for statistical health information release. Journal of the American Medical Informatics Association. 2013, 20 (1): 109-116. 10.1136/amiajnl-2012-001032.PubMedPubMed CentralView ArticleGoogle Scholar
- Mohammed N, Jiang X, Chen R, Fung BC, Ohno-Machado L: Privacy-preserving heterogeneous health data sharing. Journal of the American Medical Informatics Association. 2012,Google Scholar
- Wasserman L: Minimaxity, statistical thinking and differential privacy. Journal of Privacy and Confidentiality. 2012, 4 (1): 51-63.Google Scholar
- Wu Y, Jiang X, Ohno-machado L: Preserving Institutional Privacy in Distributed Binary Logistic Regression. AMIA Annual Symposium Proceedings. 2012, 1450-1458.Google Scholar
- Elkan C: Preserving privacy in data mining via importance weighting. Privacy and Security Issues in Data Mining and Machine Learning. 2011, 15-21.View ArticleGoogle Scholar
- Pathak MA, Rane S, Raj B: Multiparty differential privacy via aggregation of locally trained classifiers. NIPS. 2010, 1876-1884.Google Scholar
- Rajkumar A, Agarwal S: A differentially private stochastic gradient descent algorithm for multiparty classification. Journal of Machine Learning Research - Proceedings Track. 2012, 933-941.Google Scholar
- Dwork C, McSherry F, Nissim K, Smith A, Smith A: Calibrating noise to sensitivity in private data analysis. TCC. 2006, 265-284.Google Scholar
- McSherry FF: Privacy integrated queries: an extensible platform for privacy-preserving data analysis. Proceedings of the 35th SIGMOD international conference on Management of data. 2009, New York, NY, USA, 53 (9): 19-30.Google Scholar
- Macleod A: A generalization of Newton-Raphson. International Journal of Mathematical Education in Science and Technology. 1984, 15 (1): 117-120. 10.1080/0020739840150116.View ArticleGoogle Scholar
- Boyd S, Vandenberghe L: Convex optimization. 2004, Cambridge university pressView ArticleGoogle Scholar
- Minka T: A comparison of numerical optimizers for logistic regression. CMU Technical Report. 2003, 2003: 1-18.Google Scholar
- Hosmer DW, Lemeshow S: Applied logistic regression. 2000, New York: Wiley-InterscienceView ArticleGoogle Scholar
- Schumacher M, Bastert G, Bojar H, Huebner K, Olschewski M, Sauerbrei W, Schmoor C, Beyerle C, Neumann R, Rauschecker H: Randomized 2 × 2 trial evaluating hormonal treatment and the duration of chemotherapy in nodepositive breast cancer patients. german breast cancer study group. Journal of Clinical Oncology. 1994, 12 (10): 2086-2093.PubMedGoogle Scholar
- El-Kareh R, Roy C, Williams D, Poon E: Impact of automated alerts on follow-up of post-discharge microbiology results: A cluster randomized controlled trial. Journal of general internal medicine. 2012, 1-8.Google Scholar
- Howlader N, Noone AM, Krapcho M, Garshell J, Neyman N, Altekruse SF, Kosary CL, Yu M, Ruhl J, Tatalovich Z, Cho H, Mariotto A, Lewis DR, Chen HS, Feuer EJ, Cronin KA: SEER Cancer Statistics Review. 2013, National Cancer Institute. Bethesda, MD, 1975-2010. based on November 2012 SEER data submission, posted to the SEER web site, April, [http://seer.cancer.gov/csr/1975_2010/]Google Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.