 Research
 Open Access
 Published:
An embedded method for gene identification problems involving unwanted data heterogeneity
Human Genomics volume 13, Article number: 45 (2019)
Abstract
Background
Modern applications such as bioinformatics collecting data in various ways can easily result in heterogeneous data. Traditional variable selection methods assume samples are independent and identically distributed, which however is not suitable for these applications. Some existing statistical models capable of taking care of unwanted variation were developed for gene identification involving heterogeneous data, but they lack model predictability and suffer from variable redundancy.
Results
By accounting for the unwanted heterogeneity effectively, our method have shown its superiority over several stateofthe art methods, which is validated by the experimental results in both unsupervised and supervised gene identification problems. Moreover, we also applied our method to a pancancer study where our method can identify the most discriminative genes best distinguishing different cancer types.
Conclusions
This article provides an alternative gene identification method that can accounting for unwanted data heterogeneity. It is a promising method to provide new insights into the complex cancer biology and clues for understanding tumorigenesis and tumor progression.
Background
Many variable selection methods have been proven to be successful in handling highdimensional data in the past few decades. However, most of these methods assume the data are independent and identically distributed (i.i.d.), whose applications are severely limited. In most realworld applications, we are quite often facing the heterogeneous data violating the assumptions [1]. For instance, the gene expression data are usually found confounded by “batch effects” in bioinformatics [2–4], which refers to the systematic error generated while the samples are probed by multiple batches of platforms [5–7]. In many highthroughput biological experiments, technical noise will significantly influence the subsequent data analyses [8]. Using technical replicate measurements is a common way to reduce the impact of technical noise on gene expression analysis. However, technical replicates or batch effects will definitely cause the data samples have group structures, resulting in heterogeneous data for which the traditional variable selection methods designed for generic data are not suitable. These experimental or technical factors are usually known and can be easily adjusted. Nevertheless, gene expression data typically suffer from other additional unknown factors that can induce more complex data heterogeneity. It is not the truth that all of these heterogeneous factors are unwanted in association analysis or variable selection problems. Sometimes a certain known/observed heterogeneous factor could be of research interest and the others are unwanted. For example, when one would like to know the variables that can best describe the data grouping structure. In this case, only those heterogeneous factors except the grouping factor are unwanted. While in some other circumstances, the heterogeneous factors are unwanted if none of them is of interest. For example, some gene expression data are confounded by batched effects and some other unknown factors, and the corresponding gene identification problem is to seek those genes discriminative in the related disease phenotype. In this case, the factor of research interest is the disease phenotype and all the other heterogeneous factors confounding the gene expression data are unwanted. Therefore, the specific problems and the factors of interest determine whether the gene expression heterogeneity factors are wanted or unwanted in the association analysis. In either case, if the unwanted heterogeneity is not taken into account, it would lead to spurious results by directly applying conventional variable selection methods to gene identification problems involving heterogeneous data. Moreover, it is even more challenging to adjust the unwanted heterogeneity if unknown heterogeneity are involved.
Existing methods proposed for feature selection on heterogeneous data deal with those data generated from multiple views or sources that describe the same set of samples [9–12]. The corresponding data heterogeneity generated from multiple views or sources are taken advantage to improve the results learned from any single data source. This is usually completed by formulating variable selection problem as an multipletask learning or integrative learning problem. The variable selection strategies in most of these methods are inspired by sparse learning based models [13, 14] or sparse principal component analysis [15, 16]. However, this article studies different problems in which the performance of variable selection is hurt by the data heterogeneity that should be removed. Most previous related work addressing the heterogeneity issue focuses on adjusting the observed data heterogeneity. However, the data analysis results in bioinformatics have proven to be significantly influenced by the complex data heterogeneity arising from unknown heterogeneity factors [17, 18]. Rare attentions have been paid to the adjustment of this kind of data heterogeneity. Surrogate variable analysis (SVA) [17] have been proposed by Leek and Store to adjust the unknown data heterogeneity for statistical analysis. This method lacks model predictability and has the variable redundancy issue as a filtering variable selection method [1]. Therefore, this article proposes an effective embedded variable selection method that allows for model prediction from a sparse learning perspective [19]. Our method is capable of adjusting the unwanted heterogeneity no matter they are known or unknown. Moreover, it is also suitable for both unsupervised and supervised variable selection problems. We will study three different unsupervised and supervised gene identification problems to investigate its performance based on either gene expression benchmark data or realworld RNASeq gene expression data.
Methods
Data modelling
We denote the gene expression data by X, which contains n samples and p variables. The group information causing the heterogeneity is represented by a binary indicator matrix \(G\in \mathbb {R}^{n\times g}\) whose element G_{ij} with value ‘1’ indicates that the ith sample belongs to the jth group and vice versa. In different kinds of variable selection problems, the group factor can play different roles contingent on the corresponding interested factors. For instance, we have the unsupervised variable selection problems aiming to select the variables that can best determine the data group structure. In this problem, our factor of interest is G and we model X by:
where ε∈N(0,σ^{2}I) represents the i.i.d. noise and \(\mathbf {\eta }\in \mathbb {R}^{g\times g}\) reflects the group factor effect on each variable. While in the supervised variable selection problems, additional observed variables denoted by Y are introduced and affected by the variables of X. The factors of interest is Y here and the association between Y and the variables of X is what we want to study. In this case, we model X by :
in which γ reflects the effect of Y on each variable.
In both types of the above problems, there may also involves some other unaware factors causing unknown heterogeneity sometimes. The unknown heterogeneity refers to the variation patterns caused by the unknown factors that are not explicitly included in the data model. Let F={F_{k}:1≤k≤K} denote the K unknown factors. Certainly, they should be regarded as uninterested factors and included in the data models. Thus, we correct the model for unsupervised problems as:
and the model for supervised problems as:
Apparently, the analysis of true relations between the factors of interest and each variable should take the unwanted heterogeneity into account. In unsupervised problems, the unwanted heterogeneity is caused by those uninterested factors referring to F. While in supervised problems, those uninterested factors refer to {G,F}.
Variable selection
Our variable selection strategy contains two main stages: (1) adjust the unwanted heterogeneity and create a new data set X_{a}. The effects of the factors of interest are thus represented by the variables of the new data set; (2) based on the adjusted data set X_{a}, select the variables associated with the factors of interest.
It is impossible to directly estimate the unknown factors F for the first stage. We take the procedure similar to SVA [17] to estimate the unknown factors. After removing the effects from the uninterested factors, we obtain a new data set whose variation is only determined by the factors of interest. We show the detailed steps of the first stage as below.
Stage 1: Creating a new adjusted data set X _{ a } .
1. Remove the effects from the known factors.
We first reformulate both model (1) and (2) in a general form:
where N dentotes the known factors which refer to G for model (1) and {Y,G} for model (2). Let ℜ_{N} denote the column space of N. Correspondingly, let R_{N} denote the residual operator of N that projects onto the orthogonal complement of ℜ_{N}, which is denoted by I−N(N^{T}N)^{−1}N^{T}. By multiplying R_{N} to both sides of model (3), we have:
Through projecting X onto the orthogonal complement of ℜ_{N}, we can remove the effects from the known factors N.
2. Extract surrogate factors h_{k}.
It is challenging to directly estimate F. In this step, we estimate surrogate factors h_{k}(1≤k≤K) instead to represent the residual heterogeneity R_{N}Fϕ. We can apply any factor analysis method on R_{N}X to produce h_{k}. We consider Singular Value Decomposition (SVD) here to remove the arbitrary.
3. Estimate unknown factors F_{k}.
We estimate each surrogate factor h_{k} based on the following steps: 1) select several variables of X most associated with h_{k};2) conduct SVD on the set constructed by the selected variables and return the eigenvectors e_{j}(1≤j≤n);3) find j^{∗}=argmax_{1≤j≤n}cor(e_{j},h_{k}) and set \(\phantom {\dot {i}\!}F_{k}=e_{j^{*}}\).
4. Determine the uninterested factors U.
We set U to F for the unsupervised cases and to {G,F} for the supervised cases.
5. Build a new adjusted data set X_{a}.
Both model (1) and (2) are represented by a general form:
where M represents the factors of interest. It refers to G for model (1) and Y for model (2) respectively. Then, we assume that the true relations between the variables and the factors of interest hide in an adjusted data set X_{a}. By removing the unwanted heterogeneity, we calculate X_{a} as X−Uα.
Stage 2: Scoring variables
We employed an optimal scoring idea [20] to estimate the true effects of variables to score the data variables associated with the factors of interest M. We formulate the following variable selection problem with model robustness:
where φ sums over a vector of row norms (i.e, b=[b_{1}_{2},…,b_{p}_{2}]^{T} where b_{j}(1≤j≤p) denotes the jth row of the p×l projection matrix B). The l_{2} norm of the jth row vector of B reflects the influence of the jth variable of X on M. The variable selection is allowed based on the row norms of B with an appropriate choice of λ. We update all the unknown variables iteratively until convergence to find the solutions of problem (5).
In the t+1th iteration, we update Θ^{t+1} given fixed B^{t} by solving the subproblem:
We introduce a new variable Θ^{′} to denote \(\left (M^{T}M\right)^{\frac {1}{2}}\Theta \) and then solve the corresponding optimization problem with respect to Θ^{′}. By doing SVD of \(\left (M^{T}M\right)^{\frac {1}{2}}M^{T}X_{a}B^{t}\), we obtain the optimal Θ^{′}.
Given Θ^{t}, we can update B by solving the subproblem:
Since the φ function is a nondecreasing and concave function of the l_{2} norm of row vectors of B, this is a regularized regression problem which can be solved by iterative reweighted l_{2} algorithm [21]. Despite the convergence rate may be slow, the updating rules of this algorithm are closedform and simple. In detail, the iterative updates are:
in which \(W=\text {diag}\left (w_{1}^{1},\ldots,w_{p}^{1}\right)\). Since the influences of the variables on the factors of interest are reflected by the l_{2} norm of the corresponding row vectors of B, we select variables by ranking all the variables in a decreasing order based on them.
In summary, the whole detailed procedure for variable selection of our method is given by Algorithm 1. If it needs r iterations to converge, the time complexity of this algorithm is O(r(n^{2}p+n^{3})). Usually, it takes a few iterations to converge based on our experience. It is worth to note that the algorithm is desirable for highdimensional data because it scales linearly in the number of variables. We name our method as sparse optimal scoring with adjustment (SOSA) considering we use a sparse optimal scoring strategy to select variables based on the adjusted data.
Results
We studied three different unsupervised or supervised gene expression analysis problems to investigate the performance of our method. The first problem is unsupervised and studied on spikein gene expression data, while the last two problems are supervised and studied on realworld gene expression data. The performance in gene identification while taking care of unwanted heterogeneity is evaluated by comparing with stateoftheart methods for each problem.
Evaluation criteria
We evaluate and compare the gene identification performance of all the methods based on the specificity and sensitivity of the method’s ability to detect the differentially expressed genes. Particularly, we consider the number of true positives and false positives as well as the receiver operating characteristic (ROC) curve. These criteria are only suitable for the experiments in which the true differentially expressed genes are known.
Given a topN ranking list of the identified genes obtained from a method, ♯true positives is the number of true genes that appear in the topN ranking list and ♯false positives is the number of false genes that appear in the topN ranking list. Correspondingly, the true positive rate and false positive rate are:
where ♯true genes is the number of true genes which are known truly associated with the factor of interest and ♯false genes is the number of false genes which are known falsely associated with the factor of interest. The receiver operating characteristic (ROC) curve is plotted based on the calculated true positive rate (sensitivity) and false positive rate (1specificity). A method with the largest area under curve (AUC) achieves both the highest sensitivity and specificity.
Unsupervised differential gene expression analysis in a spikein study
This study attempts to find the genes that are differentially expressed in different tissues based on the benchmark data set from the spikein experiment [22]. It belongs to unsupervised differential gene expression analysis. In this spikein experiment, a series of 14 human cRNA fragments were spikedin at known concentration ratings from 0 to 1024 pM, which are the spiked probe sets/genes. For each of the 14 array groups, the cRNAs were spikedin at different concentrations. In each array group, there were three replicates except that there are two array groups having 12 replicates. The detailed arrangement of concentrations of the 14 cRNAs for each group can be found in [23]. In total, there are 59 arrays belonging to 14 array groups in the spikein data. Each array contains 12,626 probe sets and 14 of them are spikedin. Since Cope et al. found two more genes with similar patterns to the spiked genes [22], our goal here is to detect those 16 genes differentiating the 14 array groups. In other words, we want to discover those genes most representative of the biological variation from different tissues and meanwhile most resilient to the technical noise hidden among the technical replicates.
We compared three methods accounting for the heterogeneity or not. Robust Multiarray Average (RMA) is the most popular method for gene expression analysis but taking no heterogeneity into account. It summarizes the replicate measurements to a gene expression value [24], which however loses the technical variation information hidden in the replicates. We ranked the genes based on their average foldchange in RMAsummarized gene expression values. Another two representative methods taking care of the heterogeneity we considered here are SVA and SOSA. SVA uses the estimated uninterested factors as covariates and conducts association analysis for each individual gene. The genes are ranked in the increasing order of their pvalues. In contrast, our SOSA ranked the genes based on the l_{2} norm of their corresponding row vectors of the estimated B. We then calculated the respective true positives and false positives based on the rankings of the three methods. The number of their true positives appearing in the top 10, top 16, top 25 and top 50 ranking lists are shown by Fig. 1a. The reason we investigate the top 16 ranking list is that the number of the groundtruth genes is 16. We found that SOSA always identified the most true positives in all the cases. Fig. 1b further showed a more comprehensive investigation by ROC curves, in which our SOSA achieved the highest sensitivity and specificity while RMA the lowest. The superiority of our method is further validated. We select variables via choosing an appropriate λ to force the corresponding rownorms of B for most genes close to 0. The choice of λ was done by cross validation in which we sequentially select one array out of each array group to build the testing data and the rest to build training data. The average objective values calculated for the testing data using the corresponding estimated parameters at 20 different values of λ are shown by Fig. 2a. We chose the best λ as 0.001 that corresponds to the smallest objective value. Usually, the range of the best λ is expected falling in [0.0005,0.01]. Figure 2b showed the corresponding top100 ranking of genes based on their corresponding rownorms of B. We found that the topranked 16 genes matched the groundtruth exactly.
Supervised differential gene expression analysis in a gender study
In this experiment, we study a supervised gene identification problem which is first conducted by Vawter et al. [25]. It tries to find those genes differentially expressed in human brain with respect to the gender. In this problem, the samples are clearly clustered by laboratory and chip type because they were sent to different laboratories measured by one of these two types of chips. The interested effects related to the gender were blurred by these strong batch effects and consequently the analysis results will be significantly influenced. For this problem, our goal is to find those genes that can best differentiate the samples from different genders. Moreover, these genes should also be resilient to the unknown effects and the known batch effects. The samples were taken postmortem from the brains of 10 individuals including 5 men and 5 women in this study. Specifically, for each individual, the experiment takes three samples from different regions of its brain. Each sample was sent to three laboratories for analysis based on either Affymetrix HGU95Av2 or Affymetrix HGU95A platform. Since there are 6 missing chips in the 90 combinations, we have 84 collected chips in total. The processed data set keeps the 12,600 probe sets shared by the two platforms. RMA is used to measure the gene expression values for the 84 chips.
We compared six representative methods: individual gene association analysis (IndAss), penalized logistic regression (PLR), the standard sparse optimal scoring (SOS), linear model adjusting batch effects (Lm _batch), SVA and our SOSA, to identify those genes that can best differentiate the samples from different genders. Their sensitivity and specificity in gene selection were investigated. PLR, IndAss and SOS run directly on the gene expression data without accounting for the unwanted heterogeneity. The ‘glmnet’ package was employed to implement PLR with the parameter chosen by cross validation. We ranked the genes based on the absolute values of the regression coefficients in a decreasing order. Linear model was also employed here to adjust the known batch effects, which however is not able to adjust the unknown heterogeneous factors. We also considered two alternative methods accounting for the heterogeneity: SVA and our SOSA. We considered three unknown factors here. By taking the lab/chip factor and the estimated unknown factors as covariates, SVA did the association analysis for each individual gene and then ranked the genes based on their pvalues in an increasing order. In contrast, our SOSA ranked the genes based on the l_{2} norm of their corresponding row vectors of the estimated B. The parameter λ in SOSA was chosen by leaveoneout cross validation. Figure 3a showed the plot of average objective values for the testing data at 10 different values of λ falling in the range between 0.0001 and 0.05. The best λ was chosen as 0.004 which corresponds to the minimum average objective value across all the folds. We treat all the 488 genes from the X and Y chromosomes as candidate positive controls considering most genes differentially expressed with respect to the gender will be located on the X and Y chromosomes. For the negative controls, we use the 799 housekeeping genes claimed by [26]. The ROC curves for all the six methods are shown by Fig. 3b, in which the superiority of our SOSA is validated since it almost always reached the highest sensitivity and specificity. Table 1 showed the number of true positives and false positives in the top 20, 60, 80, 180 for each method. We observed that those methods with adjustment can identify more X/Y genes or less housekeeping genes than the unadjusted ones, which demonstrates their significant improvement on the gene selection performance. Benefiting from the model stability accomplished by the regularization of the gene coefficients, our SOSA achieved better performance than SVA approach. We also observed that SOSA is superior to Lm _batch since it can adjust both the known batch effects and the unknown heterogeneous factors. For those methods without adjustment, SOS and IndAss achieved higher sensitivity than PLR by correctly detecting more X/Y genes, but IndAss has lower specificity since it wrongly identified much more housekeeping genes than the other two methods.
Supervised differential gene expression analysis in pancancer study
The two studies above investigated the gene identification performance of our method on heterogeneous gene expression benchmark data with groundtruth genes known. In this experiment, we explore the gene identification performance of our method as well as the interpretation of those identified genes in the application to a pancancer study. This experiment aims to identify a set of genes differentially expressed across five common types of cancer that are clinically attractive as diagnostic biomarkers or therapeutic targets based on The Cancer Genome Atlas (TCGA) gene expression data. Particularly, we use the publicly available RNASeq (HiSeq) PANCAN data set from the UCI machine learning repository (https://archive.ics.uci.edu/ml/datasets/gene+expression+cancer+RNASeq##), which is a random extraction of RNASeq gene expressions of patients having five types of cancer: Breast (BRCA), Colon (COAD), Kidney (KIRC), Lung (LUAD) and Prostate (PRAD) from the TCGA data portal. This extracted and preprocessed data set contains 801 patients with 20,531 genes measured in total. Among them, 300, 78, 146, 141 and 136 patients have BRCA, COAD, KIRC, LUAD and PRAD respectively, which implies that the samples are relatively balanced across different types of cancer.
To identify the set of genes that can best classify these five types of tumor, we considered five methods including three typical methods without accounting for any unwanted heterogeneity: PLR, SOS, IndAss and two methods that are able to adjust for the unwanted heterogeneity: SVA and SOSA. We separated the data samples into training samples and testing samples by a 5fold cross validation. The differentially expressed genes were identified using these methods on the training data respectively. Their utility was investigated by evaluating the corresponding classification performance on the test samples using them based on 1NN classifier. The average classification testing error rates for each method using their identified top 10, 30, 50, 100, 200, 300, 400 genes are shown by Table 2. We also show the classification testing error rate using all the genes on the test data as the baseline. The different types of cancer can be well classified with a 99.5% classification accuracy using all genes. The utility of the respective top 30 genes identified by PLR, SOS and SOSA are competitive to the baseline. As filtering methods, IndAss and SVA identified pretty worse sets of top N genes than the others for N less than 50 due to lack of considering the combining effects of the identified genes. With N larger than 200 but not too large, the top N genes identified by each of the methods can lead to competitive or higher classification performance than the baseline. Among these methods, our SOSA can even achieve 99.9% classification accuracy with its top 100 identified genes. Overall, our SOSA can identify the most discriminative genes comparing to other methods at each fixed number of selected genes. And the superiority is even higher as the number of selected genes goes small.
In addition to examination of the classification performance of the identified genes, we care much more about the interpretation of the role of these genes in distinguishing the five cancer types as well as their involved biological processes. Specifically, we found the union of the top 200 genes identified from each fold of the partitions. The final top 20 genes in the union set were: MAB21L1, HAND1, SFTA2, SFTPA2, CDV3, NBL1, SFTA3, KLK1, HNF1A, NKX12, NOV, KLK2, SFTPC, PQLC3, TCF20, NAPRT1, POU3F2, CDH15, SCGB3A1 and GP9.
MAB21L1 belongs to the conserved male abnormal gene family 21, which is described as a transcription factor in cell fate determination [27]. Heart and neural crest derivatives expressed 1 (HAND1) is a basic helixloophelix transcription factor, and plays a very important role in the development and differentiation of heart and nervous system. As a developmental regulator, HAND1 is silenced in over 90% of human primary colorectal tumors [28]. Four surfactant genes (SFTA2/3, SFTPA2, and SFTPC) were identified in the top 20. They were highly expressed in lung cancer and low in all other tumors. CDV3 was documented as an unidentified gene in breast cancer [29], whose expression correlated to the expression of Her2 and the sensitivity of photonirradiation and simultaneous PTXtreatment in breast cancer [30]. Three kallikrein genes (KLK1/2/3) were identified in the top 30. The risk of prostate cancer has been reported associated with singlenucleotide polymorphisms (SNPs) located in the genes coding for PSA (KLK3) [31] and hK2 (KLK2) [32]. The classical cadherins (CDHs) are a superfamily of transmembrane glycoproteins involved in calciumdependent cell–cell adhesion in embryonic development and epithelial tissues. CDHs are also associated with signaling, mechanotransduction, cancer progression, and tissue morphogenesis, many of which are related to cancer [33].
We also performed gene ontology (GO) analysis of the top 200 genes using WebGestalt tool [34]. The top 10 highly enriched biological processes with GO terms are listed in Table 3 and the corresponding topology is shown by Fig. 4. Each gene ontology category is a node in the graph. GO categories in blue nodes are the top 10 enriched GO categories while the others are their nonenriched parent categories. It is clear that many of these genes are involved in the biological processes of development or morphogenesis of multicellular organism, organ, embryo and tissue, which may provide some clues for understanding tumorigenesis and tumor progression as well as new insights into the complex cancer biology.
Discussion
The results from our experiments on three unsupervised and supervised gene identification problems have shown the superiority of our method in handling unwanted data heterogeneity issue. The first problem studied a representative application of finding the differentially expressed genes from gene expression data with batch effects in unsupervised circumstance. Our method performs better than RMA and SVA since it is capable of taking care of the unwanted heterogeneity as well as variable redundancy. The second problem studied a typical application of identifying differentially expressed genes from gene expression data with batch effects in supervised circumstance. Our method performs better in the sensitivity and specificity in variable selection than other stateoftheart methods including penalized logistic regression model, simple statistical model for individual gene association analysis, sophisticated statistical model capable of accounting for data heterogeneity and linear model accounting for batch effects. This benefit from our twostage embedded variable selection strategy. The third problem studied a general case of supervised gene expression analysis where the batch effects are unknown. Our method can achieve the highest classification accuracy comparing to other methods using the same number of discriminative genes. In summary, our method has wide applications in various kinds of gene identification problems.
Conclusions
We proposed sparse optimal scoring with adjustment (SOSA) for gene identification on heterogeneous data involving unwanted heterogeneity. Particularly, our method is able to account for the unknown and unwanted heterogeneity that blurs the true signals of genes. The results from the first two studies on the batched gene expression data demonstrate that those methods accounting for the unwanted heterogeneity can significantly improve the gene identification performance. Moreover, our method is also superior to the statistical models capable of adjusting data heterogeneity because of the model stability and the ability of prediction on new samples. The results from pancancer study further validated the superiority of our method in identifying the genes discriminative in different cancer types to other stateoftheart methods. The biological interpretation of the results provides new insights into complex cancer biology and clues for understanding tumorigenesis. In view of the commonly existing data heterogeneity issue such as batch effects in bioinformatics, our SOSA can also be adapted and serves as a promising tool to remove the unwanted heterogeneity prior to downstream analysis for many other biomedical applications such as cell population diversity in singlecell landscapes and so on.
Abbreviations
 AUC:

Area under curve
 FDR:

False discovery rate
 IndAss:

Individual gene association analysis
 Lm _batch:

linear model adjusting batch effects
 PLR:

Penalized logistic regression
 RMA:

Robust multiarray average
 ROC:

Receiver operating characteristic
 SOS:

The standard sparse optimal scoring
 SOSA:

Sparse optimal scoring with adjustment
 SVA:

Surrogate variable analysis
 SVD:

Singular value decomposition
References
 1
Li J, Cheng K, Wang S, Morstatter F, Trevino RP, Tang J, Liu H. Feature selection: A data perspective. arXiv preprint arXiv:1601.07996. 2016.
 2
Chen C, Grennan K, Badner J, Zhang D, Gershon E, Jin L, Liu C. Removing batch effects in analysis of expression microarray data: An evaluation of six batch adjustment methods. PLoS ONE. 2011; 6(2):17238.
 3
Almeida A, Paul JT, Magdelenat H, Radvanyi F. Gene expression analysis by realtime reverse transcription polymerase chain reaction: influence of tissue handling. Anal Biochem. 2004; 328(2):101–8.
 4
Ma Y, Dai H, Kong X. Impact of warm ischemia on gene expression analysis in surgically removed biosamples. Anal Biochem. 2012; 423(2):229–35.
 5
Bakay M, Chen YW, Borup R, Zhao P, Nagaraju K, Hoffman EP. Sources of variability and effect of experimental approach on expression profiling data interpretation. BMC Bioinformatics. 2002; 3(1):4.
 6
Boedigheimer MJ, et al.Sources of variation in baseline gene expression levels from toxicogenomics study control animals across multiple laboratories. BMC Genomics. 2008; 9(1):285.
 7
Fare TL, et al.Effects of atmospheric ozone on microarray data quality. Anal Chem. 2003; 75(17):4672–5.
 8
Glaab E, Schneider R. Repexplore: addressing technical replicate variance in proteomics and metabolomics data analysis. Bioinformatics. 2015; 31(13):2235–7.
 9
Zhao Z, Liu H. Multisource feature selection via geometrydependent covariance analysis. JMLR Work Conf Proc. 2008; 4:36–47.
 10
Tang J, Hu X, Gao H, Liu H. Unsupervised feature selection for multiview data in social media. In: Proceedings of the 2013 SIAM International Conference on Data Mining.2013. p. 270–8.
 11
Feng Y, Xiao J, Zhuang Y, Liu X. Adaptive unsupervised multiview feature selection for visual concept recognition. In: Computer Vision–ACCV 2012. Berlin: Springer: 2013. p. 343–57.
 12
Wang H, Nie F, Huang H. Multiview clustering and feature learning via structured sparsity. In Proceedings of the 30th International Conference on Machine Learning. 2013:352–60.
 13
Friedman J, Hastie T, Tibshirani R. A note on the group lasso and a sparse group lasso. arXiv preprint arXiv:1001.0736. 2010.
 14
Peng J, et al.Regularized multivariate regression for identifying master predictors with application to integrative genomics study of breast cancer. Ann Appl Stat. 2010; 4(1):53.
 15
d’Aspremont A, Ghaoui LE, Jordan MI, Lanckriet GR. A direct formulation for sparse pca using semidefinite programming. SIAM Rev. 2007; 49(3):434–48.
 16
Lu M, Huang JZ, Qian X. Sparse exponential family principal component analysis. Pattern Recog. 2016; 60:681–91.
 17
Leek JT, Store JD. Capturing heterogeneity in gene expression studies by surrogate variable analysis. PLoS Genet. 2007; 3:161.
 18
GagnonBartsch JA, Speed TP. Using control genes to correct for unwanted variation in microarray data. Biostatistics. 2012; 13(3):539–52.
 19
Lu M. An embedded method for gene identification in heterogenous data involving unwanted heterogeneity. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine. 2018:242–7.
 20
Trevor H, Robert T, Andreas B. Flexible discriminant analysis by optimal scoring. J Am Stat Assoc. 1994; 89(428):1255–70.
 21
David W, Srikantan N. Iterative reweighted l1 and l2 methods for finding sparse solutions. IEEE J Sel Top Sign Process. 2010; 4(2):317–29.
 22
Cope LM, Irizarry RA, Jaffee HA, Wu Z, Speed TP. A benchmark for affymetrix genechip expression measures. Bioinformatics. 2004; 20(3):323–31.
 23
Irizarry RA, et al.Summaries of affymetrix genechip probe level data. Nucleic Acids Res. 2003; 31(4):e15.
 24
Irizarry RA, Hobbs B, Collin F, BeazerBarclay YD, Antonellis KJ, Scherf U, Speed TP. Exploration, normalization, and summaries of high density oligonucleotide array probe level data. Biostatistics. 2003; 4:249–64.
 25
Vawter MP, et al.Genderspecific gene expression in postmortem human brain: Localization to sex chromosomes. Neuropsychopharmacology. 2004; 29(2):373–84.
 26
Eisenberg E, Levanon EY. Human housekeeping genes are compact. TRENDS Genet. 2003; 19(7):362–5.
 27
Baird S, Fitch D, Kassem I, Emmons S. Pattern formation in the nematode epidermis: determination of the arrangement of peripheral sense organs in the c.elegans male tail. Development. 1991; 113:515–26.
 28
Tan J, et al.Integrative epigenome analysis identifies a polycombtargeted differentiation program as a tumorsuppressor event epigenetically inactivated in colorectal cancer. Cell Death Dis. 2014; 5(7):1324.
 29
Agus DB, Bunn PA, Franklin W, et al.Her2/neu as a therapeutic target in nonsmall cell lung cancer, prostate cancer, and ovarian cancer. Semin Oncol. 2000; 27(6):53–63.
 30
Oh JJ, Grosshans DR, Wong SG, et al.Identification of differentially expressed genes associated with her2/neu overexpression in human breast cancer cells. Nucleic Acids Res. 1999; 27(20):4008–17.
 31
Pal P, Xi H, Sun G, Kaushal R, Meeks J, Thaxton C, et al.Tagging snps in the kallikrein genes 3 and 2 on 19q13 and their associations with prostate cancer in men of european origin. Hum Genet. 2007; 122:251–9.
 32
Nam R, Zhang W, Trachtenberg J, Diamandis E, Toi A, Emami M, et al.Single nucleotide polymorphism of the human kallikrein2 gene highly correlates with serum human kallikrein2 levels and in combination enhances prostate cancer detection. J Clin Oncol. 2003; 21:2312–9.
 33
Zhu C, Feng X, Ye G, Huang T. Metaanalysis of possible role of cadherin gene methylation in evolution and prognosis of hepatocellular carcinoma with a prisma guideline. Med (Baltimore). 2017; 96(16):6650.
 34
Zhang B, Kirov S, Snoddy J. Webgestalt: an integrated system for exploring gene sets in various biological contexts. Nucleic Acids Res. 2005; 33(Web Server issue):741–8.
Acknowledgements
Not applicable.
About this supplement
This article has been published as part of Human Genomics Volume 13 Supplement 1, 2019: Selected articles from the IEEE BIBM International Conference on Bioinformatics & Biomedicine (BIBM) 2018: human genomics. The full contents of the supplement are available online at https://humgenomics.biomedcentral.com/articles/supplements/volume13supplement1.
Funding
Publication costs were funded by the faculty startup fund for ML from Tianjin University, China.
Availability of data and materials
All datasets are publicly available specified in the manuscript.
Author information
Affiliations
Contributions
ML is in charge of idea development, experiment design and implementation, result analysis and manuscript drafting. The author read and approved the final manuscript.
Corresponding author
Ethics declarations
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The author declares that he has no competing interests.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
About this article
Cite this article
Lu, M. An embedded method for gene identification problems involving unwanted data heterogeneity. Hum Genomics 13, 45 (2019). https://doi.org/10.1186/s4024601902280
Published:
Keywords
 Unwanted heterogeneity
 Gene identification
 Embedded variable selection