About MVIAevalConstruction of MVIAevalUsage of MVIAeval

Motivation of MVIAeval

What is MVIAeval?



Motivation of MVIAeval

Missing value imputation is important for microarray data analyses because microarray data with missing values would significantly degrade the performance of the downstream analyses. Although many microarray missing value imputation algorithms have been developed, an objective and comprehensive performance comparison framework is still lacking. Therefore, in our previous paper (Chiu et al. 2013), we proposed a framework which can perform a comprehensive performance comparison of different existing algorithms. Our performance comparison framework can also be applied to evaluate the performance of a newly developed algorithm. However, constructing our framework is not an easy task for the interested researchers. To save researchers time and effort, here we present an easy-to-use web tool named MVIAeval (Missing Value Imputation Algorithm evaluator) which implements our performance comparison framework.

TOP


What is MVIAeval?

MVIAeval provides a user-friendly interface allowing users to upload the R code of their new algorithm and select

(i) the test datasets among 20 benchmark microarray (time series or non-time series) datasets
(ii) the compared algorithms among 12 existing algorithms
(iii) the performance indices from three existing ones
(iv) the comprehensive performance scores from two possible choices
(v) the number of simulation runs

The comprehensive performance comparison results are then generated and shown as both figures and tables.
TOP


Twenty benchmark microarray datasets and twelve existing algorithms used for performance comparison

In MVIAeval, we collected 20 benchmark microarray datasets of different species and different types.


GEO DatasetDimTypeOrganismTitle
GDS332345101*6Non-time SeriesMus musculusNa+/H+ exchanger 3 deficiency effect on the colon
GDS321512625*6Non-time SeriesHomo sapiens13-cis retinoic acid effect on SEB-1 sebocyte cell line
GDS348545011*6Non-time SeriesMus musculusZinc transporter SLC39A13 deficiency effect on chondrocytes
GDS347645011*6Non-time SeriesMus musculusNF-E2-related factor 2 Nrf2 activation effect on the liver
GDS319745101*6Non-time SeriesMus musculusTranscriptional coactivator PGC-1beta hypomorphic mutation effect on the liver
GDS314945101*6Non-time SeriesMus musculusSuppressor of cytokine signaling 3 deficiency effect on the regenerating liver
GDS210715923*6Non-time SeriesRattus norvegicusLong-term ethanol consumption effect on pancreas
GDS346415617*6Non-time SeriesDanio rerioSPT5 mutant embryos
GDS342623015*6Non-time SeriesStaphylococcus epidermidisStaphylococcus epidermidis SarZ mutant
GDS342110208*6Non-time SeriesEscherichia coliFrag1 cells response to ionic and non-ionic hyperosmotic stress
GDS336022575*8Time SeriesHomo sapiensChlamydia pneumoniae infection effect on HL epithelial cells: time course
GDS286331099*6Time SeriesRattus norvegicusTienilic acid effect on the liver: time course
GDS505734760*8Time SeriesMus musculusMepenzolate bromide effect on lung: time course
GDS505545307*10Time SeriesMus musculusHistone demethylase KDM1A deficiency effect on 3T3-L1 preadipocytes: time course
GDS342822283*9Time SeriesHomo sapiensImmature dendritic cell response to butanol fraction of Echinacea purpurea: time course
GDS448445101*8Time SeriesMus musculusCerebellar neuronal cell response to thyroid hormone: time course
GDS378517589*8Time SeriesHomo sapiensOsteoarthritic chondrocytes and healthy mesenchymal stem cell during chondrogenic differentiation: time course
GDS39308799*9Time SeriesRattus norvegicusBone morphogenic protein effect on cultured sympathetic neurons: time course
GDS432110208*8Time SeriesEscherichia coliEscherichia coli O157:H7 response to cinnamaldehyde: time course
GDS303222277*8Time SeriesHomo sapiensQuercetin effect on intestinal cell differentiation in vitro: time course




In addition, we implemented 12 existing algorithms including two global approach algorithms and 10 local approach algorithms.


AlgorithmCategoryYear of PublishedReference
SVDGobal2001[ Troyanskaya et al. 2001 ]
BPCAGobal2003[ Oba et al. 2003 ]
KNNLocal2001[ Troyanskaya et al. 2001 ]
SKNNLocal2004[ Kim et al. 2004 ]
IKNNLocal2007[ Brás et al. 2007 ]
LSLocal2004[ Bø et al. 2004 ]
LLSLocal2005[Kim et al. 2005 ]
ILLSLocal2006[Cai et al. 2006 ]
SLLSLocal2008[Zhang et al. 2008 ]
Shrinkage LLSLocal2013[Wang et al. 2013 ]
Shrinkage SLLSLocal2013[Wang et al. 2013 ]
Shrinkage ILLSLocal2013[Wang et al. 2013 ]
TOP


Three existing performance indices used for performance evaluation

In MVIAeval, we used three existing performance indices for performance evaluation.
First, the inverse of the normalized root mean square error (1/NRMSE) is used to measure the numerical similarity between the imputed matrix (generated by an imputation algorithm) and the original complete matrix. Therefore, the higher the 1/NRMSE value is, the better the performance of an imputation algorithm is.
Second, the cluster pair proportions (CPP) is used to measure the similarity of the gene clustering results of the imputed matrix and the complete matrix. High CPP value means that the imputed matrix (generated by an imputation algorithm) has very similar gene clustering results as the complete matrix does. Therefore, the higher the CPP value is, the better the performance of an imputation algorithm is.
Third, the biomarker list concordance index (BLCI) is used to measure the similarity of the differentially expressed genes identification results of the imputed matrix and the complete matrix. High BLCI value means that differentially expressed genes identified using the imputed matrix (generated by an imputation algorithm) are very similar to those identified using the complete matrix. Therefore, the higher the BLCI value is, the better the performance of an imputation algorithm is.
In summary, 1/NRMSE measures the numerical similarity, while CPP and BLCI measure the similarity of downstream analysis results (gene clustering and differentially expressed genes identification) of the imputed matrix and the complete matrix.
TOP


Evaluating the performance of an algorithm for a benchmark microarray data matrix using a specific performance index

The simulation procedure for evaluating the performance of an imputation algorithm (e.g. KNN) for a given complete benchmark microarray data matrix using a performance index (e.g. CPP) is divided into four steps:

Step 1: randomly generate five testing matrices having missing values with different percentages (1%, 3%, 5%, 8% and 10%) from the complete matrix
Step 2: generate five imputed matrices by imputing the missing values in the five testing matrices using KNN
Step 3: calculate five CPP scores using the complete matrix and five imputed matrices
Step 4: repeat Steps 1-3 for B times, where B is the number of simulation runs per missing percentage.

Then the final CPP score of KNN for the given benchmark microarray data matrix is defined as the average of the 5*B CPP scores. The following figure illustrates the whole simulation procedure.

TOP


Two existing comprehensive performance scores

In MVIAeval, we implemented two existing comprehensive performance scores to summarize the overall performance comparison results for the selected performance indices and benchmark microarray datasets.
(I) Overall Ranking Score (ORS): the sum of the rankings of an algorithm for the selected performance indices and benchmark microarray datasets.
The ranking of an algorithm for a specific performance index and a specific benchmark microarray dataset is d if its performance ranks #d among all the compared algorithms. For example, the ranking of the best performing algorithm is 1. Therefore, the smaller the ORS is, the better the overall performance of an algorithm is.

(II) Overall Normalized Score (ONS): the sum of the normalized scores for the selected performance indices and benchmark microarray datasets.
The ONS of the algorithm $k$ is calculated as follows:

$$ONS(k) = \sum_{i=1}^I\sum_{j=1}^JN_{ij}(k)=\sum_{i=1}^I\sum_{j=1}^J(\frac{S_{ij}}{max(S_{ij}(1),S_{ij}(2),...,S_{ij}(m))})$$ where $N_{ij}(k)$ and $S_{ij}(k)$ is the normalized score and the original score of the algorithm $k$ for the selected performance index $i$ ($i$=1 for $\frac{1}{NRMSE}$, 2 for CPP, and 3 for BLCI) and benchmark microarray dataset $j$; $I$ is the number of the selected indices; $J$ is the number of the selected benchmark microarray dataset and $m$ is the number of the algorithms being compared. Note that $0≤N_{ij}(k)≤1$ and $N_{ij}(k)=1$ if and only if the algorithm $k$ is the best performing algorithm for the selected performance index $i$ and benchmark microarray dataset $j$ (i.e. $S_{ij}(k) = max(S_{ij}(1),S_{ij}(2),...,S_{ij}(m))$ ). The larger the $ONS$ is, the better the overall performance of an algorithm is.
TOP


Usage

The usage of MVIAeval is shown in the following figure.

The friendly web interface allows users to upload the R code of their newly developed algorithm. Then five kinds of settings of MVIAeval need to be specified. Users have to

(1) choose the test datasets from 20 benchmark microarray datasets
(2) choose the compared algorithms from 12 existing algorithms
(3) choose the performance indices from three existing ones (1/NRMSE, CPP, and BLCI)
(4) choose the comprehensive performance scores from two existing ones (ORS and ONS)
(5) determine the number of simulation runs

After submission, MVIAeval will conduct a comprehensive performance comparison of the user’s algorithm to the compared algorithms using the selected performance indices and benchmark datasets. Then a webpage of the comprehensive performance comparison results will be generated.
TOP


A case study

In MVIAeval, the R code of a sample algorithm is provided. For demonstration purpose, we regard the sample algorithm as the user’s newly developed algorithm and would like to use MVIAeval to conduct a comprehensive performance comparison of this new algorithm (denoted as USER) to various existing algorithms. For example, users upload the R code of the new algorithm,
select two benchmark datasets,
select 12 existing algorithms,
select three performance indices,
select the overall ranking score as the comprehensive performance score,
and use 25 simulation runs.
After submission, the comprehensive comparison results are generated and shown as both tables and bar charts. The overall performance of the new algorithm ranks six among all the 13 algorithms being compared.
Comprehensive Ranking Score: Sum of the rankings of each algorithm in the selected performance indices and selected datasets

Rankings of each algorithm in the selected performance indices and selected datasets

Performance Index DataType Dataset Algorithms Details
SLLSILLSLSLLSKNNUSERSVDBPCAIShrLLSShrLLSIKNNShrSLLSSKNN
1/NRMSEnontimeGDS321521836574910121113Details
1/NRMSEtimeGDS378512537684910121113Details
BLCInontimeGDS321534216684910121113Details
BLCItimeGDS3785342685111101171319Details
CPPnontimeGDS321515382457109121113Details
CPPtimeGDS378542341768119121310Details
Comprehensive Ranking Score14182325303335385859677071-
Final Ranking12345678910111213-
Actually, MVIAeval can provide the performance comparison results in many scenarios.
Performance IndexBenchmark DatasetsRanking of USER Using ORSRanking of USER Using ONSDetail
1/NRMSEFive Time Series: GDS3360,GDS2863,GDS5057,GDS5055,GDS342856Detail
Five Non-time Series: GDS3323,GDS3215,GDS3485,GDS3476,GDS319766Detail
CPPFive Time Series: GDS3360,GDS2863,GDS5057,GDS5055,GDS342879Detail
Five Non-time Series: GDS3323,GDS3215,GDS3485,GDS3476,GDS3197118Detail
BLCIFive Time Series: GDS3360,GDS2863,GDS5057,GDS5055,GDS342834Detail
Five Non-time Series: GDS3323,GDS3215,GDS3485,GDS3476,GDS319777Detail
1/NRMSE+CPP+BLCIFive Time Series: GDS3360,GDS2863,GDS5057,GDS5055,GDS342867Detail
Five Non-time Series: GDS3323,GDS3215,GDS3485,GDS3476,GDS319766Detail
It can conclude that the new algorithm is mediocre because its performance is always in the middle of all the compared algorithms in different data types (time series or non-time series), different performance indices (1/NRMSE, BLCI or CPP) and different comprehensive performance scores (ORS or ONS). Receiving the comprehensive comparison results from MVIAeval, researchers immediately know that there is plenty of room to improve the performance of their new algorithm.
TOP