Nothing Special   »   [go: up one dir, main page]

Skip to content

sharispe/sm-tools-evaluation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

31 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Evaluation of Software for Semantic Measures

This project aims to compare tools dedicated to the computation of semantic measures (software, source code libraries). It has been initiated in order to evaluate the Semantic Measures Library and corresponding Toolkit (http://www.semantic-measures-library.org), see the competing interest section for more information. However, note that the project doesn't focus on the Semantic Measures Library and includes comparison of several tools. It can therefore be useful for:

  • (i) end-users who want to select a tool to compute semantic measures;
  • (ii) developers who want to compare their tools.

The project currently focuses on graph-based semantic measures, e.g. measures used to assess the similarity of class/concepts/terms defined in ontologies (also denoted as measures based on the relational setting). As an example, this project doesn't evaluate tools related to distributional semantic measures or semantic measures relying on description logic.

We currently propose domain-specific evaluations. Indeed, numerous libraries and tools have been developed for domain-specific applications (e.g. ontologies/terminologies) and (to our knowledge) no project compares those software solutions. In this project, considering specific domain of use, we compare tools regarding their speed performance (running time) and their capacity to handle large datasets.

The domain for which tests are available are:

  • Molecular Biology and Biomedicine
    • Gene Ontology

More evaluation relative to other domain or application context will be added (WordNet, UMLS, Disease Ontology). This community project is public and open source - do not hesitate to contribute (see above).

Important: this project does not aim to criticize tools or denigrate the work made by their developers - we only define a strict evaluation protocol to provide objective metrics which can be relevant to compare tools. Keep in mind that tools which do not perform well on the tests defined herein may have other advantages that are not discussed in this evaluation. Please contact us to help us improve upon it. In addition, this evaluation doesn't pretend to cover all aspects which could be useful to discuss in order to evaluate software. Here we focus on objective metrics and mainly aim at evaluating the speed of the program given specific resource constraints (memory allocated to the tool, computational time). Moreover, we only provide tests and results which are strictly reproducible given the source code and information shared through this project. Our aim is not to discuss aspects relative to the (subjective) individual user experience or other important aspects such as documentation and code quality, usability, overall sustainability, community support, release updates... Please refer to corresponding tool documentations and websites to evaluate those aspects. More general information related to software evaluation can be found at http://www.software.ac.uk/software-evaluation-guide and http://software.ac.uk/sites/default/files/SSI-SoftwareEvaluationCriteria.pdf is a good start to guide software evaluation (credit to the Software Sustainability Group and more particularly to the authors Mike Jackson, Steve Crouch and Rob Baxter).

The source code used in this project is open source, it can be used to reproduce the tests. Details regarding the materials and methods used in the evaluations are also specified for each test. However, due to github constraints regarding file sizes, some datasets may not be included in this community project; they are made available through other websites.

Do not hesitate to help us improve the benchmarks and to upgrade the tests with new tools or new versions of considered tools. The issue tracker can be used to discuss any problems regarding the source code or the test protocols.

You can also contribute to the project by:

  • (i) adding new tools or upgrade existing tools or datasets;
  • (ii) adding new tests;
  • (iii) fixing bugs, improve the source code (e.g. tool parameters).

Approved contributions will be subsequently incorporated into this project.

How to Reproduce Tests and Results

The source code has been developed in order to be executed on any 64-Bit Linux operating system considering that tested tools dependencies are installed (e.g. Python, Java, R - refer to evaluated tools). I personally used the easy-to-use Ubuntu distribution (LTS 12.04) but other distributions can be used (Windows users can deploy a virtual machine to reproduce the tests). The tests can therefore be reproduced; simply download the repository and follow the instructions. Please start a discussion if you encounter any problems.

Note that results may vary considering your hardware configuration but rankings must be the same.

Molecular Biology and Biomedicine

This section contains all tests relative to computation of semantic measures scores for Molecular Biology and Biomedicine studies.

Gene Ontology

The Gene Ontology, or GO, is a major bioinformatics initiative for unifying the representation of gene and gene product attributes across all species.More specifically, the GO aims to:

  • Maintain and develop its controlled vocabulary of gene and gene product attributes;
  • Annotate genes and gene products, and assimilate and disseminate annotation data;
  • Provide tools for easy access to all aspects of the data provided by the project.
  • The GO is part of a larger classification effort, the Open Biomedical Ontologies (OBO).

source: Wikipedia http://en.wikipedia.org/wiki/Gene_Ontology

Here we focus on the evaluation of tools to compute semantic measure scores considering the GO (www.geneontology.org).

Two tests have been performed:

  • Term to Term similarity computations, in which tools are compared by computing the semantic similarity of pairs of terms defined in the Gene Ontology.
  • Gene Product to Gene Product similarity computations, in which tools are compared by computing the semantic similarity of genes annotated by terms defined in the Gene Ontology.

Compared Tools

See the section dedicated to tools for more information (tool versions). We currently compare:

  • Semantic Measures Library Toolkit (SML)
  • GOSim
  • GOSemSim
  • FastSemSim

Quick Results

Below are the digested results. Details can be found in the following subsections.

  • 'X' means that the tests have not been performed, due to the performance of the tools.
  • '!' means that the at least one of the constraints has been reached and that the computation failed (constraints are specified, e.g., running time, amount of memory the tool can use). SML Par(4) corresponds to the SML configured with 4 threads, i.e. to enable parallel computation on multi-core CPU.
Term to Term test
1K 10K 1M 100M
FastSemSim 0m12.3 0m12.83 0m31.68 ! > 6Go memory
SML 0m9.23 0m9.76 0m19.55 16m30.24
SML Par(4) 0m9.22 0m9.56 0m14.47 8m58.29
GOSim 0m49.46 3m21.5 X X
GOSemSim 1m34.69 16m21.34 X X
Gene to Gene test
1K 10K 1M 100M
FastSemSim 0m13.36 0m16.79 7m8.14 !
SML 0m10.01 0m11.18 1m38.87 133m27.44
SML Par(4) 0m9.80 0m10.24 0m47.62 58m00.74
GOSim ! ! ! !
GOSemSim 27m02.66 ! ! !

Dataset

The datasets relative to this test are available at resources/data/go:

  • ontologies go/onto
  • annotations go/annot
  • benchmarks go/benchmarks (must be downloaded)
Gene Ontology

The version of the Gene Ontology used for this test is the lite version of 2013 03 02 (contained in the dataset but can also be downloaded at ftp://ftp.geneontology.org/pub/go/godatabase/archive/lite/2013-03-02/)

GOSim & GoSemSim depends on the GO.db package proposed by Bioconductor and are therefore constrained to use the GO loaded in the GO.db package. The version of Bioconductor which has been used in this test is version 2.12 see http://www.bioconductor.org/packages/2.12/data/annotation/manuals/GO.db/man/GO.db.pdf for details regarding the information relative to the GO loaded in Bioconductor 2.12.

FastSemSim loads any GO specified in OBO-XML and the SML is able to load the OWL/RDF-XML and OBO formatted versions.

The file which has been used for this tests is go_20130302-termdb.obo-xml.gz. It can be found in the directory /data/go/onto or at ftp://ftp.geneontology.org/pub/go/godatabase/archive/lite/2013-03-02/

Also note that we converted the obo.xml format to obo using go-perl package http://search.cpan.org/~cmungall/go-perl/ ./scripts/go2fmt.pl -w obo OBO_XML_FILE > OBO_FILE . The OBO file has been saved under the name go_20130302-termdb.obo. The hold_over_chain tags of typedef definitions have been commented as go-perl doesn't support them.

Annotations

GOSim and GOSemSim rely on the GO annotations defined in the R package org.Hs.eg.GO. See http://www.bioconductor.org/packages/2.12/data/annotation/manuals/org.Hs.eg.db/man/org.Hs.eg.db.pdf for more information on the version of the annotations used.

FastSemSim and the SML rely on GAF or plain annotation files. We encountered difficulties in downloading the GAF file associated to the annotations corresponding to the one loaded in the R package (on which rely GOSim and GOSemSim). In order to ensure that the annotations used for the evaluation are the same, we created a dump of org.Hs.eg.GO using the R script named dump_orgHsegGO.R. (/scripts/go/). This script generates a TSV file containing the GO annotation for all human genes specified in org.Hs.eg.GO.

The plain format required for the SML to import annotation from TSV file is slightly different from the one used by FastSemSim. The conversion of the dump generated by the R script has been made using the python script changeDumpAnnotationTSVFormat.py (/scripts/go/)

The annotation dumps used in the tests can be found at /resources/data/go/annot:

  • dump_orgHsegGO.tsv
  • dump_orgHsegGO_sml.tsv

Test 1: Term to Term computations

This test aims to compare the tools for the computation of semantic similarities between a pair of terms defined in the Gene Ontology. Four tests have been designed. Each test is composed of a set of pair of terms for which we want the semantic similarity to be computed. Four tests of different sizes have been generated:

  • 1k pairs of terms
  • 10k pairs of terms
  • 1M pairs of terms
  • 100M pairs of terms

For each test of size x, 3 random samples of size x have been generated in order to reduce the probability that the evaluation of the performance is biased by abnormal sampling. As an example, the test composed of 1K pairs of terms is composed of three different samples r0, r1, r2. For each sample (e.g. r1), three runs (r1.0, r1.1, r1.2) have been performed. This is to reduce the probability that results are biased by abnormal operating system behavior or material lags.

The sets of pairs of terms composing the 3 samples of each test have been generated using the tool sml-tools-evaluation-generate-go-benchmarks.jar. Both the tool and its source code can be found at scripts/go/. The command line used to generate the tests is:

java -jar sml-tools-evaluation-generate-go-benchmarks.jar ../../resources/data/go/onto/go_20130302-termdb.obo ../../resources/data/go/annot/dump_orgHsegGO_sml.tsv ../../resources/data/go/benchmarks/

This tool is used to generate benchmarks composed of pairs of GO terms. It has been used to generate TSV files containing pairs of GO terms identifiers (one per line). Four sizes of benchmarks are considered (1K, 10K, 1M and 100M). As we said, for each size, three sets of pairs are generated. The benchmarks have been built selecting random pairs of terms specified in the Biological Process aspect of the GO (all pairs of terms are composed of terms subsumed by the term GO:0008150). In addition, all terms which appear in the test are at least used to annotate a gene defined in dump_orgHsegGO.tsv. Indeed, some libraries cannot compute the similarity of terms which are not used to annotate at least one gene (This is due to the computation of Resnik's Information Content).

The samples used for the tests can be downloaded at http://www.semantic-measures-library.org/sml/downloads/evaluations/sm-tools-evaluation/resources/data/go/benchmarks/ They are expected to be decompressed and located in resources/data/go/benchmarks directory.

We selected Lin Information Content based (IC-based) measure to evaluate the performance of the tools. Lin is a commonly used measure to compare two concepts/terms defined in an ontology. It requires the Most Informative Common Ancestor of the compared terms and (by default) Resnik IC to be computed. These two treatments are the most time consuming of all IC-based measures (e.g. Resnik, Lin, SimRel) and IC-based measures are the most commonly used measures.

The script which is used to perform the test can be found at /scripts/go/run.sh. This script is used to launch the tests considering the tools have been installed and the dataset downloaded. If you try to reproduce the results, please edit the variables at the beginning of the script (e.g. installation and output directory). The script also specifies two constraints that can be modified by editing the script. This must be required depending on your hardware configuration. The constraint we considered are:

  • memory consumption: processes cannot use more than 6Go of memory
  • Time constraint: processes cannot take more than two hours

If these constraints are not respected, the execution of the program stops.

Due to their performance, GOSim and GOSemSim are not considered for the large tests. This can be modified editing the script run.sh.

GOSim and GOSemSim do not have command line interfaces. We therefore developed two scripts which can be used to compute all the similarities for the pairs of entries (terms or gene products) contained in a file. See scripts/go/GOSimWrapper.R and scripts/go/GOSemSimWrapper.R.

Perform the Test

Execute the script scripts/go/run.sh (see above for details regarding the script). We consider /results/go/output_go_benchmark_T2T.log as our log file (logging information will be printed in the file and the console).

./scripts/go/run.sh 2>&1 | tee /results/go/output_go_benchmark_T2T.log

This can take several hours depending on the constraints you specified and the hardware configuration. The script execution can be consulted using:

grep "\*\|user\|real" /results/go/output_go_benchmark_T2T.log

To extract the information relative to execution time and to store it into /tmp/output_go_benchmark.log.reduce

grep "\*\|user\|real" /tmp/output_go_benchmark.log  > /results/go/output_go_benchmark_T2T.log

The average values have been computed using a spreadsheet application.

Results

The tests have been performed on a Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz with 6Go allocated to the tools.

The detailed results for each run can be consulted at results/go/benchmark_result_pairwise.xlsx (ods). They have been extracted from the log file results/go/output_go_benchmark_GT2.log. Below the digested results, results for all runs are available in the spreadsheet.

  • 'X' means that the tests have not be performed (due to the performance of the tools).
  • '!' means that one of the constraints has been reached and that the computation failed. SML Par(4) corresponds to the SML configured with 4 threads, i.e. to enable parallel computation on multi-core CPU (only adding -threads 4 to the classical SML command line).
1K 10K 1M 100M
FastSemSim 0m12.3 0m12.83 0m31.68 ! > 6Go memory
SML 0m9.23 0m9.76 0m19.55 16m30.24
SML Par(4) 0m9.22 0m9.56 0m14.47 8m58.29
GOSim 0m49.46 3m21.5 X X
GOSemSim 1m34.69 16m21.34 X X
Correlations

We evaluated the Pearson correlations between the results obtained by the various libraries considering Lin measure. The correlations have been computed taking term to term 10000 r_0_0 sample into consideration. The details can be found in results/go/correlations_tools_r0_0.ods

FSS ISA corresponds to the results obtained using a special build of the FastSemSim library only considering is-a relationships, version 0.7.1.1 (see resources/tools/). This version is not an official release supported by Marco Mina, the developer of FastSemSim. This build has been made in order to change undesired behavior relative to the way version 0.7.1 compute parents/ancestors. Indeed, version 7.1 considers all types of relationship as isa/rdfs:subClassOf relationships when parents are computed. This behavior changes the common ancestors or the most informative common ancestor of two terms which will be considered by the measures. The results of this version for sample r0_0 can be found at results/go/r_10000_FastSemSim_7.1.1_0_0.tsv (then can also be reproduced modifying run.sh script).

GOSim and GOSemSim relies on GO.db R package http://www.bioconductor.org/packages/2.12/data/annotation/html/GO.db.html They also consider as ancestors of a concept x, concepts that are not subsuming x according to is-a relationships. See http://www.bioconductor.org/packages/2.13/data/annotation/manuals/GO.db/man/GO.db.pdf: GOBPPARENTS details:

"Each GO BP term is mapped to a named vector of GO BP terms. The name associated with the parent term will be either isa, hasa or partof, where isa indicates that the child term is a more specific version of the parent, and hasa and partof indicate that the child term is a part of the parent. For example, a telomere is part of a chromosome.".

We therefore suspect GOSim and GOSemSim to not differentiate the type of relationships when the common ancestors are computed.

The Pearson correlations between the results produced by the tools are:

FastSemSim FSS ISA SML GOSIM GOSEMSIM
FastSemSim 1 0.68 0.69 0.85 0.86
FSS ISA 1 0.99 0.58 0.58
SML 1 0.57 0.58
GOSim 1 0.99
GOSemSim 1

We observe that GOSIM and GOSemSim have a maximal Pearson correlation (0.99). Both tools rely on GO.db package. They also both have a strong correlation with FSS (0.85). The differences between GOSIM/GOSemSim and FastSemSim can be explained by the way the tools compute the information content.

The SML however produces scores which are faintly correlated to FSS, GOSim and GOSemSim. We investigated the results to understand which are the causes of the differences. We found that FSS, GOSim and GOSemSim perform treatments which are not in accordance with the original definition of Information Content based measures. Indeed, IC-based measures clearly rely on the taxonomic graph in order to be computed. The taxonomic graph is the subgraph of the ontology which only contains isa relationships (rdfs:subClassOf). This graph is considered to compute the ancestors of a term and is therefore important to compute the Most Informative Common Ancestor (or Disjoint Common Ancestors) in Information Content based measures. FastSemSim, GOSim and GOSemSim consider other relationships than taxonomic ones to compute the ancestors, which explains the variation obtained. They also consider part-of relationships (or even regulates in FSS) to define ancestors. To ensure that the poor correlations were due to this difference, we built a modified version the FastSemSim library (see build 0.7.1.1 in /resources/tools). This version can be used to compute the similarities using FSS source code and only considering is-a relationships when ancestors are computed. Considering this modification we obtained the expected correlation between FastSemSim and the SML (0.99).

Most result variations can be explained by differences between the various interpretations and implementations of measures proposed by the libraries.

Observations
  • Comparing random pairs of terms, FastSemSim produces a large number of values set to None, which cannot be exploited by other algorithms. This is not an error as it is due to the way FastSemSim computes Resnik Information Content. Remember that IC(term1) = -log(p(term1)) with p(term1) the probability term1 is used in the annotation repository. Considering that some terms (e.g. terminal terms) are not used in the repository, the probability that they occur is nil and their IC is set to infinity. By applying a strict implementation, FastSemSim refuses to process those pairs and is therefore limited to computing pairwise similarity measures (not that the IC cannot be changed). Refer to the class TermSemSim method int_validate_single_term of the module TermSemSim (if self.util.IC[term] == None: return None). However, this is not a problem for Gene to Gene comparisons as all pairwise computations involve terms which have been used by at least one gene.

  • Due to their performance GOSim and GOSemSim have been excluded from the large tests

  • GOSim and GOSemSim precompute the IC and can be used to handle large quantity of annotations. Loading all UniprotKB annotations to compute the IC using FastSemSim or the SML is currently not possible. This is due to the fact that both the SML and FastSemSim use in-memory loaded annotations. Note, however, that the IC must be computed regarding the application context. Indeed, if you study human genes, the IC must be computed only considering human gene annotations. FastSemSim and the SML are perfectly adapted to handle such use cases.


Test 2: Gene products comparison

This test aims to compare the tools to compute semantic similarities between pairs of gene products annotated by terms defined in the Gene Ontology. The protocol is similar to the one used for the comparison based on the computation of term similarity. Four tests have been designed. Each test is composed of a set of pairs of gene products for which we want the semantic similarity to be computed. Four sizes have been considered, 10k, 100k, 1M and 100M pairs of gene products.

The sets of pairs of gene products have been generated using the tool sml-tools-evaluation-generate-go-benchmarks.jar (open source, see above for more information).

No restriction is applied on the Evidence Code associated to the annotations linked to the considered gene product (e.g., IEA annotations are considered). In addition, only BP annotations were used during this test.

The constraints considered are:

  • memory consumption: processes cannot use more than 6Go of memory.
  • time constraint: processes cannot take more than four hours.
Results

The tests have been performed on a Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz with 6Go allocated to the tools.

The detailed results for each run can be consulted at results/go/benchmark_result_groupwise.ods. They have been extracted from the log file results/go/output_go_benchmark_G2G.log. Note that the first log /results/go/output_go_benchmark_G2G_O.log stores the errors for GOSim and GOSemSim.

Below the digested results. 'X' means that the tests have not been performed (due to the performance of the tools). '!' means that the constraints have been reached and that the computation failed. SML Par(4) corresponds to the SML configured with 4 threads (add -threads 4 to the classical SML command line).

1K 10K 1M 100M
FastSemSim 0m13.36 0m16.79 7m8.14 !
SML 0m10.01 0m11.18 1m38.87 133m27.44
SML Par(4) 0m9.80 0m10.24 0m47.62 58m00.74
GOSim ! ! ! !
GOSemSim 27m02.66 ! ! !

List of Tools

This listing contains the tools which have been included in the evaluation. The tools are available at ./resources/tools, please refer to the corresponding documentation for the installation.

Semantic Measures Library (and Toolkit)

url: http://www.semantic-measures-library.org

version: 0.7

Dependencies

Java 1.7. In the tests we used:

java -version
java version "1.7.0_21"
Java(TM) SE Runtime Environment (build 1.7.0_21-b11)
Java HotSpot(TM) 64-Bit Server VM (build 23.21-b01, mixed mode)
Installation

None. The SML-Toolkit is ready to use.

GOSim

url: http://cran.r-project.org/src/contrib/Archive/GOSim/

version: 1.2.7.7

Dependencies

In the tests we used:

R --version
R version 3.0.1 (2013-05-16) -- "Good Sport"
Copyright (C) 2013 The R Foundation for Statistical Computing
Platform: x86_64-pc-linux-gnu (64-bit)

Upgrade R version if required (Ubuntu users: http://cran.cnr.berkeley.edu/bin/linux/ubuntu/).

Installation

In R console:

First install dependencies, they are specified when you try to install the software.

source("http://bioconductor.org/biocLite.R")
biocLite(c("annotate", "topGO", ...))

Next install the corresponding package (change the location of the package) install.packages(".../GOSim_1.2.7.7.tar.gz", repos = NULL, type ="source")

GOSemSim

url: http://www.bioconductor.org/packages/release/bioc/html/GOSemSim.html

version: 1.18.0

Dependencies

In the tests we used:

R --version
R version 3.0.1 (2013-05-16) -- "Good Sport"
Copyright (C) 2013 The R Foundation for Statistical Computing
Platform: x86_64-pc-linux-gnu (64-bit)

Upgrade R version if required (Ubuntu users: http://cran.cnr.berkeley.edu/bin/linux/ubuntu/).

Installation

In R console:

First install dependencies, they are specified when you try to install the software.

install.packages(c("Rcpp","igraph","flexmix", "RBGL", "graph", "corpcor", "org.Hs.eg.db"))
source("http://bioconductor.org/biocLite.R")
biocLite(c("GO.db", "AnnotationDbi", "annotate", "topGO"))

Next install the corresponding package (change the location of the package). install.packages(".../GOSemSim_1.18.0.tar.gz", repos = NULL, type ="source")

FastSemSim

url: http://sourceforge.net/projects/fastsemsim/

version: 0.7.1

Dependencies

We used:

python --version
Python 2.7.3
Installation

Considering Python is already installed, unzip the archive and as root run ./install.sh

Competing interests

This project has been initiated in order to evaluate the Semantic Measures Library www.semantic-measures-library.org. As the developers of these tests are also developers of the SML we cannot ensure that this evaluation is free from bias. Indeed we have a better knowledge of how to configure and use the SML than other tools. Do not hesitate to help us improve these tests!

Contributors

  • Sébastien Harispe - PhD. candidate LGI2P research center
  • Sylvie Ranwez - PhD. LGI2P research center
  • Stefan Janaqi - PhD. LGI2P research center
  • Jacky Montmain - Professor LGI2P research center

About

Evaluation of tools for Semantic Measures computation

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published