Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Using Workflows to Explore and Optimise Named Entity Recognition for Chemistry

  • BalaKrishna Kolluru ,

    balakrishna.kolluru@manchester.ac.uk

    Affiliation National Centre for Text Mining, Manchester Interdisciplinary Biocentre, University of Manchester, Manchester, United Kingdom

  • Lezan Hawizy,

    Affiliation Unilever Centre for Molecular Informatics, University of Cambridge, Cambridge, United Kingdom

  • Peter Murray-Rust,

    Affiliation Unilever Centre for Molecular Informatics, University of Cambridge, Cambridge, United Kingdom

  • Junichi Tsujii,

    Affiliation National Centre for Text Mining, Manchester Interdisciplinary Biocentre, University of Manchester, Manchester, United Kingdom

  • Sophia Ananiadou

    Affiliation National Centre for Text Mining, Manchester Interdisciplinary Biocentre, University of Manchester, Manchester, United Kingdom

Abstract

Chemistry text mining tools should be interoperable and adaptable regardless of system-level implementation, installation or even programming issues. We aim to abstract the functionality of these tools from the underlying implementation via reconfigurable workflows for automatically identifying chemical names. To achieve this, we refactored an established named entity recogniser (in the chemistry domain), OSCAR and studied the impact of each component on the net performance. We developed two reconfigurable workflows from OSCAR using an interoperable text mining framework, U-Compare. These workflows can be altered using the drag-&-drop mechanism of the graphical user interface of U-Compare. These workflows also provide a platform to study the relationship between text mining components such as tokenisation and named entity recognition (using maximum entropy Markov model (MEMM) and pattern recognition based classifiers). Results indicate that, for chemistry in particular, eliminating noise generated by tokenisation techniques lead to a slightly better performance than others, in terms of named entity recognition (NER) accuracy. Poor tokenisation translates into poorer input to the classifier components which in turn leads to an increase in Type I or Type II errors, thus, lowering the overall performance. On the Sciborg corpus, the workflow based system, which uses a new tokeniser whilst retaining the same MEMM component, increases the F-score from 82.35% to 84.44%. On the PubMed corpus, it recorded an F-score of 84.84% as against 84.23% by OSCAR.

Introduction

Text mining for the domain of chemistry is a very challenging task because of the several semantic and syntactic styles in which domain texts are usually expressed. Different aspects such as named entity recognition (NER), tokenisation and acronym detection require bespoke approaches because the complex nature of such texts [1][5]. Chemical compounds such as:

17--hydroxy-16--methyl-3,20-dioxopregna-1,4-dien-21-yl acetate

P(Cy)3

1-cyclopropyl-6-fluoro-4-oxo-7-(piperazin-1-yl)-1,

4-dihydroquinoline-3-carboxylic acid hydrochloride

illustrate the complexity of the mining task. Typical word delimiters such as spaces, brackets, hyphens and commas cease to bear the same meaning as in a natural language. As a consequence, the normal text mining approaches such as tokenisers, part-of-speech (POS) taggers and parsers will need to be re-calibrated for this domain as already done for other domains such as biochemistry, biomedicine etc., [6], [7].

In the chemistry domain, researchers have presented a few successful approaches to handle some tasks such as named entity recognition [8][13]. However, these approaches usually require reconfiguring and sometimes rewriting everytime a new training corpus or dictionary is released [14], [15]; typically this could be due to different data format or additional information in the new resource. For example, if the new resource is in a different format, the whole system or at least a part of it may need to be rewritten. With the growing number of freely available resources such as Chemspider (http://www.chemspider.com/), Chemlist [14] and [5], [16][18] etc., the ability to reconfigure the systems becomes more acute. Such reconfiguring takes time and the subtle changes in the throughputs of these components, which may seem innocuous, could result in the lowering of the net performance of a system; this could be a direct consequence of a suboptimal composition of the workflow. Therefore, it is imperative to configure the optimal set by exploring the various manifestations of the different components [19]. To be able to arrive at an optimum combination of components, one has to substitute one component for another in a workflow and then assess if the performance has indeed improved. This warrants an understanding of inter-component relations working together as a system. It would also be desirable if components using different machine learning techniques could easily be replaced to observe differences in performance. This ability to reconfigure an approach has the advantage of allowing scientists to concentrate more on science rather than format conversion and code refactoring. Usage of workflows for chemistry and its related disciplines has been pursued very actively in the community [20][23]. Thus, there is already good familiarity, if not expectation, of this methodology. For the experiments discussed in this paper, we implement reconfigurable workflows that are interchangeable by drag-&-drop on the graphical user interface. To do this we employ U-Compare [24]: an open UIMA-based [25] framework (http://incubator.apache.org/uima/[26]) which allows shareable components, using a common type system, to be used together to form different workflows. In doing so, we also design an interoperable type system for UIMA-compliant systems.

Figure 1 (b) illustrates the composition of a reconfigurable workflow system, wherein one component can be substituted by another component from a repository.

thumbnail
Figure 1. Showing a normal workflow and a reconfigurable workflow as can be built by using U-Compare.

https://doi.org/10.1371/journal.pone.0020181.g001

U-Compare framework provides a platform for reconfigurable workflow experiments. Its UIMA-based framework provides the necessary component repository, consisting of several shareable components such as Genia tagger [7], Stepp tagger [27] and OpenNLP [28] sentence splitters, for other UIMA-based components. This extensive repository readily allows for several combinations of components as workflows.

As a consequence, we use a set of individual components to handle the different aspects of text mining, which together form a workflow.

Related work for workflows

The use of several individual components to construct workflows is quite prevalent amongst the scientific community with interdisciplinary sciences [29]. Bespoke workflows have been employed in several domains such as bio-informatics and earth sciences. Some studies have introduced workflow tools as a Lego-like setup [30], wherein several simple components form a complex workflow which can be easily deployed, modified and tested without the overhead of implementing it into a monolithic application. Taverna [31] is such a workflow management suite for building scientific workflows which offers loosely-coupled services. Kepler [32] is another workflow management system for designing, executing, processing and sharing scientific workflows. The workflows in this context are directed graphs where the nodes represent components, the edges represent data paths along which data and results can flow between components.

There are also commercial products which provide environments to create and manage workflows. Pipeline Pilot [33] is an example of a commercial application that combines workflows with data analysis to represent information visually for informatics and scientific business intelligence needs. Pipeline Pilot has been extended to track bibliography in chemistry literature using a web-based graphical user interface [34].

The UIMA platform introduced a new framework for developing shareable components into a repository. Mellebeek et al. show the usage of UIMA and text mining applications for curation purposes in the domain of bio-informatics [35]. In doing so, they demonstrate the possible synergy from a combination of diverse expertise in biology, computer science and linguistics. Their application was fundamental to the development of a successful curation tool. The U-Compare [36], based on the UIMA Framework, is an integrated text mining system which provides a graphical user interface for easy drag-&-drop workflow creation. It has built-in tools for evaluation and visualisations of components and also has a number of syntactic and semantic tools to generate workflows. Kano et al. showed the advantages of using workflows in U-Compare framework by developing a protein-protein interaction extraction system [37].

Our paper presents a similar workflow to [37] but in the domain of chemistry. As a first step, we used Oscar3 [38] to extract chemical named entities from the literature. Subsequently, we segregated Oscar3 into separate components. Townsend et al. have developed a methodology and a workflow (CHIC) for the automatic semantic enrichment and structuring of legacy scientific documents by using Oscar3[38]. U-Compare has a plug-in for Taverna [37] which implicitly means the workflows discussed here can be ported to Taverna, which increases the audience and applicability of our workflows.

In the area of chemistry and text mining, Wilbur et al. employed two approaches with an aim of separating chemical terms from non-chemical terms [39]:

  1. thesaurus-based lexical text analysis using chemical patterns
  2. Bayesian classification using n-grams

They found that the Bayesian approach had an overall classification accuracy of 97%, while the thesaurus-based method had an accuracy of 84%. While the work by Wilbur et al. operates on individual words (or entities) based on thesaurus-style lists [39], the work described here processes full papers (and abstracts), tokenizes them for analysis and classifies chemical compounds found in the text.

Materials and Methods

In this paper, our principal task was to elicit chemical compounds from free-flowing text in the chemistry literature. We have used the Sciborg [40] and PubMed [41] corpora for this task.

Sciborg Corpus

This corpus was compiled as part of the Sciborg project [42]. It consists of 42 articles (full papers) published in the chemical literature which were provided by the Royal Society for Chemistry (RSC). It was curated for linguistic analysis by [41]. This corpus was split randomly into two groups of 14 and 28 papers, such that they form two disjoint testing and training sets respectively. MEMM models (discussed later in paper) were trained on the set of 28 papers having 4102 manually annotated chemical compounds and the 14 papers were used as a test set. The test set was hand-annotated by three chemistry experts and an inter-annotator agreement () of 0.91 was observed on this set.

PubMed Corpus

This corpus was compiled for linguistic analysis by Corbett et al. [41]. It had 500 abstracts from the PubMed [43] collection. This corpus was randomly split into 400 and 100 abstracts for training and test sets respectively. MEMM models were trained on the 400 abstracts consisting of 4048 annotations of chemical compounds. The test set was hand-annotated by one expert.

Models

For the MEMM-based component of our approach we experimented with two models,

  • chempaper-M: trained on Sciborg training data (28 papers)
  • pubmed-M: trained on 400 PubMed abstracts

Overview of Oscar3

Oscar3 is an open extensible system for the automated annotation of chemical entities in scientific articles [9]; it was created as part of the Sciborg project [40]. The overall architecture of Oscar3 is shown in Figure 2 and the individual components are discussed below:

SciXML.

SciXML is the interface used when working with Oscar3, all forms of input (such as XML, HTML and plain text) are converted into this format before any processing is done. It is a form of XML markup used for providing logical structure to scientific papers. Further information about SciXML and its schema can be found in [40].

Tokeniser.

The tokenisation with Oscar3 is chemistry specific; chemical names are fragile to common methods of tokenisation as they contain potential inter token and intra token characters such as space, hyphens, brackets and comma. The tokeniser here also refers back to the SciXML document to store information about the start and end points of a token as well as its content. For example, some of the tokens in data are:

aztreonamMetallo--lactamase

Cu2

C2(MONO)

CO/CN

ZnO3S(monobactam)

Chemical Entity Recognisers.

Oscar3 contains two types of chemical entity recognisers, each producing a list of named entities as an output containing chemical annotations, such as token, types and likelihood scores (where applicable).

Pattern Recogniser.

This recogniser was initially used before the machine learning component was introduced. It uses deterministic finite state automata alongside ontologies (such as CHEBI [44]), dictionaries and n-gram models to recognise the named entities. As it relies on the regular expression based rules, it does not use any mathematical models for classification.

MEMM Recogniser.

This recogniser uses MEMM and character level n-grams to recognise chemical entities based on their likelihoods. The MEMM was trained using the annotated corpora discussed earlier. Corbett et al. reported an F-score of 80.7% for a model trained on Sciborg and PubMed training sets at a confidence threshold of 0.3 [41].

Why the new Oscar?

Oscar3 is an efficient annotation tool and is widely used within the chemistry domain. However, the architecture is rigid and, due to its dependency on the SciXML format and the interdependency within the different components, it is difficult to modularise and it does not readily adapt to new and emerging trends in annotation and corpora. This puts a limitation on enabling and refactoring reusable components.

Oscar3 as a Workflow of Reconfigurable Components

Conceptually, Oscar3 [9] is a named entity recogniser which classifies tokens into chemical entities based on either likelihoods or a pattern match. Therefore, Oscar3 was divided into the following components as shown in Figure 3. This is just one of the many possible manifestations of the workflows; other configurations such as different tokenisers and components implementing machine learning techniques can be easily accommodated to make a new workflow.

thumbnail
Figure 3. Oscar3 refactored as a workflow of different components.

https://doi.org/10.1371/journal.pone.0020181.g003

  • A tokeniser: This tokeniser is a white-space delimited, word eliciting component which reads content from files in text, XML or HTML and yields tokens similar to the syntactic token of the U-Compare type system [45]. It must be noted here that any tokeniser that yields the syntactic tokens for a given source file can be used as the first stage of the Oscar workflow.
  • A MEMM Component: trained on two chemistry-specific corpora (as mentioned in the data section).
  • A Pattern matching Component: based on a finite state automaton driven regular expression matcher; the rules of which were designed after several observations of the training data. As a consequence, this component does not use any statistical models for classification purposes.

Shown in Figure 4 is one of the workflows (left in the figure) using three components (right in the figure): a file system component to read the files which are then split into individual tokens by the OscarTokeniser component which subsequently feeds into the OscarMER component to classify the tokens into chemical names.

thumbnail
Figure 4. U-Compare view of Oscar workflow.

Right side of the figure shows a workflow made from the Oscar components shown on the left.

https://doi.org/10.1371/journal.pone.0020181.g004

Results

The experiments described earlier with Oscar3 and its refactored version were designed to study the effect of tokenisation on chemical element identification. We present the results of modularising Oscar3 and compare it with the existing version. Also, to present the robustness of the workflow, we compare the performance of Oscar3 on the corpora described in the data section.

Reconfiguring Oscar3: a confidence-driven approach

The machine learning components used by the two variations of Oscar3 (Oscar3 stand-alone version and Oscar workflow) yield a confidence score, which is a likelihood estimate (see [41] for more details), to show the confidence in that annotation. In order to arrive at an optimum threshold for each of the corpora, we have plotted the ROC (A ROC curve is a receiver operating characteristic which plots the rate of true positives against the rate of false positives.) curves for each of the data sets. Shown in Figure 5 are the ROC curves for different combinations of data sets and Oscar3 variants.

thumbnail
Figure 5. ROC curves comparing the performance of various Oscar variants.

In all the four different experiments, Oscar workflow has a slight edge over the Oscar 3 variant.

https://doi.org/10.1371/journal.pone.0020181.g005

Oscar vs. Oscar: One Variant against Another

As described earlier, currently there are two types of named entity recognisers, a MEMM-based and a pattern matching one. The MEMM-based versions were tested with two models; chempaper-M and pubmed-M.

The results are presented in terms of percentages of precision (P), recall (R) and F score (F). Table 1. shows the overall performance of both the variants of Oscar3 on Sciborg test data. The MEMM-driven systems were tested using both models (chempaper-M and pubmed-M) which are trained on Sciborg and PubMed training data respectively.

thumbnail
Table 1. Performance (%) of different variants of Oscar on Sciborg test data using the models trained on Sciborg data and PubMed data.

https://doi.org/10.1371/journal.pone.0020181.t001

Table 2. shows the performances of Oscar3 with pattern recogniser (Oscar3 (PAT)) and as a workflow with pattern recogniser (Oscar workflow (PAT)) on the Sciborg test data.

thumbnail
Table 2. Performance of different Oscar pattern recogniser versions on Sciborg.

https://doi.org/10.1371/journal.pone.0020181.t002

As described in Corbett et al. [41], Oscar3 can be tuned to filter out some false positives (Type I) errors based on a confidence score derived from the logit scores (see [41] for more details). At a confidence score of 0.42, Oscar3 as a workflow with MEMM recorded an F-score of 84.84% while Oscar3 with MEMM recorded 82.35% on the Sciborg data. It is also noteworthy that although the pattern recognition variants were less accurate than their MEMM counterparts, the workflow variant still outperforms its monolithic parent.

Table 3 shows the performance of the Oscar3 variants when used on the PubMed test set against models trained on Sciborg and PubMed training data sets. It can be observed that a different tokeniser gives an extra boost of 0.61% (84.84 by the workflow MEMM variant as against 84.23 by the Oscar3 variant) whilst retaining the same machine learning component and the model.

thumbnail
Table 3. Performance of different variants of Oscar on PubMed test data using the models trained on Sciborg data and PubMed data.

https://doi.org/10.1371/journal.pone.0020181.t003

Table 4 shows the performance of pattern-recognition based variants of Oscar3 and Oscar3 workflow on the Pubmed data. Although having lower scores than the MEMM variants, Oscar workflow (PAT) outperforms the Oscar3 (PAT) by 1.7%.

thumbnail
Table 4. Performance of different Oscar pattern recogniser versions on Pubmed.

https://doi.org/10.1371/journal.pone.0020181.t004

Wren used the single-order Markov models to distinguish between chemical and non-chemical terms on Medline [46] corpus with an average precision of about 82.7% [10]. The work described here uses maximum entropy Markov models on 2 different corpora: Sciborg and PubMed. For this corpus, our approach recorded a precision of 90.31% and a recall of 85.66%. As the work by Wren ([10] ) and our approach vary on the corpora and methodology, we do not see it fair to compare head-to-head; however, our system does perform as well, if not better.

Discussion

To observe the efficacy of workflows, we have used two sets of workflows, one each in pattern recognition based variant and the MEMM variant. As shown in Tables 1 and 2, the workflow variants of Oscar3 achieve better performance the two variants.

On the Sciborg, the workflow-based MEMM model achieved an F-score of 84.44% as opposed to 82.35% for Oscar3. We observed that this increase was due to removal of the dependency on SciXML conversions within the workflow.

A reconfigurable approach enabled us to identify the erroneous (or underperforming) component and relate some of the errors to severe dependency on SciXML conversions, when using chempaper-M. We infer that there was a net increase in false positives due to the noise in several inter-conversions of formats in SciXML. It should be noted here that the dependence of Oscar3 on SciXML was due to them both being part of the Sciborg project. This dependency could make it difficult to adapt to newer corpora. It was observed that the new tokenisation identified more chemical words such as

-lactamzn2+

bis-monodentate

goldsulfur

which automatically led to a decrease in false positives (Type I errors). It also avoided wrongly tokenizing a few words such as,

diimine

mono-bidentate

which were subsequently omitted as non-chemistry words by the Oscar3 but accurately identified by the workflow variant. In this example, the complete chemical was ruthenium (ii) diimine, but Oscar3 returned only ruthenium(ii) as CM, whilst the workflow version got the complete entity. by the Oscar3 but accurately identified by the workflow variant. This led to fewer false negatives (Type II errors) and hence a better recall.

As the machine learning classifiers and the models they used were exactly the same for all experiments, we infer that tokenisation on the Sciborg test avoided partial entities for recognition and this helped reduce both Type I and Type II errors.

Figure 6 shows the chemical names as annotated by the Oscar workflow in the U-Compare framework. When these entities (which are underlined) are clicked, more information about the entity such as confidence scores, metadata etc. is available to the user.

thumbnail
Figure 6. U-Compare output for a test document.

Chemical names (underlined) as identified by the MEMM-based workflow.

https://doi.org/10.1371/journal.pone.0020181.g006

Table 3 shows a decrease in precision of 3% with an increase in recall of 4% for the PubMed test data. Perhaps, this could be attributed to the possible shortcomings in the ability of the new tokeniser to adapt to the biochemical entities; we are working on enhancing the tokeniser to suit multiple domains where chemistry plays an important role.

The current version of Oscar3, which can be downloaded from http://sourceforge.net/projects/oscar3-chem/, had an F-score of 82.35% on Sciborg as against 80.7% achieved by [41]. This could be due to the fact that [41] used a 3-fold cross validation, whilst we used only 1 combination. The usage of one training set and one test set, instead of multi-fold cross validation, was guided by the focus of our paper, namely: the advantages of workflows for text mining in chemistry. Also, as described in the Data section, the test set comprised of 14 full papers, manually annotated by three experts, whilst, the training set was annotated by a single expert. We conjecture that this test set had enough data points to support our inferences.

On the Sciborg (Table 1), the pattern-recognition based workflow achieved a precision of 66.32% while the Oscar3 using the pattern recognition module achieved a precision of 44.65%. Again, it seems the only difference between the two variants was the tokenisation which stems from issues relating to SciXML conversions. This could be perceived as an example of having an optimal combination in a workflow to derive a better performance.

The results indicate the success of workflows described in our experiments discussed earlier. Currently, we are in the process of converting implementations of other machine learning algorithms such as Conditional Random Fields (CRF) into the U-Compare framework. This will enable us to compare the performance of different algorithms on data sets. Every time a new annotation scheme is announced, it obliges the existing applications to adapt, sometimes subtly and at times extensively. We have shown that a reconfigurable system (or application) is better for such adaptation.

Conclusions

We have shown that, using a reconfigurable workflow, it is possible to assess different components in a system to elicit the best combination. As a consequence, it helps users to focus less on system implementation issues. Using these workflows, we studied the impact of using different tokenisation techniques on the task of named entity recognition in chemistry. The potential for expanding the scope of inter-component analysis is immense and more so, with complex systems involving several components. We have demonstrated the impact of tokenisation in recognising complex named entities in chemistry, wherein a named entity may contain two, three or even four words with numerals, Greek letters, punctuation marks, etc. Work is currently underway to make a CRF component so that one can freely replace MEMM models with a CRF model and thus benefit from a pool of machine learning algorithms for various tasks, named entity recognition being one of them. We are also working on workflows to combine a set of taggers and named entity recognisers for application in the domain of chemistry, biochemistry and biological sciences [47].

Acknowledgments

We would like to thank Richard Kidd and Colin Batchelor of the Royal Society for Chemistry for providing the Sciborg corpus, as well as their insight into the annotation guidelines. Also, we would like to thank Dr. Paul Dobson for valuable comments, and Dr. John McNaught for the numerous edits of our paper. Last, but not least, we would like to thank Dr. Yoshinobu Kano for his help with U-Compare.

Author Contributions

Conceived and designed the experiments: BK LH. Performed the experiments: BK LH. Analyzed the data: BK LH SA PM JT. Wrote the paper: BK LH. Supervised the experiments: PM SA JT.

References

  1. 1. Kemp N, Lynch M (1998) Extraction of information from the text of chemical patents. 1. identification of specific chemical names. Journal of Chemical Information and Computer Sciences 4: 544–551.
  2. 2. Murray-Rust P, Rzepa H (1999) Chemical markup, xml, and the worldwide web. 1. basic principles. Journal of Chemical Information and Computer Sciences 39: 928–942.
  3. 3. Murray-Rust P, Mitchell J, Rzepa H (2005) Chemistry in bioinformatics. BMC Bioinformatics 6: 141.
  4. 4. Banville D (2006) Mining chemical structural information from the drug literature. Drug Discovery Today 11: 35–42.
  5. 5. Kolrik C, Hofmann-Apitius M, Zimmermann M, Fluck J (2007) Identification of new drug classification terms in textual resources. Bioinformatics 13: 264–272.
  6. 6. Miyao Y, Tsujii J (2005) Probabilistic disambiguation models for wide-coverage hpsg parsing. ACL-2005. pp. 83–90.
  7. 7. Tsuruoka Y, Tateishi Y, Kim J, Ohta T, McNaught J, et al. (2005) Developing a robust part-ofspeech tagger for biomedical text. In: Bozanis P, Houstis EN, editors. volume 3746, chapter 36. Berlin Heidelberg: Advances in Informatics, Springer. pp. 382–392. https://doi.org/http://dx.doi.org/10.1007/11573036
  8. 8. Steinbeck C, Hoppe C, Kuhn S, Floris M, Guha R, et al. (2006) Recent developments of thechemistry development kit (cdk) - an open-source java library for chemo- and bioinformatics. Current Pharmaceutical Design. pp. 2111–2120.
  9. 9. Corbett P, Murray-Rust P (2006) High-throughput identification of chemistry in life science texts.In: Computational Life Sciences II.107–118.
  10. 10. Wren J (2006) A scalable machine learning approach to recognize chemical names within large textdatabases. BMC Bioinformatics 7: S3.
  11. 11. Florian B, Juan MTM, El-Bèze M (2008) Mixing statistical and symbolic approaches for chemicalnames recognition. In: CICLing. pp. 334–343.
  12. 12. Klinger R, Kolarik C, Fluck J, Hofmann-Apitius M, Friedrich C (2008) Detection of IUPAC and IUPAC-like Chemical Names. Bioinformatics 24: i268–276.
  13. 13. Jiao D, Wild DJJ (2009) Extraction of cyp chemical interactions from biomedical literature using natural language processing methods. Journal of chemical information and modeling 49: 263–269.
  14. 14. Hettne K, Stierum R, Schuemie M, Hendriksen P, Schijvenaars B, et al. (2009) A dictionary to identify small molecules and drugs in free text. Bioinformatics 25: 2983–2991.
  15. 15. Hettne K, Williams A, van Mulligen E, Kleinjans J, Tkachenko V, et al. (2010) Automatic vs. manual curation of a multi-source chemical dictionary: The impact on text mining. Journal of Cheminformatics 2: 4.
  16. 16. Kolarik C, Klinger R, Friedrich C, Hofmann-Apitius M, Fluck J (2008) Chemical names: Terminological resources and corpora annotation. In: Workshop on Building and evaluating resources for biomedical text mining, 6th edition LREC.
  17. 17. Klinger R, Friedrich C, Hofmann-Apitius M, Fluck J, Birlinghoven S (2009) Chemical names: Terminological resources and corpora annotation.
  18. 18. Muller B, Klinger R, Gurulingappa H, Mevissen H, Hofmann-Apitius M, et al. (2010) Abstractsversus full texts and patents: A quantitative analysis of biomedical entities. Advances in Multidisciplinary Retrieval. volume 6107, chapter 12. Berlin, Heidelberg: Springer Berlin Heidelberg,. pp. 152–165. https://doi.org/http://dx.doi.org/10.1007/978-3-642-13084-7_12
  19. 19. Rupp CJ, Copestake A, Corbett P, Waldron B (2007) Integrating general-purpose and domainspecific components in the analysis of scientific text.
  20. 20. Hassan M, Brown R, Varma-O'brien S, Rogers D (2006) Cheminformatics analysis and learning in a data pipelining environment. Molecular diversity 10: 283–299.
  21. 21. Tiwari A, Sekhar A (2008) Workflow based framework for life science informatics. Computational Biology and Chemistry 31: 306–319.
  22. 22. Shon J, Ohkawa H, Hammer J (2008) Scientific workflows as productivity tools for drug discovery. Current opinion in drug discovery and development 11: 381–388.
  23. 23. Kuhn T, Willighagen E, Zielesny1 A, Steinbeck C (2010) Cdk-taverna: An open workflow environment for cheminformatics. Bioinformatics 11:
  24. 24. Kano Y, Baumgartner W, McCrohon L, Ananiadou S, Cohen K, et al. (2009) U-compare: Share and compare text mining tools with uima. Bioinformatics 25: 1997–1998.
  25. 25. Ferrucci D, Lally A, Gruhl D, Epstein E, Schor M, et al. (2006) Towards an interoperabilitystandard for text and multi-modal analytics. Technical report, IBM.
  26. 26. Apache (2010) Unstructured information management. URL http://uima.apache.org/. [last accessed on 02-May-2011].
  27. 27. Tsuruoka Y, Tsujii J, Ananiadou S (2009) Fast full parsing by linear-chain conditional random fields. In: EACL '09: Proceedings of the 12th Conference of the European Chapter of the Association for Computational Linguistics. Morristown, NJ, USA: Association for Computational Linguistics. pp. 790–798.
  28. 28. OpenNLP (2010) Opennlp. URL http://incubator.apache.org/opennlp/. [last accessed on 02-May-2011].
  29. 29. Taylor I, Deelman E, Gannon D, Shields ME (2007) Workflows for e-Science: Scientific Workflows for Grids. Springer.
  30. 30. Kuhn T, Zielesny A, Steinbeck C (2009) Creating chemo- and bioinformatics workflows, further developments within the cdk-taverna project. Chemistry Central Journal 3: 42.
  31. 31. Oinn T, Addis M, Ferris J, Marvin D, Senger M, et al. (2004) Taverna: A tool for the composition and enactment of bioinformatics workflows. Bioinformatics 20: 3045–3054.
  32. 32. Ludascher B, Altintas I, Berkley C, Higgins D, Jaeger-Frank E, et al. (2006) Scientific workflow management and the kepler system. Special Issue: Workflow in Grid Systems Concurrency and Computation: Practice & Experience 18: 1039–1065.
  33. 33. Accelrys' (2010) Pipeline pilot. URL http://accelrys.com/products/pipeline-pilot/. [last accessed on 02-May-2011].
  34. 34. Vellay SG, Latimer ME, Paillard G (2009) Interactive text mining with pipeline pilot: A bibliographic web-based tool for pubmed. Infectious disorders drug targets 9: 366–374.
  35. 35. Mellebeek B, Rodriguez-Penagos C, Furlong LI (2009) Uima in the biocuration workflow: A coherent framework for cooperation between biologists and computational linguists. Nature Precedings.
  36. 36. Kano Y, N N, R S, Fukamachi K, Kazuhiro Y, et al. (2008) Sharable type system design for tool inter-operability and combinatorial comparison. In: Proceedings of the First International Conference on Global Interoperability for Language Resources (ICGL). Hong Kong, 122-129:
  37. 37. Kano Y, Dobson P, Nakanishi JM, Tsujii , Ananiadou S (2010) Text mining meets workflow: Linking u-compare with taverna. Bioinformatics.
  38. 38. Townsend JA, Downing J, Murray-Rust P (2009) Chic - converting hamburgers into cows. In: Proceedings of the 2009 Fifth IEEE International Conference on e-Science. Washington, DC, USA: IEEE Computer Society, E-SCIENCE '09. pp. 337–343. https://doi.org/http://dx.doi.org/10.1109/e-Sciencehttp://dx.doi.org/10.1109/e-Science.2009.54
  39. 39. Wilbur WJ, Hazard GF, Divita G, Mork JG, Aronson AR, et al. (1999) Analysis of biomedical text for chemical names: A comparison of three methods. AMIA Symposium. pp. 176–180.
  40. 40. Rupp CJ, Copestake A, Teufel S, Waldron B (2006) Flexible interfaces in the application of language technology to an escience corpus.
  41. 41. Corbett P, Copestake A (2008) Cascaded classifiers for confidence-based chemical named entity recognition. BMC Bioinformatics 9: S4.
  42. 42. Copestake A, Teufel S, Murray-Rust P, Parker A (2010) Extracting the science from scientific publications. URL http://www.cl.cam.ac.uk/research/nl/sciborg/www/. [last accessed on 02-May-2011].
  43. 43. NIH (2010) Pubmed. URL http://www.ncbi.nlm.nih.gov/pubmed/. [last accessed on 02-May-2011].
  44. 44. EBI (2010) Chemical entities of biological interest (chebi). URL http://www.ebi.ac.uk/chebi/. [last accessed on 02-May-2011].
  45. 45. Kano Y, McCrohon L, Ananiadou S, Tsujii J (2009) Integrated nlp evaluation system for pluggable evaluation metrics with extensive interoperable toolkit. In: SETQA-NLP '09: Proceedings of the Workshop on Software Engineering, Testing, and Quality Assurance for Natural Language Processing. Morristown, NJ: Association for Computational Linguistics. pp. 22–30.
  46. 46. Medline (2010) Medline. URL http://www.nlm.nih.gov/databases/databases_medline.html. [last accessed on 02-May-2011].
  47. 47. Nobata C, Sasaki Y, Okazaki N, Rupp CJ, Tsujii J, et al. (2009) Semantic search on digital document repositories based on text mining results.