Question

I want to choose a natural language processing tool to do common tasks such as tokenization, sentence detection,various tagging (Name Entity Recognition, POS tagging, ... ). my question has two parts:

  1. What are the criteria for choosing a natural language processing tool?
  2. Among (UIMA, LingPipe, Lucene, Gate, Stanford), which one satisfy these criteria better?

and what is your suggestion ?

Was it helpful?

Solution

Some general Criteria:

  1. how many tasks can I perform with the provided models (e.g. does the tool contains models for my tasks like spanish tokenisation or protein NER)?
  2. how easy is it for me to add the missing tools.

BTW, I would add NLTK to your list, and its excellent, free accompanying book.

Licensed under: CC-BY-SA with attribution
Not affiliated with StackOverflow
scroll top