I am currently using Freeling from Python API to perform sentiment analysis in a research project. More presicely, I use Freeling for tokenization, lemmatization and sense disambiguation.
Freeling gives synsets as WordNet offsets (e. g., '01123148-a'), if I am not wrong. Then, I need to look in the SentiWordNet database for synsets polarities, but to do so I need to represent synsets with their name (e.g. : 'good.a.01') to use them in SentiWordNet, as explained there : http://www.nltk.org/howto/sentiwordnet.html
The WordNet interface of nltk provides a function to do that, of2ss : http://www.nltk.org/_modules/nltk/corpus/reader/wordnet.html
Most of the time it works well, but sometimes Freeling comes with synsets that seem to not be found by of2ss, like '80000779-n' for lemma 'array', and this makes the program crash :
Traceback (most recent call last):
File "/usr/lib/python3.6/site-packages/nltk/corpus/reader/wordnet.py", line 1363, in synset_from_pos_and_line
columns_str, gloss = data_file_line.split('|')
ValueError: not enough values to unpack (expected 2, got 1)
I installed WordNet-3.0 to check array senses, and no one corresponds to the one returned by Freeling. However, both nltk WordNet Interface and Freeling should use the WordNet 3.0 database according to their documentation.
Any idea about this issue?