Potara is a multi-document summarization system that relies on Integer Linear Programming (ILP) and sentence fusion.
Its goal is to summarize a set of related documents in a few sentences. It proceeds by fusing similar sentences in order to create sentences that are either shorter or more informative than those found in the documents. It then uses ILP in order to choose the best set of sentences, fused or not, that will compose the resulting summary.
It relies on state-of-the-art (as of 2014) approaches introduced by Gillick and Favre for the ILP strategy, and Filippova for the sentence fusion.
It is compatible and tested with Python 3.5 and 3.6.
You should be able to install potara and its dependencies with pip
pip install potara
You can also clone this repo and use the requirements.txt file to install dependencies
You will also need GLPK, which is used to obtain an optimal summary (example for Debian-based distro)
$ sudo apt-get install glpk
For Ubuntu-based distros you can use:
$ sudo apt-get install libglpk40
You can check that the install run successfully by cloning the repo and running
$ python setup.py test
If you have issues with install, you can check the .travis.yml file of the repo, which corresponds to a working build.
Basically, you can use the following
from potara.summarizer import Summarizer
from potara.document import Document
s = Summarizer()
# Adding docs, preprocessing them and computing some infos for the summarizer
s.setDocuments([Document('data/' + str(i) + '.txt')
for i in range(1,10)])
# Summarizing, where the actual work is done
s.summarize()
# You can then print the summary
print(s.summary)
There's some preprocessing involved and a sentence fusion step, but I made it easily tunable. Preprocessing may take a while (a few minutes) since there is a lot going on under the hood. Default parameters are currently set for summarizing ~10 documents. You can summarize a smaller amount of documents by tweaking the "minbigramcount" parameter of the summarizer :
s = Summarizer(minbigramcount=2)
Summarizing less than 4 documents would probably yield a bad summary.
Potara relies on similarity scores between sentences. These scores can be shallow using a cosine similarity, or "deep" using gensim Word2Vec semantic representation of words. For the second use case, you'll want to train your own model or use pretrained models. You may contact me if you want to use potara that way, and I may create a tutorial on the matter for the occasion.