The DKET project has the goal to devise a Neural Networks based Ontology Learning system that
doesn't rely on hand-crafted rules and it is trained in an end-to-end fashion. The approach, the datasets, and the experimental evaluation are described in this paper.
The best way to install and use the DKET package is to clone the git repository, and set up the proper virtual environment. Once cloned the repository, just move into the directory and run the proper script to create and setup the proper Python 3 virtual environment.
:~$ git clone git@github.com:dkmfbk/dket.git
:~$ cd dket
:~$ ./bin/dket-venv-setup gpu
or just run ./bin/dket-venv-setup if you don't have a GPU card on your
machine. This will create a .py3venv directory and install all the dependencies
you need. Otherwise, You can install DKET as a regular Pyhton package via
pip. Since DKET uses the LiTeFlow
library, you must resolve the link dependency during the installation via the
--process-dependency-links directive:
:~$ pip install --process-dependency-links https://github.com/dkmfbk/dket.git
But after installed DKET you need to install the proper TensorFlow version,
for GPU or CPU, by yourself.
All the experimental settings are stored as .json files in the experiments
folder. To run them, after activating the .py3venv virtual environment, just run:
:~$ ./bin/dket-experiment-run --config experiments/<EXP>.json
where <EXP> is the name of the experimental setting that you want to use. The full options for the ./bin/dket-experiment-run are available at:
:~$ ./bin/dket-experiment-run --help
The folder resources contains:
- the manually curated dataset: 500 sentence-formula pairs, 75 used for training (
reference_set.train.tsv) and 425 used for testing (reference_set.test.tsv); - the grammar (
grammar.cfg) used for generating the synthetic datasets.