Skip to content

A C++ Lightweight Neural Machine Translation Toolkit

License

Notifications You must be signed in to change notification settings

duyvuleo/Mantidae

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

24 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Mantidae - A C++ Lightweight Neural Machine Translation Toolkit

Mantidae is a successor of Mantis (https://github.com/trevorcohn/mantis), but with more functionalities. With Mantidae, you can build a complete neural machine translation system with ease. Mantidae works fast and its performance is competitive with lamtram (https://github.com/neubig/lamtram) or nematus (https://github.com/rsennrich/nematus).

Dependencies

Before compiling Mantidae, you need:

Building

First, clone the repository

git clone https://github.com/duyvuleo/Mantidae.git

As mentioned above, you'll need the latest version of eigen (3.3.x or higher)

hg clone https://bitbucket.org/eigen/eigen/

A modified version of dynet (https://github.com/clab/dynet/tree/master/dynet) is already included (e.g., dynet folder).

CPU build

Compiling to execute on a CPU is as follows

mkdir build_cpu
cd build_cpu
cmake .. -DEIGEN3_INCLUDE_DIR=eigen [-DBoost_NO_BOOST_CMAKE=ON]
make -j 2

Boost note. The "-DBoost_NO_BOOST_CMAKE=ON" can be optional but if you have a trouble of boost-related build error(s), adding it will help to overcome.

MKL support. If you have Intel's MKL library installed on your machine, you can speed up the computation on the CPU by:

cmake .. -DEIGEN3_INCLUDE_DIR=EIGEN [-DBoost_NO_BOOST_CMAKE=ON] -DMKL=TRUE -DMKL_ROOT=MKL

substituting in different paths to EIGEN and MKL if you have placed them in different directories.

This will build the 3 binaries

build_cpu/src/attentional
build_cpu/src/biattentional
build_cpu/src/relopt-decoder

GPU build

Building on the GPU uses the Nvidia CUDA library, currently tested against version 7.5. The process is as follows

mkdir build_gpu
cd build_gpu
cmake .. -DBACKEND=cuda -DEIGEN3_INCLUDE_DIR=EIGEN -DCUDA_TOOLKIT_ROOT_DIR=CUDA [-DBoost_NO_BOOST_CMAKE=ON]
make -j 2

substituting in your Eigen and CUDA folders, as appropriate.

This will result in some binaries, including:

build_gpu/src/attentional
build_gpu/src/biattentional
build_gpu/src/relopt-decoder

Using the model

(to be updated)

Contacts

Hoang Cong Duy Vu ([email protected] or [email protected]), Trevor Cohn and Reza Haffari


Updated Mar 2017

About

A C++ Lightweight Neural Machine Translation Toolkit

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published