This is an implementation for the deep attention fusion LSTM memory network presented in the paper "Long Short-Term Memory Networks for Machine Reading".
This code requires Torch7 and nngraph. It is updated to use torch version around May 2016. Minimum preprocessing is needed to obtain a good accuracy, including lower-casing and tokenization.
@article{cheng2016,
author = {Cheng, Jianpeng and Dong, Li and Lapata, Mirella,
title = {Long Short-Term Memory Networks for Machine Reading},
journal = {EMNLP},
year = {2016},
pages = {551--562}
}