DeepQA is a library for doing high-level NLP tasks with deep learning, particularly focused on various kinds of question answering. DeepQA is built on top of Keras and TensorFlow, and can be thought of as a better interface to these systems that makes NLP easier.
Specifically, this library provides the following benefits over plain Keras / tensorflow:
- It is hard to get NLP right in Keras. There are a lot of issues around padding sequences and masking that are not handled well in the main Keras code, and we have well-tested code that does the right thing for, e.g., computing attentions over padded sequences, padding all training instances to the same lengths (possibly dynamically by batch, to minimize computation wasted on padding tokens), or distributing text encoders across several sentences or words.
- We provide a nice, consistent API around building NLP models in Keras. This API has functionality around processing data instances, embedding words and/or characters, easily getting various kinds of sentence encoders, and so on. It makes building models for high-level NLP tasks easy.
- We provide a nice interface to training, validating, and debugging Keras models. It is very easy to experiment with variants of a model family, just by changing some parameters in a JSON file. For example, the particulars of how words are represented, either with fixed GloVe vectors, fine-tuned word2vec vectors, or a concatenation of those with a character-level CNN, are all specified by parameters in a JSON file, not in your actual code. This makes it trivial to switch the details of your model based on the data that you're working with.
- We have implemented a number of state-of-the-art models, particularly focused around question answering systems (though we've dabbled in models for other tasks, as well). The actual model code for these systems is typically 50 lines or less.
To train or evaluate a model using DeepQA, the recommended entry point is to use the
run_model.py
script. That script takes one argument, which is a
parameter file. You can see example parameter files in the examples
directory. You can get some notion of what parameters are available by
looking through the documentation.
Actually training a model will require input files, which you need to provide. We have a companion library, DeepQA Experiments, which was originally designed to produce input files and run experiments, and can be used to generate required data files for most of the tasks we have models for. We're moving towards putting the data processing code directly into DeepQA, so that DeepQA Experiments is not necessary, but for now, getting training data files in the right format is most easily done with DeepQA Experiments.
To implement a new model in DeepQA, you need to subclass TextTrainer
. There is
documentation on what is
necessary for this; see in particular the Abstracts
methods
section. For a simple example of a fully functional model, see the simple sequence
tagger, which has about 20 lines of actual
implementation code.
One snag is that if you're doing a new task, or a new variant of a task with a different
input/output specification, you probably also need to implement an
Instance
type. The Instance
handles reading data from
a file and converting it into numpy arrays that can be used for training and evaluation. This
only needs to happen once for each input/output spec.
DeepQA is organised into the following main sections:
common
: Code for parameter parsing, logging and runtime checks.contrib
: Related code for experiments and untested layers, models and features. Generally untested.data
: Indexing, padding, tokenisation, stemming, embedding and general dataset manipulation happens here.layers
: The bulk of the library. Use these Layers to compose new models. Some of these Layers are very similar to what you might find in Keras, but altered slightly to support arbitrary dimensions or correct masking.models
: Frameworks for different types of task. These generally all extend the TextTrainer class which provides training capabilities to a DeepQaModel. We have models for Sequence Tagging, Entailment, Multiple Choice QA, Reading Comprehension and more. Take a look at the READMEs undermodel
for more details - each task typically has a README describing the task definition.tensors
: Convenience functions for writing the internals of Layers. Will almost exclusively be used inside Layer implementations.training
: This module does the heavy lifting for training and optimisation. We also wrap the Keras Model class to give it some useful debugging functionality.
The data
and models
sections are, in turn, structured according to what task they are intended
for (e.g., text classification, reading comprehension, sequence tagging, etc.). This should make
it easy to see if something you are trying to do is already implemented in DeepQA or not.
DeepQA has implementations of state-of-the-art methods for a variety of tasks. Here are a few of them:
- The attentive reader, from Teaching Machines to Read and Comprehend, by Hermann and others
- Gated Attention Reader from Gated Attention Readers for Text Comprehension,
- Bidirectional Attention Flow, from Bidirectional Attention Flow for Machine Comprehension,
- Decomposable Attention, from A Decomposable Attention Model for Natural Language Inference,
- The original MemNN, from Memory Networks, by Weston, Chopra and Bordes
- End-to-end memory networks, by Sukhbaatar and others
- Dynamic memory networks, by Kumar and others
- DMN+, from Dynamic Memory Networks for Visual and Textual Question Answering, by Xiong, Merity and Socher
This code allows for easy experimentation with the following datasets:
- AI2 Elementary school science questions (no diagrams)
- The Facebook Children's Book Test dataset
- The Facebook bAbI dataset
- The NewsQA dataset
- The Stanford Question Answering Dataset (SQuAD)
- The Who Did What dataset
Note that the data processing code for most of this currently lives in DeepQA Experiments, however.
If you use this code and think something could be improved, pull requests are very welcome. Opening an issue is ok, too, but we're a lot more likely to respond to a PR. The primary maintainer of this code is Matt Gardner, with a lot of help from Pradeep Dasigi (who was the initial author of this codebase), Mark Neumann and Nelson Liu.
This code is released under the terms of the Apache 2 license.