Skip to content

Implementation of Semantic Parsing with BERT and compositional pre-training on GeoQuery

License

Notifications You must be signed in to change notification settings

simhag/Compositional-Pre-Training-for-Semantic-Parsing-with-BERT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

CS 224N: Natural Language Processing with Deep Learning

Compositional Pre-Training for Semantic Parsing with BERT

Arnaud Autef, Simon Hagege

Abstract

Semantic parsing - the conversion of natural language utterances to logical forms - is a typical natural language processing task. Its applications cover a wide variety of topics, such as question answering, machine translation, instruction following or regular expression generation. In our project, we investigate the performance of Transformer-based sequence-to-sequence models for semantic parsing. We compare a simple Transformer Encoder-Decoder model built on the work of (Vaswani et al., 2017) and a Encoder-Decoder model where the Encoder is BERT (Devlin et al., 2018), a Transformer with weights pre-trained to learn a bidirectional language model over a large dataset. Our architectures have been trained on semantic parsing data using data recombination techniques described in (Jia et al., 2016). We illustrated how Transformer-based Encoder Decoder architectures could be applied to semantic parsing. We evidenced the benefits of BERT as the Encoder of such architectures and pointed to some of the difficulties in fine-tuning it on limited data while the Decoder has to be trained from scratch. We have been mentored by Robin Jia from Stanford University's Computer Science Department.

Course

This project was part of the course CS 224N: Natural Language Processing with Deep Learning taught at Stanford University. http://cs224n.stanford.edu

About

Implementation of Semantic Parsing with BERT and compositional pre-training on GeoQuery

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published