Advances in Neural Information Processing Systems 28 (NIPS 2015)
http://media.nips.cc/Conferences/2015/NIPS-2015-Conference-Book.pdf
https://nips.cc/Conferences/2015/Schedule
https://papers.nips.cc/
https://papers.nips.cc/book/advances-in-neural-information-processing-systems-28-2015
Grammar as a Foreign Language
Oriol Vinyals∗
Google
vinyals@google.com
Lukasz Kaiser∗
Google
lukaszkaiser@google.com
Terry Koo
Google
terrykoo@google.com
Slav Petrov
Google
slav@google.com
Ilya Sutskever
Google
ilyasu@google.com
Geoffrey Hinton
Google
geoffhinton@google.com
Abstract
Syntactic constituency parsing is a fundamental problem in natural language processing
and has been the subject of intensive research and engineering for decades.
As a result, the most accurate parsers are domain specific, complex, and inefficient.
In this paper we show that the domain agnostic attention-enhanced
sequence-to-sequence model achieves state-of-the-art results on the most widely
used syntactic constituency parsing dataset, when trained on a large synthetic corpus
that was annotated using existing parsers. It also matches the performance
of standard parsers when trained only on a small human-annotated dataset, which
shows that this model is highly data-efficient, in contrast to sequence-to-sequence
models without the attention mechanism. Our parser is also fast, processing over
a hundred sentences per second with an unoptimized CPU implementation.
https://papers.nips.cc/paper/5635-grammar-as-a-foreign-language.pdf
https://papers.nips.cc/paper/5857-inferring-algorithmic-patterns-with-stack-augmented-recurrent-nets
Inferring Algorithmic Patterns with
Stack-Augmented Recurrent Nets
Armand Joulin
Facebook AI Research
770 Broadway, New York, USA.
ajoulin@fb.com
Tomas Mikolov
Facebook AI Research
770 Broadway, New York, USA.
tmikolov@fb.com
Abstract
Despite the recent achievements in machine learning, we are still very far from
achieving real artificial intelligence. In this paper, we discuss the limitations of
standard deep learning approaches and show that some of these limitations can be
overcome by learning how to grow the complexity of a model in a structured way.
Specifically, we study the simplest sequence prediction problems that are beyond
the scope of what is learnable with standard recurrent networks, algorithmically
generated sequences which can only be learned by models which have the capacity
to count and to memorize sequences. We show that some basic algorithms can be
learned from sequential data using a recurrent network associated with a trainable
memory.
https://papers.nips.cc/paper/5857-inferring-algorithmic-patterns-with-stack-augmented-recurrent-nets.pdf