Яндекс Метрика
Языковая модель

Multipop Adaptive Continuous Stack (PTB)

DeepMind,University of Oxford
Языковое моделирование

Multipop Adaptive Continuous Stack — продвинутая языковая модель с уникальной архитектурой стековой памяти. Этот ИИ от DeepMind показывает отличные результаты на датасете PTB, эффективно справляясь со сложными долгосрочными зависимостями в текстах.

We compare and analyze sequential, random access, and stack memory architectures for recurrent neural network language models. Our experiments on the Penn Treebank and Wikitext-2 datasets show that stack-based memory architectures consistently achieve the best performance in terms of held out perplexity. We also propose a generalization to existing continuous stack models (Joulin & Mikolov,2015; Grefenstette et al., 2015) to allow a variable number of pop operations more naturally that further improves performance. We further evaluate these language models in terms of their ability to capture non-local syntactic dependencies on a subject-verb agreement dataset (Linzen et al., 2016) and establish new state of the art results using memory augmented language models. Our results demonstrate the value of stack-structured memory for explaining the distribution of words in natural language, in line with linguistic theories claiming a context-free backbone for natural language.

Что такое Multipop Adaptive Continuous Stack (PTB)?+
Кто разработал Multipop Adaptive Continuous Stack (PTB)?+
Какие задачи решает Multipop Adaptive Continuous Stack (PTB)?+