Author of the publication

LSTMs Can Learn Syntax-Sensitive Dependencies Well, But Modeling Structure Makes Them Better.

, , , , , and . ACL (1), page 1426-1436. Association for Computational Linguistics, (2018)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

Hybrid computing using a neural network with dynamic external memory, , , , , , , , , and 10 other author(s). Nature, 538 (7626): 471--476 (October 2016)Counterfactual Data Augmentation for Neural Machine Translation., , and . NAACL-HLT, page 187-197. Association for Computational Linguistics, (2021)Learning with Stochastic Guidance for Navigation., , , , , , , and . CoRR, (2018)LSTMs Can Learn Syntax-Sensitive Dependencies Well, But Modeling Structure Makes Them Better., , , , , and . ACL (1), page 1426-1436. Association for Computational Linguistics, (2018)Relational Memory-Augmented Language Models., , and . Trans. Assoc. Comput. Linguistics, (2022)Understanding In-Context Learning in Transformers and LLMs by Learning to Learn Discrete Functions., , , and . ICLR, OpenReview.net, (2024)Human Feedback is not Gold Standard., , and . ICLR, OpenReview.net, (2024)Mutual Information Constraints for Monte-Carlo Objectives to Prevent Posterior Collapse Especially in Language Modelling., , and . J. Mach. Learn. Res., (2022)Hidden Markov Models. (2004)Teaching Machines to Read and Comprehend, , , , , , and . (2015)cite arxiv:1506.03340Comment: Appears in: Advances in Neural Information Processing Systems 28 (NIPS 2015). 14 pages, 13 figures.