From post

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

Rethinking embedding coupling in pre-trained language models., , , , и . CoRR, (2020)Demystifying the Better Performance of Position Encoding Variants for Transformer., , , , , и . CoRR, (2021)Charformer: Fast Character Transformers via Gradient-based Subword Tokenization., , , , , , , , , и . ICLR, OpenReview.net, (2022)Entropy Generation of Desalination Powered by Variable Temperature Waste Heat., , , , и . Entropy, 17 (11): 7530-7566 (2015)Large Language Models Encode Clinical Knowledge., , , , , , , , , и 20 other автор(ы). CoRR, (2022)Language Models are Multilingual Chain-of-Thought Reasoners., , , , , , , , , и 2 other автор(ы). CoRR, (2022)Do Transformer Modifications Transfer Across Implementations and Applications?, , , , , , , , , и 6 other автор(ы). CoRR, (2021)UL2: Unifying Language Learning Paradigms., , , , , , , , , и 3 other автор(ы). ICLR, OpenReview.net, (2023)Scale Efficiently: Insights from Pretraining and Finetuning Transformers., , , , , , , , , и . ICLR, OpenReview.net, (2022)Transcending Scaling Laws with 0.1% Extra Compute., , , , , , , , , и 6 other автор(ы). EMNLP, стр. 1471-1486. Association for Computational Linguistics, (2023)