From post

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

LLM in a flash: Efficient Large Language Model Inference with Limited Memory., , , , , , , и . ACL (1), стр. 12562-12584. Association for Computational Linguistics, (2024)ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models, , , , , , , и . (2023)ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models., , , , , , , и . CoRR, (2023)LLM in a flash: Efficient Large Language Model Inference with Limited Memory., , , , , , , и . CoRR, (2023)eDKM: An Efficient and Accurate Train-time Weight Clustering for Large Language Models., , , , , , , и . CoRR, (2023)Butterfly Transform: An Efficient FFT Based Neural Architecture Design., , , и . CVPR, стр. 12021-12030. Computer Vision Foundation / IEEE, (2020)DKM: Differentiable k-Means Clustering Layer for Neural Network Compression., , , и . ICLR, OpenReview.net, (2022)ReLU Strikes Back: Exploiting Activation Sparsity in Large Language Models., , , , , , , и . ICLR, OpenReview.net, (2024)eDKM: An Efficient and Accurate Train-Time Weight Clustering for Large Language Models., , , , , , , и . IEEE Comput. Archit. Lett., 23 (1): 37-40 (января 2024)