Author of the publication

η-LSTM: Co-Designing Highly-Efficient Large LSTM Training via Exploiting Memory-Saving and Architectural Design Opportunities.

, , , , , , and . ISCA, page 567-580. IEEE, (2021)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

Enabling Highly Efficient Capsule Networks Processing Through Software-Hardware Co-Design., , , , and . IEEE Trans. Computers, 70 (4): 495-510 (2021)DynamAP: Architectural Support for Dynamic Graph Traversal on the Automata Processor., , , , and . ACM Trans. Archit. Code Optim., 19 (4): 60:1-60:26 (2022)Randomness In Neural Network Training: Characterizing The Impact of Tooling., , , and . CoRR, (2021)An Efficient End-to-End Deep Learning Training Framework via Fine-Grained Pattern-Based Pruning., , , , , , , , , and 1 other author(s). CoRR, (2020)Randomness in Neural Network Training: Characterizing the Impact of Tooling., , , and . MLSys, mlsys.org, (2022)Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity., , , , , , , , and . CoRR, (2023)FP6-LLM: Efficiently Serving Large Language Models Through FP6-Centric Algorithm-System Co-Design., , , , , , , , , and 3 other author(s). CoRR, (2024)ClickTrain: efficient and accurate end-to-end deep learning training via fine-grained architecture-preserving pruning., , , , , , , , , and 1 other author(s). ICS, page 266-278. ACM, (2021)An efficient uncertain graph processing framework for heterogeneous architectures., , , , , , , and . PPoPP, page 477-479. ACM, (2021)η-LSTM: Co-Designing Highly-Efficient Large LSTM Training via Exploiting Memory-Saving and Architectural Design Opportunities., , , , , , and . ISCA, page 567-580. IEEE, (2021)