Author of the publication

Flash-LLM: Enabling Low-Cost and Highly-Efficient Large Generative Model Inference With Unstructured Sparsity.

, , , , , , , , and . Proc. VLDB Endow., 17 (2): 211-224 (2023)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed. You can also use the button next to the name to display some publications already assigned to the person.

 

Other publications of authors with the same name

An Efficient End-to-End Deep Learning Training Framework via Fine-Grained Pattern-Based Pruning., , , , , , , , , and 1 other author(s). CoRR, (2020)Randomness in Neural Network Training: Characterizing the Impact of Tooling., , , and . MLSys, mlsys.org, (2022)Enabling Highly Efficient Capsule Networks Processing Through Software-Hardware Co-Design., , , , and . IEEE Trans. Computers, 70 (4): 495-510 (2021)DynamAP: Architectural Support for Dynamic Graph Traversal on the Automata Processor., , , , and . ACM Trans. Archit. Code Optim., 19 (4): 60:1-60:26 (2022)Randomness In Neural Network Training: Characterizing The Impact of Tooling., , , and . CoRR, (2021)An efficient uncertain graph processing framework for heterogeneous architectures., , , , , , , and . PPoPP, page 477-479. ACM, (2021)η-LSTM: Co-Designing Highly-Efficient Large LSTM Training via Exploiting Memory-Saving and Architectural Design Opportunities., , , , , , and . ISCA, page 567-580. IEEE, (2021)Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity., , , , , , , , and . CoRR, (2023)FP6-LLM: Efficiently Serving Large Language Models Through FP6-Centric Algorithm-System Co-Design., , , , , , , , , and 3 other author(s). CoRR, (2024)ClickTrain: efficient and accurate end-to-end deep learning training via fine-grained architecture-preserving pruning., , , , , , , , , and 1 other author(s). ICS, page 266-278. ACM, (2021)