Homepage: https://sc25.supercomputing.orgarrow-up-right
Paper list: https://sc25.conference-program.comarrow-up-right
LLM Inference
Hetis: Serving LLMs in Heterogeneous GPU Clusters with Fine-grained and Dynamic Parallelism [Paperarrow-up-right] [arXivarrow-up-right]
University of Macau & SYSU
LLM: Large Language Model
Last updated 27 days ago