PPoPP 2025

Meta Info

Homepage: https://ppopp25.sigplan.org

Paper list: https://ppopp25.sigplan.org/track/PPoPP-2025-Main-Conference-1#event-overview

Acceptance Rate

20.1% (= 38 / 189)

Papers

Large Language Models (LLMs)

  • LLM Training

    • ATTNChecker: Highly-Optimized Fault Tolerant Attention for Large Language Model Training

      • Oregon & Pacific Northwest National Laboratory & William and Mary

    • Mario: Near Zero-cost Activation Checkpointing in Pipeline Parallelism

      • ICT, CAS

    • WeiPipe: Weight Pipeline Parallelism for Communication-Effective Long-Context Large Model Training

      • THU & NUS & CETHIK & Lynxi Technology

  • LLM Inference

    • MARLIN: Mixed-Precision Auto-Regressive Parallel Inference on Large Language Models

      • ISTA & Universidade da Coruña & ETH & IST Austria

Mixture-of-Experts (MoEs)

  • MoE Training

    • Harnessing Inter-GPU Shared Memory for Seamless MoE Communication-Computation Fusion

      • WHU & NVIDIA & UMacau

Graph Neural Networks (GNNs)

  • GNN Training

    • Adaptive Parallel Training for Graph Neural Networks [Code]

      • CUHK

  • GNN Inference

    • Helios: Efficient Distributed Dynamic Graph Sampling for Online GNN Inference

      • ZJU & Alibaba

GPU Sharing

  • SGDRC: Software-Defined Dynamic Resource Control for Concurrent DNN Inference on NVIDIA GPUs

    • HKUST

Sparse Matrix-Matrix Multiplication (SpMM)

  • Acc-SpMM: Accelerating General-purpose Sparse Matrix-Matrix Multiplication with GPU Tensor Cores

    • Computer Network Information Center, CAS & RUC & Hangzhou Dianzi University

  • FlashSparse: Minimizing Computation Redundancy for Fast Sparse Matrix Multiplications on Tensor Cores

    • BUPT

Last updated

Was this helpful?