PPoPP 2025
Meta Info
Homepage: https://ppopp25.sigplan.org
Paper list: https://ppopp25.sigplan.org/track/PPoPP-2025-Main-Conference-1#event-overview
Acceptance Rate
20.1% (= 38 / 189)
Papers
Large Language Models (LLMs)
LLM Training
ATTNChecker: Highly-Optimized Fault Tolerant Attention for Large Language Model Training
Oregon & Pacific Northwest National Laboratory & William and Mary
Mario: Near Zero-cost Activation Checkpointing in Pipeline Parallelism
ICT, CAS
WeiPipe: Weight Pipeline Parallelism for Communication-Effective Long-Context Large Model Training
THU & NUS & CETHIK & Lynxi Technology
LLM Inference
MARLIN: Mixed-Precision Auto-Regressive Parallel Inference on Large Language Models
ISTA & Universidade da Coruña & ETH & IST Austria
Mixture-of-Experts (MoEs)
MoE Training
Harnessing Inter-GPU Shared Memory for Seamless MoE Communication-Computation Fusion
WHU & NVIDIA & UMacau
Graph Neural Networks (GNNs)
GNN Training
Adaptive Parallel Training for Graph Neural Networks [Code]
CUHK
GNN Inference
Helios: Efficient Distributed Dynamic Graph Sampling for Online GNN Inference
ZJU & Alibaba
GPU Sharing
SGDRC: Software-Defined Dynamic Resource Control for Concurrent DNN Inference on NVIDIA GPUs
HKUST
Sparse Matrix-Matrix Multiplication (SpMM)
Acc-SpMM: Accelerating General-purpose Sparse Matrix-Matrix Multiplication with GPU Tensor Cores
Computer Network Information Center, CAS & RUC & Hangzhou Dianzi University
FlashSparse: Minimizing Computation Redundancy for Fast Sparse Matrix Multiplications on Tensor Cores
BUPT
Last updated
Was this helpful?