reading list

[摘自网络]

# Frameworks

  • [VLDB '20] PyTorch Distributed: Experiences on Accelerating Data Parallel Training
  • [NeurIPS '19] PyTorch: An Imperative Style, High-Performance Deep Learning Library
  • [OSDI '18] Ray: A Distributed Framework for Emerging AI Applications
  • [OSDI '16] TensorFlow: A System for Large-Scale Machine Learning

# Parallelism & Distributed Systems

  • [OSDI '22] Unity: Accelerating DNN Training Through Joint Optimization of Algebraic Transformations and Parallelization
  • [EuroSys '22] Varuna: Scalable, Low-cost Training of Massive Deep Learning Models
  • [SC ‘21’] Chimera: Efficiently Training Large-Scale Neural Networks with Bidirectional Pipelines
  • [ICML '21] PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models
  • [OSDI '20] A Unified Architecture for Accelerating Distributed DNN Training in Heterogeneous GPU/CPU Clusters
  • [ATC '20] HetPipe: Enabling Large DNN Training on (Whimpy) Heterogeneous GPU Clusters through Integration of Pipelined Model Parallelism and Data Parallelism
  • [NeurIPS '19] GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
  • [SOSP '19] A Generic Communication Scheduler for Distributed DNN Training Acceleration
  • [SOSP '19] PipeDream: Generalized Pipeline Parallelism for DNN Training
  • [EuroSys '19] Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks
  • [arXiv '18] Horovod: fast and easy distributed deep learning in TensorFlow
  • [ATC '17] Poseidon: An Efficient Communication Architecture for Distributed Deep Learning on GPU Clusters
  • [EuroSys '16] STRADS: A Distributed Framework for Scheduled Model Parallel Machine Learning
  • [EuroSys '16] GeePS: Scalable Deep Learning on Distributed GPUs with a GPU-specialized Parameter Server
  • [OSDI '14] Scaling Distributed Machine Learning with the Parameter Server
  • [NIPS '12] Large Scale Distributed Deep Networks

# GPU Cluster Management

  • [OSDI '22] Looking Beyond GPUs for DNN Scheduling on Multi-Tenant Clusters
  • [NSDI '22] MLaaS in the Wild: Workload Analysis and Scheduling in Large-Scale Heterogeneous GPU Clusters
  • [OSDI '21] Pollux: Co-adaptive Cluster Scheduling for Goodput-Optimized Deep Learning
  • [NSDI '21] Elastic Resource Sharing for Distributed Deep Learning
  • [OSDI '20] Heterogeneity-Aware Cluster Scheduling Policies for Deep Learning Workloads
  • [OSDI '20] AntMan: Dynamic Scaling on GPU Clusters for Deep Learning
  • [NSDI '20] Themis: Fair and Efficient GPU Cluster Scheduling
  • [EuroSys '20] Balancing Efficiency and Fairness in Heterogeneous GPU Clusters for Deep Learning
  • [NSDI '19] Tiresias: A GPU Cluster Manager for Distributed Deep Learning
  • [ATC '19] Analysis of Large-Scale Multi-Tenant GPU Clusters for DNN Training Workloads
  • [OSDI '18] Gandiva: Introspective cluster scheduling for deep learning

# Memory Management for Machine Learning

  • [ATC '22] Memory Harvesting in Multi-GPU Systems with Hierarchical Unified Virtual Memory
  • [MobiSys '22] Memory-efficient DNN Training on Mobile Devices
  • [HPCA '22] Enabling Efficient Large-Scale Deep Learning Training with Cache Coherent Disaggregated Memory Systems
  • [ASPLOS '20] Capuchin: Tensor-based GPU Memory Management for Deep Learning
  • [ASPLOS '20] SwapAdvisor: Push Deep Learning Beyond the GPU Memory Limit via Smart Swapping
  • [ISCA '19] Interplay between Hardware Prefetcher and Page Eviction Policy in CPU-GPU Unified Virtual Memory
  • [ISCA '18] Gist: Efficient Data Encoding for Deep Neural Network Training
  • [PPoPP '18] SuperNeurons: Dynamic GPU Memory Management for Training Deep Neural Networks
  • [MICRO '16] vDNN: Virtualized Deep Neural Networks for Scalable, Memory-Efficient Neural Network Design

# Scheduling & Resource Management

  • [arXiv '22] EasyScale: Accuracy-consistent Elastic Training for Deep Learning
  • [MLSys '22] VirtualFlow: Decoupling Deep Learning Models from the Underlying Hardware
  • [SIGCOMM '22] Multi-resource interleaving for deep learning training
  • [EuroSys '22] Out-Of-Order BackProp: An Effective Scheduling Technique for Deep Learning
  • [ATC '21] Zico: Efficient GPU Memory Sharing for Concurrent DNN Training
  • [NeurIPS '20] Nimble: Lightweight and Parallel GPU Task Scheduling for Deep Learning
  • [OSDI’ 20] KungFu: Making Training in Distributed Machine Learning Adaptive
  • [OSDI '20] PipeSwitch: Fast Pipelined Context Switching for Deep Learning Applications
  • [MLSys '20] Salus: Fine-Grained GPU Sharing Primitives for Deep Learning Applications
  • [SOSP '19] Generic Communication Scheduler for Distributed DNN Training Acceleration
  • [EuroSys '18] Optimus: An Efficient Dynamic Resource Scheduler for Deep Learning Clusters
  • [HPCA '18] Applied Machine Learning at Facebook: A Datacenter Infrastructure Perspective

# Serving Systems (& inference acceleration)

  • [EuroSys '23] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access
  • [MICRO '22] DFX: A Low-latency Multi-FPGA Appliance for Accelerating Transformer-based Text Generation
  • [ATC '22] Serving Heterogeneous Machine Learning Models on Multi-GPU Servers with Spatio-Temporal Sharing
  • [OSDI '22] Orca: A Distributed Serving System for Transformer-Based Language Generation Tasks
  • [OSDI '22] Achieving μs-scale Preemption for Concurrent GPU-accelerated DNN Inferences
  • [ATC '21] INFaaS: Automated Model-less Inference Serving
  • [OSDI '20] Serving DNNs like Clockwork: Performance Predictability from the Bottom Up
  • [ISCA '20] MLPerf Inference Benchmark
  • [SOSP '19] Nexus: A GPU Cluster Engine for Accelerating DNN-Based Video Analysis
  • [ISCA '19] MnnFast: a fast and scalable system architecture for memory-augmented neural networks
  • [EuroSys '19] μLayer: Low Latency On-Device Inference Using Cooperative Single-Layer Acceleration and Processor-Friendly Quantization
  • [EuroSys '19] GrandSLAm: Guaranteeing SLAs for Jobs in Microservices Execution Frameworks
  • [OSDI '18] Pretzel: Opening the Black Box of Machine Learning Prediction Serving Systems
  • [NSDI '17] Clipper: A Low-Latency Online Prediction Serving System

# Deep Learning Compiler

  • [PLDI '21] DeepCuts: A Deep Learning Optimization Framework for Versatile GPU Workloads
  • [OSDI '18] TVM: An Automated End-to-End Optimizing Compiler for Deep Learning

# Very Large Models

  • [arxiv '21] ZeRO-Infinity: Breaking the GPU Memory Wall for Extreme Scale Deep Learning
  • [ATC '21] ZeRO-Offload: Democratizing Billion-Scale Model Training
  • [FAST '21] Behemoth: A Flash-centric Training Accelerator for Extreme-scale DNNs

# Deep Learning Recommendation Models

  • [OSDI '22] FAERY: An FPGA-accelerated Embedding-based Retrieval System
  • [OSDI '22] Ekko: A Large-Scale Deep Learning Recommender System with Low-Latency Model Update
  • [EuroSys '22] Fleche: An Efficient GPU Embedding Cache for Personalized Recommendations
  • [ASPLOS '22] RecShard: statistical feature-based memory optimization for industry-scale neural recommendation
  • [HPCA '22] Hercules: Heterogeneity-Aware Inference Serving for At-Scale Personalized Recommendation
  • [MLSys '21] TT-Rec: Tensor Train Compression for Deep Learning Recommendation Model Embeddings
  • [HPCA '21] Tensor Casting: Co-Designing Algorithm-Architecture for Personalized Recommendation Training
  • [HPCA '21] Understanding Training Efficiency of Deep Learning Recommendation Models at Scale
  • [ISCA '20] DeepRecSys: A System for Optimizing End-To-End At-scale Neural Recommendation Inference
  • [HPCA '20] The Architectural Implications of Facebook’s DNN-based Personalized Recommendation
  • [MICRO '19] TensorDIMM: A Practical Near-Memory Processing Architecture for Embeddings and Tensor Operations in Deep Learning

# Hardware Support for ML

  • [ISCA '18] A Configurable Cloud-Scale DNN Processor for Real-Time AI
  • [ISCA '17] In-Datacenter Performance Analysis of a Tensor Processing Unit

# ML at Mobile & Embedded Systems

  • [MobiCom '20] SPINN: Synergistic Progressive Inference of Neural Networks over Device and Cloud
  • [RTSS '19] Pipelined Data-Parallel CPU/GPU Scheduling for Multi-DNN Real-Time Inference
  • [ASPLOS '17] Neurosurgeon: Collaborative Intelligence Between the Cloud and Mobile Edge

# ML Techniques for Systems

  • [ICML '20] An Imitation Learning Approach for Cache Replacement
  • [ICML '18] Learning Memory Access Patterns
作者

lcy

发布于

2023-01-14

更新于

2023-01-18

许可协议