All Stories

  1. Flex: Fast, Accurate DNN Inference on Low-Cost Edges Using Heterogeneous Accelerator Execution
  2. PACER: Accelerating Distributed GNN Training Using Communication-Efficient Partition Refinement and Caching
  3. Improving DNN Inference Throughput Using Practical, Per-Input Compute Adaptation
  4. Towards memory-efficient inference in edge video analytics
  5. A scalable distributed spatial index for the internet-of-things