Showing 1–20 of 23 results
/ Date/ Name
Aug 21, 2024Distributed-Memory Parallel Algorithms for Sparse Matrix and Sparse Tall-and-Skinny Matrix MultiplicationFeb 9, 2026Towards CXL Resilience to CPU FailuresMay 13, 2022UniHeap: Managing Persistent Objects Across Managed Runtimes for Non-Volatile MemoryJan 5, 2025TAPAS: Thermal- and Power-Aware Scheduling for LLM Inference in Cloud PlatformsMay 25, 2025DECA: A Near-Core LLM Decompression Accelerator Grounded on a 3D Roofline ModelMar 29, 2024Towards Greener LLMs: Bringing Energy-Efficiency to the Forefront of LLM InferenceNov 22, 2019SparseTrain:Leveraging Dynamic Sparsity in Training DNNs on General-Purpose SIMD ProcessorsSep 21, 201621st Century Computer ArchitectureMay 21, 2024Last-Level Cache Side-Channel Attacks Are Feasible in the Modern Public Cloud (Extended Version)May 31, 2025COGNATE: Acceleration of Sparse Tensor Programs on Emerging Hardware using Transfer LearningFeb 9, 2026HEAL: Online Incremental Recovery for Leaderless Distributed Systems Across Persistency ModelsJul 22, 2019Maya: Falsifying Power Sidechannels with Dynamic ControlDec 23, 2018Engineer the Channel and Adapt to it: Enabling Wireless Intra-Chip CommunicationJun 16, 2018Medium Access Control in Wireless Network-on-Chip: A Context AnalysisFeb 3, 2023Defensive ML: Defending Architectural Side-channels with Adversarial ObfuscationJun 27, 2023SENSEi: Input-Sensitive Compilation for Accelerating GNNsNov 28, 2025Serving Heterogeneous LoRA Adapters in Distributed LLM Inference SystemsDec 20, 2021A Method for Hiding the Increased Non-Volatile Cache Read LatencyAug 14, 2018Cache Telepathy: Leveraging Shared Resource Attacks to Learn DNN ArchitecturesAug 1, 2024DynamoLLM: Designing LLM Inference Clusters for Performance and Energy Efficiency