| 2024 | A Parallel Hash Table for Streaming Applications. Magnus Östgren, Ioannis Sourdis |
| 2024 | A Transducers-based Programming Framework for Efficient Data Transformation. Tri Nguyen, Michela Becchi |
| 2024 | ACE: Efficient GPU Kernel Concurrency for Input-Dependent Irregular Computational Graphs. Sankeerth Durvasula, Adrian Zhao, Raymond Kiguru, Yushi Guan, Zhonghan Chen, Nandita Vijaykumar |
| 2024 | Activation Sequence Caching: High-Throughput and Memory-Efficient Generative Inference with a Single GPU. Sowoong Kim, Eunyeong Sim, Youngsam Shin, Yeongon Cho, Woongki Baek |
| 2024 | BOOM: Use your Desktop to Accurately Predict the Performance of Large Deep Neural Networks. Qidong Su, Jiacheng Yang, Gennady Pekhimenko |
| 2024 | BoostCom: Towards Efficient Universal Fully Homomorphic Encryption by Boosting the Word-wise Comparisons. Ardhi Wiratama Baskara Yudha, Jiaqi Xue, Qian Lou, Huiyang Zhou, Yan Solihin |
| 2024 | Chimera: Leveraging Hybrid Offsets for Efficient Data Prefetching. Shuiyi He, Zicong Wang, Xuan Tang, Qiyao Sun, Dezun Dong |
| 2024 | Faster and More Reliable Quantum SWAPs via Native Gates. Pranav Gokhale, Teague Tomesh, Martin Suchara, Fred Chong |
| 2024 | FriendlyFoe: Adversarial Machine Learning as a Practical Architectural Defense against Side Channel Attacks. Hyoungwook Nam, Raghavendra Pradyumna Pothukuchi, Bo Li, Nam Sung Kim, Josep Torrellas |
| 2024 | GraNNDis: Fast Distributed Graph Neural Network Training Framework for Multi-Server Clusters. Jaeyong Song, Hongsun Jang, Hunseong Lim, Jaewon Jung, Youngsok Kim, Jinho Lee |
| 2024 | Improving Throughput-oriented LLM Inference with CPU Computations. Daon Park, Bernhard Egger |
| 2024 | Leveraging Difference Recurrence Relations for High-Performance GPU Genome Alignment. Alberto Zeni, Seth Onken, Marco Domenico Santambrogio, Mehrzad Samadi |
| 2024 | MIREncoder: Multi-modal IR-based Pretrained Embeddings for Performance Optimizations. Akash Dutta, Ali Jannesari |
| 2024 | MORSE: Memory Overwrite Time Guided Soft Writes to Improve ReRAM Energy and Endurance. Devesh Singh, Donald Yeung |
| 2024 | Mozart: Taming Taxes and Composing Accelerators with Shared-Memory. Vignesh Suresh, Bakshree Mishra, Ying Jing, Zeran Zhu, Naiyin Jin, Charles Block, Paolo Mantovani, Davide Giri, Joseph Zuckerman, Luca P. Carloni, Sarita V. Adve |
| 2024 | NavCim: Comprehensive Design Space Exploration for Analog Computing-in-Memory Architectures. Juseong Park, Boseok Kim, Hyojin Sung |
| 2024 | Optimizing Tensor Computation Graphs with Equality Saturation and Monte Carlo Tree Search. Jakob Hartmann, Guoliang He, Eiko Yoneki |
| 2024 | PIM-Opt: Demystifying Distributed Optimization Algorithms on a Real-World Processing-In-Memory System. Steve Rhyner, Haocong Luo, Juan Gómez-Luna, Mohammad Sadrosadati, Jiawei Jiang, Ataberk Olgun, Harshita Gupta, Ce Zhang, Onur Mutlu |
| 2024 | Parallel Loop Locality Analysis for Symbolic Thread Counts. Fangzhou Liu, Yifan Zhu, Shaotong Sun, Chen Ding, Wesley Smith, Kaave Seyed Hosseini |
| 2024 | PipeGen: Automated Transformation of a Single-Core Pipeline into a Multicore Pipeline for a Given Memory Consistency Model. An Qi Zhang, Andrés Goens, Nicolai Oswald, Tobias Grosser, Daniel J. Sorin, Vijay Nagarajan |
| 2024 | Proceedings of the 2024 International Conference on Parallel Architectures and Compilation Techniques, PACT 2024, Long Beach, CA, USA, October 14-16, 2024 |
| 2024 | Recompiling QAOA Circuits on Various Rotational Directions. Enhyeok Jang, Dongho Ha, Seungwoo Choi, Youngmin Kim, Jaewon Kwon, Yongju Lee, Sungwoo Ahn, Hyungseok Kim, Won Woo Ro |
| 2024 | Rethinking Page Table Structure for Fast Address Translation in GPUs: A Fixed-Size Hashed Page Table. Sungbin Jang, Junhyeok Park, Osang Kwon, Yongho Lee, Seokin Hong |
| 2024 | SZKP: A Scalable Accelerator Architecture for Zero-Knowledge Proofs. Alhad Daftardar, Brandon Reagen, Siddharth Garg |
| 2024 | Toast: A Heterogeneous Memory Management System. Maurice Bailleu, Dimitrios Stavrakakis, Rodrigo Caetano Rocha, Soham Chakraborty, Deepak Garg, Pramod Bhatotia |
| 2024 | Trimma: Trimming Metadata Storage and Latency for Hybrid Memory Systems. Yiwei Li, Boyu Tian, Mingyu Gao |
| 2024 | ZeD: A Generalized Accelerator for Variably Sparse Matrix Computations in ML. Pranav Dangi, Zhenyu Bai, Rohan Juneja, Dhananjaya Wijerathne, Tulika Mitra |
| 2024 | vSPACE: Supporting Parallel Network Packet Processing in Virtualized Environments through Dynamic Core Management. Gyeongseo Park, Minho Kim, Ki-Dong Kang, Yunhyeong Jeon, Sungju Kim, Hyosang Kim, Daehoon Kim |