CGO 2026
Sat 31 January - Wed 4 February 2026 Sydney, Australia
co-located with HPCA/CGO/PPoPP/CC 2026
Dates
Tracks
Plenary
You're viewing the program in a time zone which is different from your device's time zone change time zone

Sun 1 Feb

Displayed time zone: Hobart change

18:00 - 20:00
18:00
2h
Poster
Tensor Abstraction Enabling Explicit Layout Optimization in Homomorphic Encryption
Student Research Competition
Seongho Kim Yonsei University, Hanjun Kim Yonsei University
18:00
2h
Poster
UniCon: Unified Controllers for the Quantum Computers
Student Research Competition
Ercüment Kaya Technical University of München and Leibniz Supercomputing Centre, Hossam Ahmed Technical University of München and Leibniz Supercomputing Centre, Martin Schulz Technical University of Munich
18:00
2h
Poster
MDH-DSL: Reduction-Aware Data Parallelism via Multi-Dimensional Homomorphisms
Student Research Competition
Richard Schulze University of Muenster, Sergei Gorlatch University of Muenster
18:00
2h
Poster
Effective Tiling for the Snitch Cluster
Student Research Competition
Emily Sillars University of Murcia, Spain, Alexandra Jimborean University of Murcia
18:00
2h
Poster
Automated Adversarial Test Generation for Debugging Neural Compiler Optimizations
Student Research Competition
Vasu Jindal Columbia University
18:00
2h
Poster
Unlocking Vectorization Scope: Extensible Vectorization via Unified Dependence Semantics
Student Research Competition
Shihan Fang Shanghai Jiao Tong University, Wenxin Zheng Shanghai Jiao Tong University
18:00
2h
Poster
Unifying Medium Sparse Processing Frameworks
Student Research Competition
Meisam Tarabkhah University of Edinburgh, Amir Shaikhha University of Edinburgh
18:00
2h
Poster
Bridging Linalg Dialect with Gemmini Backend
Student Research Competition
Jaemin Kim Yonsei University, Hanjun Kim Yonsei University
18:00
2h
Poster
Leveraging Alias Analysis Without Porting
Student Research Competition
Ravikiran Ravindranath Reddy University of Murcia, Alberto Ros University of Murcia, Alexandra Jimborean University of Murcia

Mon 2 Feb

Displayed time zone: Hobart change

09:50 - 11:10
Compiling for ML 1Main Conference at Bronte
Chair(s): Albert Cohen Google DeepMind
09:50
20m
Talk
Enabling Spill-Free Compilation via Affine-Based Live Range Reduction Optimization
Main Conference
Pre-print
10:10
20m
Talk
GRANII: Selection and Ordering of Primitives in GRAph Neural Networks using Input Inspection
Main Conference
Damitha Lenadora University of Illinois at Urbana-Champaign, Vimarsh Sathia University of Illinois Urbana Champaign, Gerasimos Gerogiannis University of Illinois at Urbana-Champaign, Serif Yesil NVIDIA, Josep Torrellas University of Illinois at Urbana-Champaign, Charith Mendis University of Illinois at Urbana-Champaign
Pre-print
10:30
20m
Talk
Fast Autoscheduling for Sparse ML Frameworks
Main Conference
Bobby Yan Stanford University, Alexander J Root Stanford University, Trevor Gale Stanford University, David Broman KTH Royal Institute of Technology, Fredrik Kjolstad Stanford University
Pre-print
10:50
20m
Talk
Eliminating Redundancy: Ultra-compact Code Generation for Programmable Dataflow Accelerators
Main Conference
Prasanth Chatarasi IBM Research, Alex Gatea IBM, Bardia Mahjour IBM, Jintao Zhang Unaffiliated, Alberto Mannari IBM, Chris Bowler IBM, Shubham Jain IBM Research, Masoud Ataei Jaliseh IBM, Nicole Khoun IBM, Kamlesh Kumar Unaffiliated, Viji Srinivasan IBM Research, Swagath Venkataramani IBM Research
Pre-print
11:10 - 11:30
11:10
20m
Coffee break
Break
HPCA/CGO/PPoPP/CC Catering

11:30 - 12:50
SecurityMain Conference at Balmoral
Chair(s): Michael Franz University of California, Irvine
11:30
20m
Talk
PriTran: Privacy-Preserving Inference for Transformer-Based Language Models under Fully Homomorphic Encryption
Main Conference
Yuechen Mu UNSW, Guangli Li Institute of Computing Technology, Chinese Academy of Sciences, Shiping Chen Data61 at CSIRO, Australia / UNSW, Australia, Jingling Xue University of New South Wales
Pre-print
11:50
20m
Talk
FHEFusion: Enabling Operator Fusion in FHE Compilers for Depth-Efficient DNN Inference
Main Conference
Tianxiang Sui Ant Group, Jianxin Lai Ant Group, Long Li Ant Group, Peng Yuan Ant Group, Yan Liu Ant Group, Qing Zhu Ant Group, Xiaojing Zhang Ant Group, Linjie Xiao Ant Group, Mingzhe Zhang Ant Group, Jingling Xue University of New South Wales
Pre-print Media Attached
12:10
20m
Talk
Towards Path-Aware Coverage-Guided Fuzzing
Main Conference
Giacomo Priamo Sapienza University of Rome, Daniele Cono D'Elia Sapienza University of Rome, Mathias Payer EPFL, Leonardo Querzoni Sapienza University Rome
Pre-print Media Attached
12:30
20m
Talk
SecSwift, a Compiler-Based Framework for Software Countermeasures in Cybersecurity
Main Conference
François de Ferrière STMICROELECTRONICS, Yves Janin STMICROELECTRONICS, Sirine Mechmech Grenoble INP
Pre-print
11:30 - 12:50
AbstractionsMain Conference at Bronte
Chair(s): Antonino Tumeo Pacific Northwest National Laboratory
11:30
20m
Talk
Partial-Evaluation Templates: Accelerating Partial Evaluation with Pre-compiled Templates
Main Conference
Florian Huemer JKU Linz, Aleksandar Prokopec Oracle Labs, David Leopoldseder Oracle Labs, Raphael Mosaner Oracle Labs, Hanspeter Mössenböck JKU Linz
Pre-print
11:50
20m
Talk
Pyls: Enabling Python Hardware Synthesis with Dynamic Polymorphism via LCRS Encoding
Main Conference
Bolei Tong Wuhan University, Yongyan Fang Wuhan University, Wang Chaorui Wuhan University, Qingan Li Wuhan University, China, Jingling Xue University of New South Wales, YUAN Mengting School of Computer Science, Wuhan University, Wuhan, China
Pre-print
12:10
20m
Talk
SkeleShare: Algorithmic Skeletons and Equality Saturation for Hardware Resource Sharing
Main Conference
Jonathan Van der Cruysse McGill University, Tzung-Han Juang McGill University, Shakiba Bolbolian Khah McGill University, Christophe Dubach McGill University
Pre-print Media Attached
12:30
20m
Talk
Ember: A Compiler for Embedding Operations on Decoupled Access-Execute Architectures
Main Conference
Marco Siracusa Barcelona Supercomputing Center; Universitat Politècnica de Catalunya, Olivia Hsu Stanford University, Víctor Soria-Pardos Barcelona Supercomputing Center, Joshua Randall Arm, Arnaud Grasset Arm, Eric Biscondi Arm, Douglas J. Joseph Arm, Randy Allen Barcelona Supercomputing Center, Fredrik Kjolstad Stanford University, Miquel Moreto Technical Univeristy of Catalonia, Adrià Armejach Sanosa Barcelona Supercomputing Center & Universitat Politècnica de Catalunya
Pre-print Media Attached
12:50 - 14:10
12:50
80m
Lunch
Lunch
HPCA/CGO/PPoPP/CC Catering

14:10 - 15:30
MemoryMain Conference at Balmoral
Chair(s): Christophe Guillon STMicroelectronics
14:10
20m
Talk
Flow-Graph-Aware Tiling and Rescheduling for Memory-Efficient On-Device Inference
Main Conference
Yeonoh Jeong Yonsei University, Taehyeong Park Yonsei University, Yongjun Park Yonsei University
Pre-print
14:30
20m
Talk
VFlatten: Selective Value-Object Flattening using Hybrid Static and Dynamic Analysis
Main Conference
Arjun H. Kumar IIT Mandi, Bhavya Hirani SVNIT, Surat, Hang Shao IBM, Tobi Ajila IBM, Vijay Sundaresan IBM Canada, Daryl Maier IBM Canada, Manas Thakur IIT Bombay
Pre-print Media Attached
14:50
20m
Talk
FRUGAL: Pushing GPU Applications beyond Memory Limits
Main Conference
Lingqi Zhang RIKEN RCCS, Tengfei Wang Google Cloud, Jiajun Huang University of California, Riverside, Chen Zhuang Tokyo Institute of Technology, Riken Center for Computational Science, Ivan Ivanov Institute of Science Tokyo, Peng Chen RIKEN RCCS, Toshio Endo , Mohamed Wahib RIKEN Center for Computational Science
Pre-print
15:10
20m
Talk
Automatic Data Enumeration for Fast Collections
Main Conference
Tommy McMichen Northwestern University, Simone Campanoni Google / Northwestern University
Pre-print Media Attached
14:10 - 15:30
DSLsMain Conference at Bronte
Chair(s): Olivia Hsu Stanford University
14:10
20m
Talk
FORTE: Online DataFrame Query Optimizer
Main Conference
Yoonho Choi POSTECH, Kyoungtae Lee Seoul National University, Minji Kim Ewha Womans University, Hyungsoo Jung Seoul National University, Hyojin Sung Seoul National University
Pre-print
14:30
20m
Talk
LEGO: A Layout Expression Language for Code Generation of Hierarchical Mapping
Main Conference
Amir Mohammad Tavakkoli University of Utah, Cosmin E. Oancea University of Copenhagen, Denmark, Mary Hall University of Utah
Pre-print Media Attached
14:50
20m
Talk
Pushing Tensor Accelerators beyond MatMul in a User-Schedulable Language
Main Conference
Yihong Zhang University of Washington, Derek Gerstmann Adobe, Andrew Adams Adobe Research, Maaz Bin Safeer Ahmad University of Washington, Seattle
Pre-print Media Attached
15:10
20m
Talk
Tawa: Automatic Warp Specialization for Modern GPUs with Asynchronous References
Main Conference
Hongzheng Chen Cornell University, Bin Fan Nvidia, Alexander Collins NVIDIA, Bastian Hagedorn NVIDIA, Evghenii Gaburov NVIDIA, Masahiro Masuda NVIDIA, Matthew Brookhart NVIDIA, Chris Sullivan NVIDIA, Jason Knight NVIDIA, Zhiru Zhang Cornell University, USA, Vinod Grover NVIDIA
Pre-print
15:30 - 15:50
15:30
20m
Coffee break
Break
HPCA/CGO/PPoPP/CC Catering

15:50 - 17:10
Quantum / HLSMain Conference at Balmoral
Chair(s): Aaron Smith
15:50
20m
Talk
Dependence-Driven, Scalable Quantum Circuit Mapping with Affine Abstractions
Main Conference
Marouane Benbetka École Nationale Supérieure d’Informatique, Merwan BEKKAR École Nationale Supérieure d’Informatique, Riyadh Baghdadi New York University Abu Dhabi, Martin Kong Ohio State University
Pre-print Media Attached
16:10
20m
Talk
Space-Time Optimisations for Early Fault-Tolerant Quantum Computation
Main Conference
Sanaa Sharma University of Cambridge, Prakash Murali University of Cambridge
Pre-print Media Attached
16:30
20m
Talk
OpenQudit: Extensible and Accelerated Numerical Quantum Compilation via a JIT-Compiled DSL
Main Conference
Ed Younis Lawrence Berkeley National Laboratory
Pre-print Media Attached
16:50
20m
Talk
Selene: Cross-Level Barrier-Free Pipelining for Irregular Nested Loops in High-Level Synthesis
Main Conference
Sungwoo Yun Yonsei University, Seonyoung Cheon Yonsei University, Dongkwan Kim Yonsei University, Heelim Choi Yonsei University, Kunmo Jeong Yonsei University, Chan Lee Yonsei University, Yongwoo Lee DGIST, Hanjun Kim Yonsei University
Pre-print
15:50 - 17:10
Parallelization / VectorizationMain Conference at Bronte
Chair(s): V Krishna Nandivada IIT Madras
15:50
20m
Talk
Enabling Automatic Compiler-Driven Vectorization of Transformers
Main Conference
Shreya Alladi University of Murcia, Alberto Ros University of Murcia, Alexandra Jimborean University of Murcia
Pre-print Media Attached
16:10
20m
Talk
Unlocking Python Multithreading Capabilities using OpenMP-Based Programming with OMP4Py
Main Conference
César Piñeiro University of Santiago de Compostela, Juan C. Pichel University of Santiago de Compostela
Pre-print Media Attached
16:30
20m
Talk
The Parallel-Semantics Program Dependence Graph for Parallel Optimization
Main Conference
Yian Su Northwestern University, Brian Homerding Northwestern University, Haocheng Gao Northwestern University, Federico Sossai Northwestern University, Yebin Chon Princeton University, David I. August Princeton University, Simone Campanoni Google / Northwestern University
Pre-print Media Attached
16:50
20m
Talk
From Threads to Tiles: T2T, a Compiler for CUDA-to-NPU Translation via 2D Vectorization
Main Conference
Shuaijiang Li Institute of Computing Technology at Chinese Academy of Sciences, Jiacheng Zhao Institute of Computing Technology at Chinese Academy of Sciences; University of Chinese Academy of Sciences; Zhongguancun Laboratory, Ying Liu Institute of Computing Technology, Chinese Academy of Sciences, Shuoming Zhang Institute of Computing Technology at Chinese Academy of Sciences, Lei Chen University of Chinese Academy of Sciences, Yijin Li Institute of Computing Technology at Chinese Academy of Sciences, Yangyu Zhang Institute of Computing Technology,Chinese Academy of Sciences, lizhicheng Institute of Computing Technology at Chinese Academy of Sciences, Runyu Zhou Institute of Computing Technology at Chinese Academy of Sciences, Xiyu Shi Institute of Computing Technology at Chinese Academy of Sciences, Chunwei Xia University of Leeds, Yuan Wen University of Aberdeen, Xiaobing Feng ICT CAS, Huimin Cui Institute of Computing Technology, Chinese Academy of Sciences
Pre-print
17:30 - 19:00
Business MeetingMain Conference at Bronte
Chair(s): Steve Blackburn Google and Australian National University, Albert Cohen Google DeepMind, Timothy M. Jones University of Cambridge
17:30
90m
Meeting
CGO Business Meeting
Main Conference

Tue 3 Feb

Displayed time zone: Hobart change

09:50 - 11:10
Binary / JITMain Conference at Balmoral
Chair(s): Alexandra Jimborean University of Murcia
09:50
20m
Talk
Binary Diffing via Library Signatures
Main Conference
Andrei Rimsa CEFET-MG, Anderson Faustino da Silva State University of Maringá, Camilo Santana Melgaço Federal University of Minas Gerais, Fernando Magno Quintão Pereira Federal University of Minas Gerais
Pre-print Media Attached
10:10
20m
Talk
BIT: Empowering Binary Analysis through the LLVM Toolchain
Main Conference
Puzhuo Liu Ant Group & Tsinghua University, Peng Di Ant Group & UNSW, Jingling Xue University of New South Wales, Yu Jiang Tsinghua University
Pre-print
10:30
20m
Talk
Dr.avx: A Dynamic Compilation System for Seamlessly Executing Hardware-Unsupported Vectorization Instructions
Main Conference
Yue Tang East China Normal University, Mianzhi Wu East China Normal University, Yufeng Li East China Normal University, Haoyu Liao East China Normal University, Jianmei Guo East China Normal University, Bo Huang East China Normal University
Pre-print Media Attached
10:50
20m
Talk
Practical: Are Abstract-Interpreter Baseline JITs Worth It? An Empirical Evaluation through Metacompilation
Main Conference
Nahuel Palumbo Université Lille, CNRS, Centrale Lille, Inria, UMR 9189 - CRIStAL, Guillermo Polito Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 CRIStAL, Stéphane Ducasse Inria; University of Lille; CNRS; Centrale Lille; CRIStAL, Pablo Tesone Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 CRIStAL, Pharo Consortium
Pre-print
09:50 - 11:10
Code GenerationMain Conference at Bronte
Chair(s): Fredrik Kjolstad Stanford University
09:50
20m
Talk
TPDE: A Fast Adaptable Compiler Back-End Framework
Main Conference
Tobias Schwarz TU Munich, Tobias Kamm TU Munich, Alexis Engelke TU Munich
Pre-print Media Attached
10:10
20m
Talk
Synthesizing Instruction Selection Back-Ends from ISA Specifications Made Practical
Main Conference
Florian Drescher Technical University of Munich, Alexis Engelke TU Munich
Pre-print
10:30
20m
Talk
SparseX: Synergizing GPU Libraries for Sparse Matrix Multiplication on Heterogeneous Processors
Main Conference
Ruifeng Zhang North Carolina State University, Xiangwei Wang North Carolina State University, Ang Li Pacific Northwest National Laboratory, Xipeng Shen North Carolina State University
Pre-print Media Attached
10:50
20m
Talk
Compilation of Generalized Matrix Chains with Symbolic Sizes
Main Conference
Francisco López Umeå University, Lars Karlsson Umeå University, Paolo Bientinesi Umeå University
Pre-print Media Attached
11:10 - 11:30
11:10
20m
Coffee break
Break
HPCA/CGO/PPoPP/CC Catering

11:30 - 12:50
Profiling / InstrumentationMain Conference at Bronte
Chair(s): Mircea Trofin Google
11:30
20m
Talk
TRACE4J: A Lightweight, Flexible, and Insightful Performance Tracing Tool for Java
Main Conference
Haide He UC Merced, Pengfei Su University of California, Merced
Pre-print Media Attached
11:50
20m
Talk
Proton: Towards Multi-level, Adaptive Profiling for Triton
Main Conference
Keren Zhou George Mason University, Tianle Zhong University of Virginia, Hao Wu George Mason University, Jihyeong Lee George Mason University, Yue Guan University of California at San Diego, Yufei Ding University of California at Santa Barbara, Corbin Robeck Meta, Yuanwei Fang Meta, Jeff Niu OpenAI, Philippe Tillet OpenAI
Pre-print Media Attached
12:10
20m
Talk
On the Precision of Dynamic Program Fingerprints Based on Performance Counters
Main Conference
Anderson Faustino da Silva State University of Maringá, Sergio Queiroz de Medeiros Universidade Federal do Rio Grande do Norte, Marcelo Borges Nogueira Federal University of Rio Grande do Norte, Jeronimo Castrillon TU Dresden, Germany, Fernando Magno Quintão Pereira Federal University of Minas Gerais
Pre-print Media Attached
12:30
20m
Talk
PASTA: A Modular Program Analysis Tool Framework for Accelerators
Main Conference
Mao Lin University of California Merced, Hyeran Jeon University of California, Merced, Keren Zhou George Mason University
Pre-print Media Attached
12:50 - 14:10
12:50
80m
Awards
HPCA Awards Lunch
HPCA/CGO/PPoPP/CC Catering

12:50 - 14:10
12:50
80m
Lunch
Lunch
HPCA/CGO/PPoPP/CC Catering

14:10 - 15:30
AnalysisMain Conference at Bronte
Chair(s): Jose Nelson Amaral University of Alberta
14:10
20m
Talk
PIP: Making Andersen’s Points-to Analysis Sound and Practical for Incomplete C Programs
Main Conference
Håvard Rognebakke Krogstie NTNU, Helge Bahmann Independent Researcher, Magnus Själander Norwegian University of Science and Technology (NTNU), Nico Reissmann Independent Researcher
Pre-print Media Attached
14:30
20m
Talk
Thinking Fast and Correct: Automated Rewriting of Numerical Code through Compiler Augmentation
Main Conference
Siyuan Brant Qian University of Illinois at Urbana-Champaign, Vimarsh Sathia University of Illinois Urbana Champaign, Ivan Ivanov Institute of Science Tokyo, Jan Hueckelheim Argonne National Laboratory, Paul Hovland Argonne National Laboratory, William S. Moses University of Illinois Urbana-Champaign
Pre-print Media Attached
14:50
20m
Talk
PolyUFC: Polyhedral Compilation Meets Roofline Analysis for Uncore Frequency Capping
Main Conference
Nilesh Rajendra Shah Indian Institute of Technology Hyderabad, India, M V V S Manoj Kumar IIT Hyderabad, Dhairya Baxi IIT Hyderabad, Ramakrishna Upadrasta IIT Hyderabad
Pre-print
15:10
20m
Talk
Accelerating App Recompilation across Android System Updates by Code Reusing
Main Conference
Hongtao Wu Wuhan University, Yu Chen Wuhan University, Mengfei Xie Wuhan University, Futeng Yang Guangdong OPPO Mobile Telecommunications, Jun Yan Guangdong OPPO Mobile Telecommunications, Jiang Ma OPPO Electronics Corp., Jianming Fu Wuhan University, Jason Xue MBZUAI, Qingan Li Wuhan University, China
Pre-print
15:30 - 15:50
15:30
20m
Coffee break
Break
HPCA/CGO/PPoPP/CC Catering

15:50 - 17:10
Compiling for ML 2Main Conference at Bronte
Chair(s): Fabrice Rastello University Grenoble Alpes - Inria - CNRS - Grenoble INP - LIG
15:50
20m
Talk
QIGen: A Kernel Generator for Inference on Nonuniformly Quantized Large Language Models
Main Conference
Tommaso Pegolotti ETH Zürich, Dan Alistarh IST Austria, Markus Püschel ETH Zurich
Pre-print Media Attached
16:10
20m
Talk
DyPARS: Dynamic-Shape DNN Optimization via Pareto-Aware MCTS for Graph Variants
Main Conference
Hao Qian University of New South Wales, Guangli Li Institute of Computing Technology, Chinese Academy of Sciences, Qiuchu Yu Institute of Computing Technology at Chinese Academy of Sciences, Xueying Wang Beijing University of Posts and Telecommunications, Jingling Xue University of New South Wales
Pre-print Media Attached
16:30
20m
Talk
Compiler-Runtime Co-operative Chain of Verification for LLM-Based Code Optimization
Main Conference
Hyunho Kwon Yonsei University, Sanggyu Shin SAIT, Ju Min Lee Yonsei University, Hoyun Youm Yonsei University, Seungbin Song SAIT, Seongho Kim Yonsei University, Hanwoong Jung Samsung Advanced Institute of Technology, Seungwon Lee Samsung Advanced Institute of Technology, Hanjun Kim Yonsei University
Pre-print
16:50
20m
Talk
Hexcute: A Compiler Framework for Automating Layout Synthesis in GPU Programs
Main Conference
Xiao Zhang University of Toronto; NVIDIA, Yaoyao Ding University of Toronto; Vector Institute; NVIDIA, Bolin Sun University of Toronto; NVIDIA, Yang Hu NVIDIA, Tatiana Shpeisman Google, Gennady Pekhimenko University of Toronto / Vector Institute
Pre-print Media Attached
17:15 - 18:15

Wed 4 Feb

Displayed time zone: Hobart change

09:50 - 11:10
Tensor OptimizationMain Conference at Bronte
Chair(s): Bastian Hagedorn NVIDIA
09:50
20m
Talk
Multidirectional Propagation of Sparsity Information across Tensor Slices
Main Conference
Kaio Henrique Andrade Ananias Universidade Federal de Minas Gerais, Danila Seliayeu University of Alberta, Jose Nelson Amaral University of Alberta, Fernando Magno Quintão Pereira Federal University of Minas Gerais
Pre-print Media Attached
10:10
20m
Talk
Synthesizing Specialized Sparse Tensor Accelerators for FPGAs via High-Level Functional Abstractions
Main Conference
Hamza Javed McGill University, Canada, Christophe Dubach McGill University
Pre-print
10:30
20m
Talk
Progressive Low-Precision Approximation of Tensor Operators on GPUs: Enabling Greater Trade-Offs between Performance and Accuracy
Main Conference
Fan Luo Institute of Computing Technology at Chinese Academy of Sciences, Guangli Li Institute of Computing Technology, Chinese Academy of Sciences, Zhaoyang Hao Institute of Computing Technology at Chinese Academy of Sciences, Xueying Wang Beijing University of Posts and Telecommunications, Xiaobing Feng ICT CAS, Huimin Cui Institute of Computing Technology, Chinese Academy of Sciences, Jingling Xue University of New South Wales
Pre-print
10:50
20m
Talk
Tensor Program Superoptimization through Cost-Guided Symbolic Program Synthesis
Main Conference
Alexander Brauckmann University of Edinburgh, Aarsh Chaube University of Edinburgh, José Wesley De Souza Magalhães University of Edinburgh, Elizabeth Polgreen University of Edinburgh, Michael F. P. O'Boyle University of Edinburgh
Pre-print Media Attached
11:10 - 11:30
11:10
20m
Coffee break
Break
HPCA/CGO/PPoPP/CC Catering

11:30 - 12:50
OptimizationMain Conference at Bronte
Chair(s): Teresa Johnson Google
11:30
20m
Talk
A Reinforcement Learning Environment for Automatic Code Optimization in the MLIR Compiler
Main Conference
Mohammed Tirichine New York University Abu Dhabi; Ecole nationale Supérieure d'Informatique, Nassim Ameur NYU Abu Dhabi; École Nationale Supérieure d’Informatique, Nazim Bendib NYU Abu Dhabi; École Nationale Supérieure d’Informatique, Iheb Nassim Aouadj NYU Abu Dhabi, Djad Bouchama NYU Abu Dhabi; University of Science and Technology Houari Boumediene, Rafik Bouloudene NYU Abu Dhabi; University of Science and Technology Houari Boumediene, Riyadh Baghdadi New York University Abu Dhabi
Pre-print Media Attached
11:50
20m
Talk
Towards Threading the Needle of Debuggable Optimized Binaries
Main Conference
Cristian Assaiante Sapienza University of Rome, Simone Di Biasio Sapienza University of Rome, Snehasish Kumar Google LLC, Giuseppe Antonio Di Luna Sapienza University of Rome, Daniele Cono D'Elia Sapienza University of Rome, Leonardo Querzoni Sapienza University Rome
Pre-print Media Attached
12:10
20m
Talk
Compiler-Assisted Instruction Fusion
Main Conference
Ravikiran Ravindranath Reddy University of Murcia, Sawan Singh AMD, Arthur Perais CNRS, Alberto Ros University of Murcia, Alexandra Jimborean University of Murcia
Pre-print
12:30
20m
Talk
LLM-VeriOpt: Verification-Guided Reinforcement Learning for LLM-Based Compiler Optimization
Main Conference
Xiangxin Fang Queen Mary University of London; University of Edinburgh, Jiaqin Kang Queen Mary University of London, Rodrigo C. O. Rocha University of Edinburgh, Sam Ainsworth University of Edinburgh, Lev Mukhanov IMEC (Cambridge); Queen Mary University of London
Pre-print Media Attached
12:50 - 13:20

Accepted Papers

Title
Accelerating App Recompilation across Android System Updates by Code Reusing
Main Conference
Pre-print
A Reinforcement Learning Environment for Automatic Code Optimization in the MLIR Compiler
Main Conference
Pre-print Media Attached
Automatic Data Enumeration for Fast Collections
Main Conference
Pre-print Media Attached
Binary Diffing via Library Signatures
Main Conference
Pre-print Media Attached
BIT: Empowering Binary Analysis through the LLVM Toolchain
Main Conference
Pre-print
Compilation of Generalized Matrix Chains with Symbolic Sizes
Main Conference
Pre-print Media Attached
Compiler-Assisted Instruction Fusion
Main Conference
Pre-print
Compiler-Runtime Co-operative Chain of Verification for LLM-Based Code Optimization
Main Conference
Pre-print
Dependence-Driven, Scalable Quantum Circuit Mapping with Affine Abstractions
Main Conference
Pre-print Media Attached
Dr.avx: A Dynamic Compilation System for Seamlessly Executing Hardware-Unsupported Vectorization Instructions
Main Conference
Pre-print Media Attached
DyPARS: Dynamic-Shape DNN Optimization via Pareto-Aware MCTS for Graph Variants
Main Conference
Pre-print Media Attached
Eliminating Redundancy: Ultra-compact Code Generation for Programmable Dataflow Accelerators
Main Conference
Pre-print
Ember: A Compiler for Embedding Operations on Decoupled Access-Execute Architectures
Main Conference
Pre-print Media Attached
Enabling Automatic Compiler-Driven Vectorization of Transformers
Main Conference
Pre-print Media Attached
Enabling Spill-Free Compilation via Affine-Based Live Range Reduction Optimization
Main Conference
Pre-print
Fast Autoscheduling for Sparse ML Frameworks
Main Conference
Pre-print
FHEFusion: Enabling Operator Fusion in FHE Compilers for Depth-Efficient DNN Inference
Main Conference
Pre-print Media Attached
Flow-Graph-Aware Tiling and Rescheduling for Memory-Efficient On-Device Inference
Main Conference
Pre-print
FORTE: Online DataFrame Query Optimizer
Main Conference
Pre-print
From Threads to Tiles: T2T, a Compiler for CUDA-to-NPU Translation via 2D Vectorization
Main Conference
Pre-print
FRUGAL: Pushing GPU Applications beyond Memory Limits
Main Conference
Pre-print
GRANII: Selection and Ordering of Primitives in GRAph Neural Networks using Input Inspection
Main Conference
Pre-print
Hexcute: A Compiler Framework for Automating Layout Synthesis in GPU Programs
Main Conference
Pre-print Media Attached
LEGO: A Layout Expression Language for Code Generation of Hierarchical Mapping
Main Conference
Pre-print Media Attached
LLM-VeriOpt: Verification-Guided Reinforcement Learning for LLM-Based Compiler Optimization
Main Conference
Pre-print Media Attached
Multidirectional Propagation of Sparsity Information across Tensor Slices
Main Conference
Pre-print Media Attached
On the Precision of Dynamic Program Fingerprints Based on Performance Counters
Main Conference
Pre-print Media Attached
OpenQudit: Extensible and Accelerated Numerical Quantum Compilation via a JIT-Compiled DSL
Main Conference
Pre-print Media Attached
Partial-Evaluation Templates: Accelerating Partial Evaluation with Pre-compiled Templates
Main Conference
Pre-print
PASTA: A Modular Program Analysis Tool Framework for Accelerators
Main Conference
Pre-print Media Attached
PIP: Making Andersen’s Points-to Analysis Sound and Practical for Incomplete C Programs
Main Conference
Pre-print Media Attached
PolyUFC: Polyhedral Compilation Meets Roofline Analysis for Uncore Frequency Capping
Main Conference
Pre-print
Practical: Are Abstract-Interpreter Baseline JITs Worth It? An Empirical Evaluation through Metacompilation
Main Conference
Pre-print
PriTran: Privacy-Preserving Inference for Transformer-Based Language Models under Fully Homomorphic Encryption
Main Conference
Pre-print
Progressive Low-Precision Approximation of Tensor Operators on GPUs: Enabling Greater Trade-Offs between Performance and Accuracy
Main Conference
Pre-print
Proton: Towards Multi-level, Adaptive Profiling for Triton
Main Conference
Pre-print Media Attached
Pushing Tensor Accelerators beyond MatMul in a User-Schedulable Language
Main Conference
Pre-print Media Attached
Pyls: Enabling Python Hardware Synthesis with Dynamic Polymorphism via LCRS Encoding
Main Conference
Pre-print
QIGen: A Kernel Generator for Inference on Nonuniformly Quantized Large Language Models
Main Conference
Pre-print Media Attached
SecSwift, a Compiler-Based Framework for Software Countermeasures in Cybersecurity
Main Conference
Pre-print
Selene: Cross-Level Barrier-Free Pipelining for Irregular Nested Loops in High-Level Synthesis
Main Conference
Pre-print
SkeleShare: Algorithmic Skeletons and Equality Saturation for Hardware Resource Sharing
Main Conference
Pre-print Media Attached
Space-Time Optimisations for Early Fault-Tolerant Quantum Computation
Main Conference
Pre-print Media Attached
SparseX: Synergizing GPU Libraries for Sparse Matrix Multiplication on Heterogeneous Processors
Main Conference
Pre-print Media Attached
Synthesizing Instruction Selection Back-Ends from ISA Specifications Made Practical
Main Conference
Pre-print
Synthesizing Specialized Sparse Tensor Accelerators for FPGAs via High-Level Functional Abstractions
Main Conference
Pre-print
Tawa: Automatic Warp Specialization for Modern GPUs with Asynchronous References
Main Conference
Pre-print
Tensor Program Superoptimization through Cost-Guided Symbolic Program Synthesis
Main Conference
Pre-print Media Attached
The Parallel-Semantics Program Dependence Graph for Parallel Optimization
Main Conference
Pre-print Media Attached
Thinking Fast and Correct: Automated Rewriting of Numerical Code through Compiler Augmentation
Main Conference
Pre-print Media Attached
Towards Path-Aware Coverage-Guided Fuzzing
Main Conference
Pre-print Media Attached
Towards Threading the Needle of Debuggable Optimized Binaries
Main Conference
Pre-print Media Attached
TPDE: A Fast Adaptable Compiler Back-End Framework
Main Conference
Pre-print Media Attached
TRACE4J: A Lightweight, Flexible, and Insightful Performance Tracing Tool for Java
Main Conference
Pre-print Media Attached
Unlocking Python Multithreading Capabilities using OpenMP-Based Programming with OMP4Py
Main Conference
Pre-print Media Attached
VFlatten: Selective Value-Object Flattening using Hybrid Static and Dynamic Analysis
Main Conference
Pre-print Media Attached

Call for Papers

IEEE/ACM International Symposium on Code Generation and Optimization (CGO)

Co-located with HPCA, PPoPP, and CC

Sydney, Australia

The International Symposium on Code Generation and Optimization (CGO’26) will be held in Sydney, Australia. CGO is the premier venue to bring together researchers and practitioners working at the interface of hardware and software on a wide range of optimization and code generation techniques and related issues. The conference spans the spectrum from purely static to fully dynamic approaches, and from pure software-based methods to specific architectural features and support for code generation and optimization.


Important dates

CGO uses two submissions per year.

This follows the model established by other conferences in our field in recent years, such as ASPLOS and OOPSLA. Papers submitted to the first round can either be directly accepted, rejected, or invited to submit a revised version of the paper to the second round. Papers rejected in the first round may not be submitted to the second round. For papers invited to submit a revised version, authors will be given a list of revisions that should be acted on to improve the paper. We will make every effort to ensure that the revised paper is reviewed by the same reviewers (and possibly additional reviewers), who will assess whether the revisions are satisfactory. If so, the paper will be accepted. If the revised paper is rejected, the authors may submit a further revised version in a subsequent round, which will be treated as a new submission.

First Submission Deadline

  • Paper Submission: 29 May 2025
  • Author Rebuttal Period: 8–10 July 2025
  • Paper Notification: 21 July 2025

Second Submission Deadline

  • Paper Submission: 11 September 2025
  • Author Rebuttal Period: 21–23 October 2025
  • Paper Notification: 3 November 2025

Contacts:


Topics

Original contributions are solicited on, but not limited to, the following topics:

  • Code generation, translation, transformation, and optimization for performance, energy, virtualization, portability, security, or reliability concerns, and architectural support
  • Efficient execution of dynamically typed and higher-level languages
  • Optimization and code generation for emerging programming models, platforms, domain-specific languages
  • Dynamic, static, profile-guided and feedback-directed optimization
  • Machine-learning-based code generation, analysis, transformation and optimization
  • Static, dynamic, and hybrid analysis for performance, energy, memory locality, throughput or latency, security, reliability, or functional debugging
  • Program characterization methods
  • Profiling and instrumentation techniques and architectural support for them
  • Novel and efficient tools
  • Compiler design, practice and experience
  • Compiler abstraction and intermediate representations
  • Vertical integration of language features, representations, optimizations, and runtime support for parallelism
  • Solutions that involve cross-layer (HW/OS/VM/SW) design and integration
  • Deployed dynamic/static compiler and runtime systems for general purpose, embedded system and cloud / HPC platforms
  • Parallelism, heterogeneity, and reconfigurable architectures
  • Optimizations for heterogeneous or specialized targets, GPUs, SoCs, CGRA and quantum computers
  • Compiler support for vectorization, thread extraction, task scheduling, speculation, transaction, memory management, data distribution and synchronization

Paper types

Papers submitted to CGO can be standard research papers, tools papers or practical experience papers. All must be written in the IEEE conference format (use the conference mode), and may have up to 10 pages, references excluded. Supplementary materials may be included as an Appendix at the end of the submitted paper. The Appendix has no page limit, but the text of the full paper excluding the Appendix must fit within 10 pages. Reviewers are not required to read the Appendix and may do so at their discretion. In other words, papers must be self-contained without needing to read any material in the Appendix.

Please ensure that you and your co-authors obtain an ORCID ID, so you can complete the publishing process for your accepted paper. We are committed to improve author discoverability, ensure proper attribution and contribute to ongoing community efforts around name normalization; your ORCID ID will help in these efforts.

Please be mindful of the IEEE publishing policies when preparing and submitting your article.

Tool papers

Tool papers must give a clear account of a new tool’s functionality. They must describe the problem that the tool helps to solve, how the tool works and provide an evaluation of the tool. There does not have to be novel research in the way the tool works, but it does have to solve a novel problem or in a novel (and better) way compared to existing tools. A good tools paper will:

  • Frame the problem that the tool solves, showing that it is an important subject
  • Describe how the tool works to solve the problem
  • Evaluate the ability of the tool to solve the problem
  • Possibly provide case studies of using the tool, showing what improvements can be made now the problem has been ameliorated.

Tools paper authors should prepend their paper title with ‘Tool:’ to provide clarity that this is a tools paper during the review process (this can be removed for the final camera-ready paper).

The successful evaluation of an artifact is mandatory for a Tool Paper. Therefore, authors of work conditionally accepted as Tool Papers must submit an artifact to the Artifact Evaluation Committee. The successful evaluation of the artifact is a requirement for final acceptance.

The selection criteria for papers in this category are:

  • Originality: Papers should present CGO-related technologies applied to real-world problems with scope or characteristics that set them apart from previous solutions.
  • Usability: The presented tool or compiler should have broad usage or applicability. They are expected to assist in CGO-related research, or could be extended to investigate or demonstrate new technologies. If significant components are not yet implemented, the paper will not be considered.
  • Documentation: The tool or compiler should be presented on a web-site giving documentation and further information about the tool.
  • Benchmark Repository: A suite of benchmarks for testing should be provided.
  • Availability: The tool or compiler should be available for public use.
  • Foundations: Papers should incorporate the principles underpinning Code Generation and Optimization (CGO). However, a thorough discussion of theoretical foundations is not required; a summary of such should suffice.
  • Artifact Evaluation: The submitted artifact must be functional and support the claims made in the paper.

Practical experience papers

Practical experience papers should summarize a practical experience with realistic case studies. They must make it clear where the novelty in the work comes from, which could be from applying known techniques to a new system (with novel features), from finding new ways to implement existing analyses or transformations in a unique way, or from applying established schemes to draw new conclusions about them.

Practical experience paper authors should prepend their paper title with ‘Practical:’ to provide clarity that this is a practical experience paper during the review process (this can be removed for the final camera-ready paper).

Practical experience papers are encouraged, but not required, to submit an artifact to the Artifact Evaluation process.

The selection criteria for papers in this category are:

  • Originality: Papers should present CGO-related technologies applied to real-world problems with scope or characteristics that set them apart from previous solutions. Alternatively, papers may also report on scaling known techniques to significantly larger and/or more complex real-world problems.
  • Foundations: Papers should incorporate the principles underpinning Code Generation and Optimization (CGO). However, a thorough discussion of theoretical foundations is not required; a summary of such should suffice.
  • Insight: The practical experience should provide meaningful conclusions related to CGO topics that show unexpected or novel characteristics.
  • Applicability: The results of the practical experience should be applicable beyond the specific system evaluated in the paper.

Geographic Diversity and Inclusion

Authors of papers accepted for CGO 2026 are encouraged to present their work in person. However, to foster the participation of students and professionals from everywhere, CGO 2026 will allow the remote presentation of papers, if their authors are unable to travel to the conference venue for reasons beyond their control (e.g. visa issues). Additionally, the conference organization will try to make attendance of CGO 2026 affordable for as many people as possible, with a specific focus on students from universities located in under-represented countries who are paper authors.


Artifact Evaluation

The Artifact Evaluation process is run by a separate committee whose task is to assess how the artifacts support the work described in the papers. This process contributes to improved reproducibility in research that should be a great concern to all of us. There is also some evidence that papers with a supporting artifact receive higher citations than papers without artifact evaluation. Authors of accepted papers at CGO have the option of submitting their artifacts for evaluation within two weeks of paper acceptance. To ease the organization of the AE committee, we kindly ask authors to indicate at the time they submit the paper, whether they are interested in submitting an artifact. Papers that go through the Artifact Evaluation process successfully will receive a seal of approval printed on the papers themselves. Additional information is available on the CGO AE web page. Authors of accepted papers are encouraged, but not required, to make these materials publicly available upon publication of the proceedings.


AUTHORS TAKE NOTE: The official publication date is the date the proceedings are made available in the IEEE Xplore Platform. This date may be up to two weeks prior to the first day of the conference. The official publication date affects the deadline for any patent filings related to published work.

CGO will make the proceedings freely available via the IEEE Xplore platform during the period from two weeks before to two weeks after the conference. This option will facilitate easy access to the proceedings by conference attendees, and it will also enable the community at large to experience the excitement of learning about the latest developments being presented in the period surrounding the event itself.


Distinguished Paper Awards

Up to 10% of papers accepted at CGO 2026 will be designated as Distinguished Papers, following the ACM policy. This award is open to both regular and tool papers.

Submission Site

Papers can be submitted for the second round at https://cgo26-second-round.hotcrp.com/ (the first-round link is https://cgo26-first-round.hotcrp.com/).

Submission Guidelines

Please make sure that your paper satisfies ALL of the following requirements before it is submitted:

  • The paper must be original material that has not been previously published in another conference or journal, nor is currently under review by another conference or journal. Note that you may submit material presented previously at a workshop without copyrighted proceedings.

  • Your submission is limited to ten (10) letter-size (8.5″x11″), single-spaced, double-column pages, using 10pt or larger font, not including references. There is no page limit for references. We strongly encourage the use of the IEEE Conference Template. Submissions not adhering to these submission guidelines may be outright rejected at the discretion of the program chairs. (Please make sure your paper prints satisfactorily on letter-size (8.5″x11″) paper: this is especially important for submissions from countries where A4 paper is standard.)

  • Supplementary materials may be included as an Appendix at the end of the submitted paper. The Appendix has no page limit, but the text of the full paper excluding the Appendix must fit within 10 pages. Reviewers are not required to read the Appendix and may do so at their discretion. So papers must be self-contained without needing to read any material in the Appendix.

  • Papers are to be submitted for double-blind review. Blind reviewing of papers will be done by the program committee, assisted by outside referees. Author names as well as hints of identity are to be removed from the submitted paper. Use care in naming your files. Source file names, e.g., Joe.Smith.dvi, are often embedded in the final output as readily accessible comments. In addition, do not omit references to provide anonymity, as this leaves the reviewer unable to grasp the context. Instead, if you are extending your own work, you need to reference and discuss the past work in third person, as if you were extending someone else’s research. We realize in doing this that for some papers it will still be obvious who the authors are. In this case, the submission will not be penalized as long as a concerted effort was made to reference and describe the relationship to the prior work as if you were extending someone else’s research. For example, if your name is Joe Smith:

    In previous work [1,2], Smith presented a new branch predictor for …. In this paper, we extend their work by …

    Bibliography

    [1] Joe Smith, “A Simple Branch Predictor for …,” Proceedings of CGO 2019.

    [2] Joe Smith, “A More Complicated Branch Predictor for…,” Proceedings of CGO 2019.

  • Your submission must be formatted for black-and-white printers and not color printers. This is especially true for plots and graphs in the paper.

  • Please make sure that the labels on your graphs are readable without the aid of a magnifying glass. Typically the default font sizes on the graph axes in a program like Microsoft Excel are too small.

  • Please number the pages.

  • To ease reviewing, please add line numbers to your submission. In LaTeX you can use the lineno package, with the ‘switch’ option.

  • The paper must be written in English.

  • The paper must be submitted in PDF. We cannot accept any other format, and we must be able to print the document just as we receive it. We strongly suggest that you use only the four widely-used printer fonts: Times, Helvetica, Courier and Symbol. Please make sure that the output has been formatted for printing on LETTER size paper. If generating the paper using “dvips”, use the option “-P cmz -t letter”, and if that is not supported, use “-t letter”.

  • The Artifact Evaluation process is run by a separate committee whose task is to assess how the artifacts support the work described in the papers. Authors of accepted papers have the option of submitting their artifacts for evaluation within one week of paper acceptance. To ease the organization of the AE committee, we kindly ask authors to indicate at the time they submit the paper, whether they are interested in submitting an artifact. Papers that go through the Artifact Evaluation process successfully will receive a seal of approval printed on the papers themselves. Additional information is available on the CGO AE web page. Authors of accepted papers are encouraged, but not required, to make these materials publicly available upon publication of the proceedings.

Conflicts of Interest

Please read these guidelines on conflicts of interest carefully before submission.

  • Authors must register all their conflicts on the paper submission site. Conflicts are needed to ensure appropriate assignment of reviewers. If a paper is found to have an undeclared conflict that causes a problem OR if a paper is found to declare false conflicts in order to abuse or “game” the review system, the paper may be rejected.

  • Please declare a conflict of interest with the following people for any author of your paper:

    • Your Ph.D. advisor(s), post-doctoral advisor(s), Ph.D. students, and post-doctoral advisees, forever.
    • Family relations by blood or marriage, or their equivalent, forever (if they might be potential reviewers).
    • People with whom you have collaborated in the last FIVE years, including:
      • Co-authors of accepted/rejected/pending papers.
      • Co-PIs on accepted/rejected/pending grant proposals.
      • Funders (decision-makers) of your research grants, and researchers whom you fund.
    • People (including students) who shared your primary institution(s) in the last FIVE years.
    • Other relationships, such as close personal friendship, that you think might tend to affect your judgment or be seen as doing so by a reasonable person familiar with the relationship.
    • “Service” collaborations such as co-authoring a report for a professional organization, serving on a program committee, or co-presenting tutorials, do not themselves create a conflict of interest. Co-authoring a paper that is a compendium of various projects with no true collaboration among the projects does not constitute a conflict among the authors of the different projects.
    • Internships represent an institutional conflict during the time that the internship is active but not afterwards, unless another form of conflict exists (e.g. an on-going collaboration or co-authorship of a paper at any point during the previous FIVE years), in which case the conflict is no longer institutional but between individuals.
    • On the other hand, there may be others not covered by the above with whom you believe a COI exists, for example, an ongoing collaboration that has not yet resulted in the creation of a paper or proposal. Please report such COIs; however, you may be asked to justify them. Please be reasonable. For example, you cannot declare a COI with a reviewer just because that reviewer works on topics similar to or related to those in your paper. The PC Chairs may contact co-authors to explain a COI whose origin is unclear.
    • If in doubt, please contact the PC Chairs for advice on whether a conflict exists or not.