Publications
60+CCF-AOverall Pubs
20+Architecture ISCA / MICRO / HPCA/ ASPLOS
20+EDA & AutomationDAC / ICCAD
3+AwardsBest /Outstanding/ Feature Papers
{=} Equal contribution {*} Corresponding author [Google Scholar Profile]
2026
[ISCA '26]
[ISCA '26]
[ISCA '26]
[DAC '26]
[DAC '26]
[DAC '26]
[DAC '26]
[DAC '26]
[Tech Report]
HyperOffload: Graph-Driven Hierarchical Memory Management for LLMs on SuperNode Applied at HUAWEI MindSpore
[TACO '26]
[ASPLOS '26]
[ASPLOS '26]
Nebula: Infinite-Scale 3D Gaussian Splatting in VR via Collaborative Rendering (Top Conf. in Arch)CCF-AAcc: 10.6%
[PPoPP '26]
BEEMS: Boosting Machine Vision Efficiency via Computation Graph-Based Memory Smoothing (Top Conf. in Sys.)CCF-AAcc: 11.5%
[PPoPP '26]
[DATE '26]
LaMoS: Enabling Efficient Large Number Modular Multiplication through SRAM-based CiM (Top Conf. in EDA)CCF-BAcc: 25%
[AAAI '26]
SpecQuant: Spectral Decomposition and Adaptive Truncation for Ultra-Low-Bit LLMs Quantization (Top Conf. in AI)CCF-AAcc: 17.6%
[HPCA '26]
ORANGE: Exploring Ockham's Razor for Neural Rendering by Accelerating 3DGS on NPUs (Top Conf. in Arch)CCF-AAcc: 19.7%
[HPCA '26]
Splatonic: Architecture Support for 3D Gaussian Splatting SLAM via Sparse Processing (Top Conf. in Arch)CCF-AAcc: 19.7%
[ASPDAC '26]
BLADE: Boosting LLM Decoding's Communication Efficiency in DRAM-based PIM (Top Conf. in EDA)CCF-AAcc: 29%
[ASPLOS '26]
ASDR: Exploiting Adaptive Sampling and Data Reuse for CIM-based Instant Neural Rendering (Top Conf. in Arch)CCF-AAcc: 23%
[ASPLOS '26]
2025
[TACO '25]
[TACO '25]
[EMNLP '25]
FlexQuant: A Flexible and Efficient Dynamic Precision Switching Framework for LLM Quantization (Top Conf. in NLP)CCF-BApplied at Ant Group
[ACM MM '25]
ASTER: Adaptive Dynamic Layer-Skipping for Efficient Transformer Inference via MDP (Top Conf. in Multimedia/AI)CCF-A🏆 Outstanding Paper
[ICCAD '25]
QUARK: Quantization-Enabled Circuit Sharing for Transformer Acceleration via Pattern Exploiting (Top Conf. in EDA)CCF-BAcc: 24%
[ICCAD '25]
PLAIN: Leveraging High Internal Bandwidth in PIM for LLM Inference via Mixed-Precision (Top Conf. in EDA)CCF-BAcc: 24%
[ICCAD '25]
DevTrace: Efficient and Fine-grained PCIe Transaction Tracing for Edge Intelligence Workloads (Top Conf. in EDA)CCF-BAcc: 24%
[TCAD '25]
[TACO '25]
Attack and Defense: Enhancing Robustness of Binary Hyper-Dimensional Computing (Top Jour. in Arch)CCF-A
[ISCA '25]
FATE: Boosting HDC Intelligence Performance with Flexible Numerical Data Type (Top Conf. in Arch)CCF-AAcc: 21%
[ISCA '25]
Qtenon: Towards Low-Latency Architecture Integration for Accelerating Hybrid Quantum-Classical Computing (Top Conf. in Arch)CCF-AAcc: 21%
[DAC '25]
ALLMod: Exploring Area-Efficiency of LUT-based Large Number Modular Reduction via Hybrid Workloads (Top Conf. in EDA)CCF-AAcc: 23%
[DAC '25]
MILLION: MasterIng Long-Context LLM Inference Via Outlier-Immunized KV Product Quantization (Top Conf. in EDA)CCF-AApplied at HUAWEI
[DAC '25]
BLOOM: Bit-Slice Framework for DNN Acceleration with Mixed-Precision (Top Conf. in EDA)CCF-AApplied at ZTE
[DAC '25]
PISA: Efficient Precision-Slice Framework for LLMs with Adaptive Numerical Type (Top Conf. in EDA)CCF-AAcc: 23%
[DATE '25]
HyperDyn: Dynamic Dimensional Masking for Efficient Hyper-Dimensional Computing (Top Conf. in EDA)CCF-BAcc: 21%
[DATE '25]
HyperNeO: Efficient and Accurate Hyper-Dimensional Regression via Neural Optimization (Top Conf. in EDA)CCF-BAcc: 21%
[DATE '25]
TAIL: Temporal Asynchronous Execution for Efficient SNNs with Inter-Layer Parallelism (Top Conf. in EDA)CCF-BAcc: 21%
[DATE '25]
[DATE '25]
[HPCA '25]
CROSS: Compiler-Driven Optimization of Sparse DNNs Using Sparse/Dense Computation Kernels (Top Conf. in Arch)CCF-AAcc: 21%
[HPCA '25]
GSArch: Breaking Memory Barriers in 3D Gaussian Splatting Training via Architectural Support (Top Conf. in Arch)CCF-AAcc: 21%
[ASPDAC '25]
[ASPDAC '25]
2024
[TCAS-AI '24]
[TODAES '24]
[TCAD '24]
[ICCD '24]
[ICCD '24]
[ICCD '24]
[ICCD '24]
[MICRO '24]
COMPASS: SRAM-Based Computing-in-Memory SNN Accelerator with Adaptive Spike Speculation (Top Conf. in Arch)CCF-AAcc: 22%
[MICRO '24]
SRender: Boosting Neural Radiance Field Efficiency via Sensitivity-Aware Dynamic Precision Rendering (Top Conf. in Arch)CCF-AAcc: 22%
[TPDS '24]
Exploiting Temporal-Unrolled Parallelism for Energy-Efficient SNN Acceleration (Top Jour. in Arch)CCF-A
[ISLPED '24]
[ISCA '24]
[DAC '24]
INSPIRE: Accelerating Deep Neural Networks via Hardware-friendly Index-Pair Encoding (Top Conf. in EDA)CCF-AAcc: 23%
[DAC '24]
EOS: An Energy-Oriented Attack Framework for Spiking Neural Networks (Top Conf. in EDA)CCF-AAcc: 23%
[DATE '24]
RTSA: An RRAM-TCAM based In-Memory-Search Accelerator for Sub-100 μs Collision Detection (Top Conf. in EDA)CCF-BAcc: 24%
[ASPLOS '24]
CMC: Video Transformer Acceleration via CODEC Assisted Matrix Condensing (Top Conf. in Arch)CCF-AAcc: 24%
[HPCA '24]
SPARK: Scalable and Precision-Aware Acceleration of Neural Networks via Efficient Encoding (Top Conf. in Arch)CCF-AAcc: 18%
[ASPDAC '24]
PAAP-HD: PIM-Assisted Approximation for Efficient Hyper-Dimensional Computing (Top Conf. in EDA)Acc: 29%
[ASPDAC '24]
TEAS: Exploiting Spiking Activity for Temporal-wise Adaptive Spiking Neural Networks (Top Conf. in EDA)Acc: 29%
[ASPDAC '24]
TSTC: Enabling Efficient Training via Structured Sparse Tensor Compilation (Top Conf. in EDA)Acc: 29%
[ASPDAC '24]
HyperFeel: An Efficient Federated Learning Framework Using Hyperdimensional Computing (Top Conf. in EDA)Acc: 29%
[DAC '24]
HyperAttack: An Efficient Attack Framework for HyperDimensional Computing (Top Conf. in EDA)CCF-AAcc: 23%
2023
[IEEE TC '23]
[ICCD '23]
[ICCAD '23]
HyperNode: An Efficient Node Classification Framework Using HyperDimensional Computing (Top Conf. in EDA)Acc: 23%
2022
[SIGIR '22]
[TCAD '22]
[DAC '22]
[DAC '22]
[DAC '22]
[DATE '22]
Self-Terminated Write of MLC ReRAM for Efficient Neuromorphic Computing 🏆 Best Paper Award(Top Conf. in EDA)
[AAAI '22]
SpikeConverter: An Efficient Conversion Framework Zipping the Gap between ANNs and SNNs 🏆 Oral (Top 15%)(Top Conf. in AI)
[ICASSP '22]
Dynsnn: A dynamic approach to reduce redundancy in spiking neural networks (Top Conf. in Signal Processing)CCF-B
[DATE '22]
[ASPDAC '22]
[DAC '22]
[ICMR '22]
2021
[Frontiers '21]
SSTDP: Supervised Spike Timing Dependent Plasticity for Efficient SNN Training (SCI Tier 2, Impact Factor: 4.7)
[ICCD '21]
[ICCV '21]
Improving NN Efficiency via Post-training Quantization with Adaptive Floating-Point Acc: 25.9%(Top Conf. in Vision)
[ICCAD '21]
