Follow
Sheng-Chun Kao
Title
Cited by
Cited by
Year
Gamma: Automating the hw mapping of dnn models on accelerators via genetic algorithm
SC Kao, T Krishna
Proceedings of the 39th International Conference on Computer-Aided Design, 1-9, 2020
1332020
Confuciux: Autonomous hardware resource assignment for dnn accelerators using reinforcement learning
SC Kao, G Jeong, T Krishna
2020 53rd Annual IEEE/ACM International Symposium on Microarchitecture …, 2020
1072020
Magma: An optimization framework for mapping multiple dnns on multiple accelerator cores
SC Kao, T Krishna
2022 IEEE International Symposium on High-Performance Computer Architecture …, 2022
432022
FLAT: An Optimized Dataflow for Mitigating Attention Performance Bottlenecks
SC Kao, S Subramanian, G Agrawal, T Krishna
arXiv preprint arXiv:2107.06419, 2021
39*2021
Digamma: Domain-aware genetic algorithm for hw-mapping co-optimization for dnn accelerators
SC Kao, M Pellauer, A Parashar, T Krishna
2022 Design, Automation & Test in Europe Conference & Exhibition (DATE), 232-237, 2022
232022
Extending sparse tensor accelerators to support multiple compression formats
E Qin, G Jeong, W Won, SC Kao, H Kwon, S Srinivasan, D Das, GE Moon, ...
2021 IEEE International Parallel and Distributed Processing Symposium (IPDPS …, 2021
172021
Reinforcement learning based interconnection routing for adaptive traffic optimization
SC Kao, CHH Yang, PY Chen, X Ma, T Krishna
Proceedings of the 13th IEEE/ACM international symposium on networks-on-chip …, 2019
152019
Dynamically updatable ternary segmented aging bloom filter for openflow-compliant low-power packet processing
SC Kao, DY Lee, TS Chen, AY Wu
IEEE/ACM Transactions on Networking 26 (2), 1004-1017, 2018
142018
Demystifying map space exploration for NPUs
SC Kao, A Parashar, PA Tsai, T Krishna
2022 IEEE International Symposium on Workload Characterization (IISWC), 269-281, 2022
122022
Training recipe for n: M structured sparsity with decaying pruning mask
SC Kao, A Yazdanbakhsh, S Subramanian, S Agrawal, U Evci, T Krishna
arXiv preprint arXiv:2209.07617, 2022
92022
DNNFuser: Generative pre-trained transformer as a generalized mapper for layer fusion in dnn accelerators
SC Kao, X Huang, T Krishna
arXiv preprint arXiv:2201.11218, 2022
72022
DiGamma: domain-aware genetic algorithm for HW-mapping cooptimization for DNN accelerators. In 2022 Design, Automation & Test in Europe Conference & Exhibition (DATE)
SC Kao, M Pellauer, A Parashar, T Krishna
IEEE, 2022
72022
Domain-specific genetic algorithm for multi-tenant dnnaccelerator scheduling
SC Kao, T Krishna
arXiv preprint arXiv:2104.13997, 2021
62021
E3: A hw/sw co-design neuroevolution platform for autonomous learning in edge device
SC Kao, T Krishna
2021 IEEE International Symposium on Performance Analysis of Systems and …, 2021
52021
JaxPruner: A concise library for sparsity research
JH Lee, W Park, NE Mitchell, J Pilault, JSO Ceron, HB Kim, N Lee, ...
Conference on Parsimony and Learning, 515-528, 2024
42024
A Formalism of DNN Accelerator Flexibility
SC Kao, H Kwon, M Pellauer, A Parashar, T Krishna
Proceedings of the ACM on Measurement and Analysis of Computing Systems 6 (2 …, 2022
32022
Progressive Gradient Flow for Robust N: M Sparsity Training in Transformers
AR Bambhaniya, A Yazdanbakhsh, S Subramanian, SC Kao, S Agrawal, ...
arXiv preprint arXiv:2402.04744, 2024
12024
Training Recipe for N: M Structured Sparsity with Decaying Pruning Mask
A Yazdanbakhsh, SC Kao, S Agrawal, S Subramanian, T Krishna, U Evci
arXiv preprint arXiv:2209.07617, 2022
12022
ATTACC the Quadratic Bottleneck of Attention Layers.
SC Kao, S Subramanian, G Agrawal, T Krishna
ArXiv, abs/2107.06419, 2021
12021
Generative Design of Hardware-aware DNNs
SC Kao, A Ramamurthy, T Krishna
arXiv preprint arXiv:2006.03968, 2020
12020
The system can't perform the operation now. Try again later.
Articles 1–20