OpenFold: Retraining AlphaFold2 yields new insights into its learning mechanisms and capacity for generalization G Ahdritz, N Bouatta, C Floristean, S Kadyan, Q Xia, W Gerecke, ... Nature Methods, 1-11, 2024 | 102 | 2024 |
A length adaptive algorithm-hardware co-design of transformer on fpga through sparse attention and dynamic pipelining H Peng, S Huang, S Chen, B Li, T Geng, A Li, W Jiang, W Wen, J Bi, H Liu, ... Proceedings of the 59th ACM/IEEE Design Automation Conference, 1135-1140, 2022 | 37 | 2022 |
Et: re-thinking self-attention for transformer models on gpus S Chen, S Huang, S Pandey, B Li, GR Gao, L Zheng, C Ding, H Liu Proceedings of the international conference for high performance computing …, 2021 | 31 | 2021 |
Sparse progressive distillation: Resolving overfitting under pretrain-and-finetune paradigm S Huang, D Xu, IEH Yen, Y Wang, SE Chang, B Li, S Chen, M Xie, ... arXiv preprint arXiv:2110.08190, 2021 | 27 | 2021 |
Optimizing fpga-based accelerator design for large-scale molecular similarity search (special session paper) H Peng, S Chen, Z Wang, J Yang, SA Weitze, T Geng, A Li, J Bi, M Song, ... 2021 IEEE/ACM International Conference On Computer Aided Design (ICCAD), 1-7, 2021 | 12 | 2021 |
HMC-TRAN A Tensor-core Inspired Hierarchical Model Compression for Transformer-based DNNs on GPU S Huang, S Chen, H Peng, D Manu, Z Kong, G Yuan, L Yang, S Wang, ... Proceedings of the 2021 on Great Lakes Symposium on VLSI, 169-174, 2021 | 7 | 2021 |
Zhenglun Kong, Geng Yuan, Lei Yang, Shusen Wang, Hang Liu, and Caiwen Ding. Hmc-tran: A tensor-core inspired hierarchical model compression for transformer-based dnns on gpu S Huang, S Chen, H Peng, D Manu Proceedings of the 2021 on Great Lakes Symposium on VLSI, 169-174, 2021 | 7 | 2021 |
Motif-based graph representation learning with application to chemical molecules Y Wang, S Chen, G Chen, E Shurberg, H Liu, P Hong Informatics 10 (1), 8, 2023 | 3 | 2023 |
TANGO: re-thinking quantization for graph neural network training on GPUs S Chen, D Zheng, C Ding, C Huan, Y Ji, H Liu Proceedings of the International Conference for High Performance Computing …, 2023 | 2 | 2023 |
FP6-LLM: Efficiently Serving Large Language Models Through FP6-Centric Algorithm-System Co-Design H Xia, Z Zheng, X Wu, S Chen, Z Yao, S Youn, A Bakhtiari, M Wyatt, ... arXiv preprint arXiv:2401.14112, 2024 | 1 | 2024 |
Zeroquant (4+ 2): Redefining llms quantization with a new fp6-centric strategy for diverse generative tasks X Wu, H Xia, S Youn, Z Zheng, S Chen, A Bakhtiari, M Wyatt, Y He, ... arXiv preprint arXiv:2312.08583, 2023 | 1 | 2023 |
DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery through Sophisticated AI System Technologies SL Song, B Kruft, M Zhang, C Li, S Chen, C Zhang, M Tanaka, X Wu, ... arXiv preprint arXiv:2310.04610, 2023 | 1 | 2023 |
Kernel Fusion in Atomistic Spin Dynamics Simulations on Nvidia GPUs using Tensor Core H Chen, S Chen, JJ Turner, A Feiguin arXiv preprint arXiv:2308.07487, 2023 | | 2023 |
Motif-based Graph Representation Learning with Application to Chemical Molecules Y Wang, S Chen, G Chen, E Shurberg, H Liu, P Hong arXiv preprint arXiv:2208.04529, 2022 | | 2022 |