Ptq4vit: Post-training quantization for vision transformers with twin uniform quantization Z Yuan, C Xue, Y Chen, Q Wu, G Sun European conference on computer vision, 191-207, 2022 | 102 | 2022 |
PTQ4ViT: Post-training quantization framework for vision transformers with twin uniform quantization Z Yuan, C Xue, Y Chen, Q Wu, G Sun arXiv preprint arXiv:2111.12293, 2021 | 36 | 2021 |
Latency-aware spatial-wise dynamic networks Y Han, Z Yuan, Y Pu, C Xue, S Song, G Sun, G Huang Advances in Neural Information Processing Systems 35, 36845-36857, 2022 | 22 | 2022 |
Llm inference unveiled: Survey and roofline model insights Z Yuan, Y Shang, Y Zhou, Z Dong, C Xue, B Wu, Z Li, Q Gu, YJ Lee, ... arXiv preprint arXiv:2402.16363, 2024 | 20 | 2024 |
The dawn of ai-native eda: Promises and challenges of large circuit models L Chen, Y Chen, Z Chu, W Fang, TY Ho, Y Huang, S Khan, M Li, X Li, ... arXiv preprint arXiv:2403.07257, 2024 | 11 | 2024 |
Ptq-sl: Exploring the sub-layerwise post-training quantization Z Yuan, Y Chen, C Xue, C Zhang, Q Wang, G Sun arXiv preprint arXiv:2110.07809, 2021 | 3 | 2021 |
Theseus: Towards High-Efficiency Wafer-Scale Chip Design Space Exploration for Large Language Models J Zhu, C Xue, Y Chen, Z Wang, G Sun arXiv preprint arXiv:2407.02079, 2024 | 1 | 2024 |