Follow
Janghwan Lee
Title
Cited by
Cited by
Year
Token-Scaled Logit Distillation for Ternary Weight Generative Language Models
M Kim, S Lee, J Lee, S Hong, DS Chang, W Sung, J Choi
Advances in Neural Information Processing Systems 36, 2023
62023
Finding optimal numerical format for sub-8-bit post-training quantization of vision transformers
J Lee, Y Hwang, J Choi
ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and …, 2023
32023
Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization
J Lee, M Kim, S Baek, SJ Hwang, W Sung, J Choi
The 2023 Conference on Empirical Methods in Natural Language Processing, 2023
22023
PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices
M Lee, H Kim, S Park, M Yoon, J Lee, J Choi, M Kang, J Choi
arXiv preprint arXiv:2305.07522, 2023
22023
Range-Invariant Approximation of Non-Linear Operations for Efficient BERT Fine-Tuning
J Kim, J Lee, J Choi, JH Han, S Lee
2023 60th ACM/IEEE Design Automation Conference (DAC), 1-6, 2023
12023
Optimizing Exponent Bias for Sub-8bit Floating-Point Inference of Fine-tuned Transformers
J Lee, J Choi
2022 IEEE 4th International Conference on Artificial Intelligence Circuits …, 2022
12022
SPADE: Sparse Pillar-based 3D Object Detection Accelerator for Autonomous Driving
M Lee, S Park, H Kim, M Yoon, J Lee, JW Choi, NS Kim, M Kang, J Choi
2024 IEEE International Symposium on High-Performance Computer Architecture …, 2024
2024
Searching Optimal Floating-Point Format for Sub-8-Bit Large Language Model Inference
Y Hwang, J Lee, J Park, J Lim, J Choi
2024 International Conference on Electronics, Information, and Communication …, 2024
2024
The system can't perform the operation now. Try again later.
Articles 1–8