Shizhe Diao
Cited by
Cited by
Raft: Reward ranked finetuning for generative foundation model alignment
H Dong, W Xiong, D Goyal, Y Zhang, W Chow, R Pan, S Diao, J Zhang, ...
arXiv preprint arXiv:2304.06767, 2023
ZEN: Pre-training Chinese text encoder enhanced by n-gram representations
S Diao, J Bai, Y Song, T Zhang, Y Wang
Findings of EMNLP 2020, 2019
Active prompting with chain-of-thought for large language models
S Diao, P Wang, Y Lin, T Zhang
arXiv preprint arXiv:2302.12246, 2023
Black-Box Prompt Learning for Pre-trained Language Models
S Diao, Z Huang, R Xu, X Li, Y Lin, X Zhou, T Zhang
Transactions on Machine Learning Research (TMLR), 2022
Automatic prompt augmentation and selection with chain-of-thought from labeled data
KS Shum, S Diao, T Zhang
arXiv preprint arXiv:2302.12822, 2023
Taming pre-trained language models with n-gram representations for low-resource domain adaptation
S Diao, R Xu, H Su, Y Jiang, Y Song, T Zhang
Proceedings of the 59th Annual Meeting of the Association for Computational …, 2021
Detgpt: Detect what you need via reasoning
R Pi, J Gao, S Diao, R Pan, H Dong, J Zhang, L Yao, J Han, H Xu, L Kong, ...
arXiv preprint arXiv:2305.14167, 2023
Lmflow: An extensible toolkit for finetuning and inference of large foundation models
S Diao, R Pan, H Dong, KS Shum, J Zhang, W Xiong, T Zhang
arXiv preprint arXiv:2306.12420, 2023
Efficient neural network training via forward and backward propagation sparsification
X Zhou, W Zhang, Z Chen, S Diao, T Zhang
Advances in neural information processing systems 34, 15216-15229, 2021
R-tuning: Teaching large language models to refuse unknown questions
H Zhang, S Diao, Y Lin, YR Fung, Q Lian, X Wang, Y Chen, H Ji, T Zhang
arXiv preprint arXiv:2311.09677, 2023
Vlue: A multi-task multi-dimension benchmark for evaluating vision-language pre-training
W Zhou, Y Zeng, S Diao, X Zhang
International Conference on Machine Learning, 27395-27411, 2022
TILGAN: transformer-based implicit latent GAN for diverse and coherent text generation
S Diao, X Shen, K Shum, Y Song, T Zhang
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 …, 2021
Mixture-of-Domain-Adapters: Decoupling and Injecting Domain Knowledge to Pre-trained Language Models Memories
S Diao, T Xu, R Xu, J Wang, T Zhang
arXiv preprint arXiv:2306.05406, 2023
Write and Paint: Generative Vision-Language Models are Unified Modal Learners
S Diao, W Zhou, X Zhang, J Wang
ICLR 2023, 0
Towards unifying medical vision-and-language pre-training via soft prompts
Z Chen, S Diao, B Wang, G Li, X Wan
Proceedings of the IEEE/CVF International Conference on Computer Vision …, 2023
Unitime: A language-empowered unified model for cross-domain time series forecasting
X Liu, J Hu, Y Li, S Diao, Y Liang, B Hooi, R Zimmermann
arXiv preprint arXiv:2310.09751, 2023
Speciality vs generality: An empirical study on catastrophic forgetting in fine-tuning foundation models
Y Lin, L Tan, H Lin, Z Zheng, R Pi, J Zhang, S Diao, H Wang, H Zhao, ...
arXiv preprint arXiv:2309.06256, 2023
Keyphrase generation with cross-document attention
S Diao, Y Song, T Zhang
arXiv preprint arXiv:2004.09800, 2020
GubaLex: Guba-oriented sentiment lexicon for big texts in finance
Y Sun, M Fang, X Wang, S Diao
2017 13th International Conference on Semantics, Knowledge and Grids (SKG …, 2017
On the Difference of BERT-style and CLIP-style Text Encoders
Z Chen, GH Chen, S Diao, X Wan, B Wang
arXiv preprint arXiv:2306.03678, 2023
The system can't perform the operation now. Try again later.
Articles 1–20