Follow
Qingru Zhang
Title
Cited by
Cited by
Year
Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning
Q Zhang, M Chen, A Bukharin, P He, Y Cheng, W Chen, T Zhao
International Conference on Learning Representation, 2023, 2023
1352023
Adashift: Decorrelation and convergence of adaptive learning rate methods
Z Zhou*, Q Zhang*, G Lu, H Wang, W Zhang, Y Yu
International Conference on Learning Representation, 2019, 2018
652018
PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance
Q Zhang, S Zuo, C Liang, A Bukharin, P He, W Chen, T Zhao
Proceedings of the 39th International Conference on Machine Learning, 2022, 2022
482022
A Biased Graph Neural Network Sampler with Near-optimal Regret
Q Zhang, D Wipf, Q Gan, L Song
Advances in Neural Information Processing Systems 35, 2021, 2021
232021
LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation
Y Li, Y Yu, Q Zhang, C Liang, P He, W Chen, T Zhao
Proceedings of the 40th International Conference on Machine Learning, 2023, 2023
222023
Less is More: Task-aware Layer-wise Distillation for Language Model Compression
C Liang, S Zuo, Q Zhang, P He, W Chen, T Zhao
Proceedings of the 40th International Conference on Machine Learning, 2023, 2022
222022
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation
S Zuo, Q Zhang, C Liang, P He, T Zhao, W Chen
North American Chapter of the Association for Computational Linguistics, 2022, 2022
212022
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs
Q Zhang, C Singh, L Liu, X Liu, B Yu, J Gao, T Zhao
International Conference on Learning Representation, 2024, 2023
42023
Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms
A Bukharin, Y Li, Y Yu, Q Zhang, Z Chen, S Zuo, C Zhang, S Zhang, ...
Advances in Neural Information Processing Systems 37, 2023, 2023
32023
Efficient Long-Range Transformers: You Need to Attend More, but Not Necessarily at Every Layer
Q Zhang, D Ram, C Hawkins, S Zha, T Zhao
The 2023 Conference on Empirical Methods in Natural Language Processing …, 2023
22023
GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM
H Kang, Q Zhang, S Kundu, G Jeong, Z Liu, T Krishna, T Zhao
arXiv preprint arXiv:2403.05527, 2024
12024
A Non-asymptotic comparison of SVRG and SGD: tradeoffs between compute and speed
Q Zhang, Y Wu, F Faghri, T Zhang, J Ba
2019
The system can't perform the operation now. Try again later.
Articles 1–12