Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning Q Zhang, M Chen, A Bukharin, P He, Y Cheng, W Chen, T Zhao International Conference on Learning Representation, 2023, 2023 | 141 | 2023 |
Adashift: Decorrelation and convergence of adaptive learning rate methods Z Zhou*, Q Zhang*, G Lu, H Wang, W Zhang, Y Yu International Conference on Learning Representation, 2019, 2018 | 65 | 2018 |
PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance Q Zhang, S Zuo, C Liang, A Bukharin, P He, W Chen, T Zhao Proceedings of the 39th International Conference on Machine Learning, 2022, 2022 | 49 | 2022 |
LoSparse: Structured Compression of Large Language Models based on Low-Rank and Sparse Approximation Y Li, Y Yu, Q Zhang, C Liang, P He, W Chen, T Zhao Proceedings of the 40th International Conference on Machine Learning, 2023, 2023 | 23 | 2023 |
Less is More: Task-aware Layer-wise Distillation for Language Model Compression C Liang, S Zuo, Q Zhang, P He, W Chen, T Zhao Proceedings of the 40th International Conference on Machine Learning, 2023, 2022 | 23 | 2022 |
A Biased Graph Neural Network Sampler with Near-optimal Regret Q Zhang, D Wipf, Q Gan, L Song Advances in Neural Information Processing Systems 35, 2021, 2021 | 23 | 2021 |
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation S Zuo, Q Zhang, C Liang, P He, T Zhao, W Chen North American Chapter of the Association for Computational Linguistics, 2022, 2022 | 21 | 2022 |
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs Q Zhang, C Singh, L Liu, X Liu, B Yu, J Gao, T Zhao International Conference on Learning Representation, 2024, 2023 | 5 | 2023 |
Robust Multi-Agent Reinforcement Learning via Adversarial Regularization: Theoretical Foundation and Stable Algorithms A Bukharin, Y Li, Y Yu, Q Zhang, Z Chen, S Zuo, C Zhang, S Zhang, ... Advances in Neural Information Processing Systems 37, 2023, 2023 | 3 | 2023 |
GEAR: An Efficient KV Cache Compression Recipe for Near-Lossless Generative Inference of LLM H Kang, Q Zhang, S Kundu, G Jeong, Z Liu, T Krishna, T Zhao arXiv preprint arXiv:2403.05527, 2024 | 2 | 2024 |
Efficient Long-Range Transformers: You Need to Attend More, but Not Necessarily at Every Layer Q Zhang, D Ram, C Hawkins, S Zha, T Zhao The 2023 Conference on Empirical Methods in Natural Language Processing …, 2023 | 2 | 2023 |
A Non-asymptotic comparison of SVRG and SGD: tradeoffs between compute and speed Q Zhang, Y Wu, F Faghri, T Zhang, J Ba | | 2019 |