Follow
Beidi Chen
Beidi Chen
Verified email at andrew.cmu.edu
Title
Cited by
Cited by
Year
Efficient streaming language models with attention sinks
G Xiao, Y Tian, B Chen, S Han, M Lewis
arXiv preprint arXiv:2309.17453, 2023
2952023
FlexGen: High-Throughput Generative Inference of Large Language Models with a Single GPU
Y Sheng, L Zheng, B Yuan, Z Li, M Ryabinin, B Chen, P Liang, C Re, ...
International Conference on Machine Learning, 2023
2772023
Deja vu: Contextual sparsity for efficient llms at inference time
Z Liu, J Wang, T Dao, T Zhou, B Yuan, Z Song, A Shrivastava, C Zhang, ...
International Conference on Machine Learning, 22137-22176, 2023
2012023
H O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models
Z Zhang, Y Sheng, T Zhou, T Chen, L Zheng, R Cai, Z Song, Y Tian, C Ré, ...
International Conference on Machine Learning, 2023
1892023
SLIDE: In Defense of Smart Algorithms over Hardware Acceleration for Large-scale Deep Learning Systems
B Chen, T Medini, J Farwell, S Gobriel, C Tai, A Shrivastava
Proceedings of Machine Learning and System 2, 291--306, 2020
1402020
Scatterbrain: Unifying sparse and low-rank attention
B Chen, T Dao, E Winsor, Z Song, A Rudra, C Ré
Advances in Neural Information Processing Systems 34, 17413-17426, 2021
1222021
Monarch: Expressive structured matrices for efficient and accurate training
T Dao, B Chen, NS Sohoni, A Desai, M Poli, J Grogan, A Liu, A Rao, ...
International Conference on Machine Learning, 4690-4721, 2022
862022
Decentralized training of foundation models in heterogeneous environments
B Yuan, Y He, JQ Davis, T Zhang, T Dao, B Chen, P Liang, C Re, C Zhang
Neural Information Processing Systems., 2022
812022
Pixelated butterfly: Simple and efficient sparse training for neural network models
B Chen, T Dao, K Liang, J Yang, Z Song, A Rudra, C Re
International Conference on Learning Representations, 2022
79*2022
MONGOOSE: A learnable LSH framework for efficient neural network training
B Chen, Z Liu, B Peng, Z Xu, JL Li, T Dao, Z Song, A Shrivastava, C Re
International Conference on Learning Representations, 2021
772021
Scan and Snap: Understanding Training Dynamics and Token Composition in 1-layer Transformer
Y Tian, Y Wang, B Chen, S Du
International Conference on Machine Learning, 2023
712023
Analyzing log analysis: An empirical study of user log mining
S Alspaugh, B Chen, J Lin, A Ganapathi, M Hearst, R Katz
28th Large Installation System Administration Conference (LISA14), 62-77, 2014
702014
Angular visual hardness
B Chen, W Liu, Z Yu, J Kautz, A Shrivastava, A Garg, A Anandkumar
International Conference on Machine Learning, 1637-1648, 2020
522020
Fast and accurate stochastic gradient estimation
B Chen, Y Xu, A Shrivastava
Advances in Neural Information Processing Systems 32, 2019
52*2019
Unique entity estimation with application to the Syrian conflict
B Chen, A Shrivastava, RC Steorts
The Annals of Applied Statistics 12 (2), 1039-1067, 2018
382018
Joma: Demystifying multilayer transformers via joint dynamics of mlp and attention
Y Tian, Y Wang, Z Zhang, B Chen, S Du
arXiv preprint arXiv:2310.00535, 2023
372023
Llm inference unveiled: Survey and roofline model insights
Z Yuan, Y Shang, Y Zhou, Z Dong, Z Zhou, C Xue, B Wu, Z Li, Q Gu, ...
arXiv preprint arXiv:2402.16363, 2024
362024
Anima Anandkumar, and Yuandong Tian. Galore: Memory-efficient llm training by gradient low-rank projection
J Zhao, Z Zhang, B Chen, Z Wang
arXiv preprint arXiv:2403.03507, 2024
332024
Galore: Memory-efficient llm training by gradient low-rank projection
J Zhao, Z Zhang, B Chen, Z Wang, A Anandkumar, Y Tian
arXiv preprint arXiv:2403.03507, 2024
322024
A tuning-free asymmetric 2bit quantization for kv cache
Z Liu, J Yuan, H Jin, S Zhong, Z Xu, V Braverman, B Chen, XH Kivi
arXiv preprint arXiv:2402.02750, 2024
312024
The system can't perform the operation now. Try again later.
Articles 1–20