Follow
Yu Emma Wang
Title
Cited by
Cited by
Year
Glam: Efficient scaling of language models with mixture-of-experts
N Du, Y Huang, AM Dai, S Tong, D Lepikhin, Y Xu, M Krikun, Y Zhou, ...
International Conference on Machine Learning, 5547-5569, 2022
3302022
A Systematic Methodology for Analysis of Deep Learning Hardware and Software Platforms
YE Wang, GY Wei, D Brooks
the 3rd MLSys Conference, 2020
64*2020
Exploiting parallelism opportunities with deep learning frameworks
YE Wang, CJ Wu, X Wang, K Hazelwood, D Brooks
ACM Transactions on Architecture and Code Optimization (TACO) 18 (1), 1-23, 2020
312020
Predicting New Workload or CPU Performance by Analyzing Public Datasets
YE Wang, V Lee, GY Wei, D Brooks
ACM Transactions on Architecture and Code Optimization 15 (4), 53, 2019
302019
A flexible approach to autotuning multi-pass machine learning compilers
PM Phothilimthana, A Sabne, N Sarda, KS Murthy, Y Zhou, ...
2021 30th International Conference on Parallel Architectures and Compilation …, 2021
242021
Exploring the limits of Concurrency in ML Training on Google TPUs
S Kumar, Y Wang, C Young, J Bradbury, N Kumar, D Chen, A Swing
Proceedings of Machine Learning and Systems 3, 81-92, 2021
162021
A Learning Algorithm for Bayesian Networks and Its Efficient Implementation on GPUs
Y Wang, W Qian, S Zhang, X Liang, B Yuan
IEEE Transactions on Parallel and Distributed Systems 27 (1), 17-30, 2016
102016
Demystifying Bayesian Inference Workloads
YE Wang, Y Zhu, GG Ko, B Reagen, GY Wei, D Brooks
IEEE International Symposium on Performance Analysis of Systems and Software, 2019
92019
AutoDistill: An end-to-end framework to explore and distill hardware-efficient language models
X Zhang, Z Zhou, D Chen, YE Wang
arXiv preprint arXiv:2201.08539, 2022
72022
Yazhou Zu, Yuanzhong Xu, and Andy Swing. Exploring the limits of concurrency in ml training on google TPUs
S Kumar, J Bradbury, C Young, YE Wang, A Levskaya, B Hechtman, ...
arXiv preprint arXiv:2011.03641 1 (1), 1.2, 2020
72020
Exploring hardware profile-guided green datacenter scheduling
W Tang, Y Wang, H Liu, T Zhang, C Li, X Liang
2015 44th International Conference on Parallel Processing, 11-20, 2015
52015
Mixed precision post training quantization of neural networks with sensitivity guided search
CJS Schaefer, E Guo, C Stanton, X Zhang, T Jablin, N Lambert-Shirzad, ...
arXiv preprint arXiv:2302.01382, 2023
12023
Hadamard Domain Training with Integers for Class Incremental Quantized Learning
M Schiemer, CJS Schaefer, JP Vap, MJ Horeni, YE Wang, J Ye, S Joshi
arXiv preprint arXiv:2310.03675, 2023
2023
Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
CJS Schaefer, N Lambert-Shirzad, X Zhang, C Chou, T Jablin, J Li, E Guo, ...
arXiv preprint arXiv:2306.04879, 2023
2023
Sparsely Activated Language Models are Efficient In-Context Learners
A Yu, A Dai, C Cui, DD Lepikhin, E Wang, K Meier-Hellstern, K Webster, ...
2022
Performance Analysis for Machine Learning Applications
YE Wang
Harvard University, 2019
2019
The system can't perform the operation now. Try again later.
Articles 1–16