Follow
In Gim
Title
Cited by
Cited by
Year
Memory-efficient DNN Training on Mobile Devices
I Gim, JG Ko
Proceedings of the 20th Annual International Conference on Mobile Systems …, 2022
222022
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
I Gim, G Chen, S Lee, N Sarda, A Khandelwal, L Zhong
arXiv preprint arXiv:2311.04934, 2023
92023
Fast Monte-Carlo Approximation of the Attention Mechanism
H Kim, JG Ko
Proceedings of the 36th AAAI Conference on Artificial Intelligence 36 (Vol …, 2022
12022
The system can't perform the operation now. Try again later.
Articles 1–3