Æȷοì
minsub kim
minsub kim
navercorp
navercorp.comÀÇ À̸ÞÀÏ È®ÀεÊ
Á¦¸ñ
Àοë
Àοë
¿¬µµ
Lut-gemm: Quantized matrix multiplication based on luts for efficient inference in large-scale generative language models
G Park, B Park, M Kim, S Lee, J Kim, B Kwon, SJ Kwon, B Kim, Y Lee, ...
arXiv preprint arXiv:2206.09557, 2022
1232022
What changes can large-scale language models bring? intensive study on hyperclova: Billions-scale korean generative pretrained transformers
B Kim
arXiv preprint arXiv:2109.04650, 2021
1212021
Dfx: A low-latency multi-fpga appliance for accelerating transformer-based text generation
S Hong, S Moon, J Kim, S Lee, M Kim, D Lee, JY Kim
2022 55th IEEE/ACM International Symposium on Microarchitecture (MICRO), 616-630, 2022
662022
Reducing tail latency of DNN-based recommender systems using in-storage processing
M Kim, S Lee
Proceedings of the 11th ACM SIGOPS Asia-Pacific Workshop on Systems, 90-97, 2020
122020
Towards scalable analytics with inference-enabled solid-state drives
M Kim, J Kung, S Lee
IEEE Computer Architecture Letters 19 (1), 13-17, 2019
42019
ÇöÀç ½Ã½ºÅÛÀÌ ÀÛµ¿µÇÁö ¾Ê½À´Ï´Ù. ³ªÁß¿¡ ´Ù½Ã ½ÃµµÇØ ÁÖ¼¼¿ä.
ÇмúÀÚ·á 1–5