Æȷοì
Seungone Kim
Á¦¸ñ
Àοë
Àοë
¿¬µµ
Prometheus: Inducing fine-grained evaluation capability in language models
S Kim, J Shin, Y Cho, J Jang, S Longpre, H Lee, S Yun, S Shin, S Kim, ...
ICLR 2024, 2023
129*2023
Personalized soups: Personalized large language model alignment via post-hoc parameter merging
J Jang, S Kim, BY Lin, Y Wang, J Hessel, L Zettlemoyer, H Hajishirzi, ...
NeurIPS 2024 AFM Workshop, 2023
792023
Prometheus 2: An open source language model specialized in evaluating other language models
S Kim, J Suk, S Longpre, BY Lin, J Shin, S Welleck, G Neubig, M Lee, ...
EMNLP 2024, 2024
692024
The cot collection: Improving zero-shot and few-shot learning of language models via chain-of-thought fine-tuning
S Kim, SJ Joo, D Kim, J Jang, S Ye, J Shin, M Seo
EMNLP 2023, 2023
682023
Flask: Fine-grained language model evaluation based on alignment skill sets
S Ye, D Kim, S Kim, H Hwang, S Kim, Y Jo, J Thorne, J Kim, M Seo
ICLR 2024, 2023
622023
Exploring the benefits of training expert language models over instruction tuning
J Jang, S Kim, S Ye, D Kim, L Logeswaran, M Lee, K Lee, M Seo
ICML 2023, 2023
592023
Mind the gap! injecting commonsense knowledge for abstractive dialogue summarization
S Kim, SJ Joo, H Chae, C Kim, S Hwang, J Yeo
COLING 2022, 2022
182022
Kmmlu: Measuring massive multitask language understanding in korean
G Son, H Lee, S Kim, S Kim, N Muennighoff, T Choi, C Park, KM Yoo, ...
arXiv preprint arXiv:2402.11548, 2024
172024
Consent in Crisis: The Rapid Decline of the AI Data Commons
S Longpre, R Mahari, A Lee, C Lund, H Oderinwale, W Brannon, ...
NeurIPS 2024, 2024
162024
Aligning to thousands of preferences via system message generalization
S Lee, SH Park, S Kim, M Seo
NeurIPS 2024, 2024
122024
Prometheusvision: Vision-language model as a judge for fine-grained evaluation
S Lee, S Kim, SH Park, G Kim, M Seo
ACL 2024 Findings, 2024
122024
Cotever: Chain of thought prompting annotation toolkit for explanation verification
S Kim, SJ Joo, Y Jang, H Chae, J Yeo
EACL 2023, 2023
82023
Langbridge: Multilingual reasoning without multilingual supervision
D Yoon, J Jang, S Kim, S Kim, S Shafayat, M Seo
ACL 2024, 2024
72024
Self-Explore to Avoid the Pit: Improving the Reasoning Capabilities of Language Models with Fine-grained Rewards
H Hwang, D Kim, S Kim, S Ye, M Seo
EMNLP 2024 Findings, 2024
6*2024
The BiGGen Bench: A Principled Benchmark for Fine-grained Evaluation of Language Models with Language Models
S Kim, J Suk, JY Cho, S Longpre, C Kim, D Yoon, G Son, Y Cho, ...
arXiv preprint arXiv:2406.05761, 2024
5*2024
Language models as compilers: Simulating pseudocode execution improves algorithmic reasoning in language models
H Chae, Y Kim, S Kim, KT Ong, B Kwak, M Kim, S Kim, T Kwon, J Chung, ...
EMNLP 2024, 2024
42024
Multi-Task Inference: Can Large Language Models Follow Multiple Instructions at Once?
G Son, S Baek, S Nam, I Jeong, S Kim
ACL 2024, 2024
32024
Pangea: A fully open multilingual multimodal llm for 39 languages
X Yue, Y Song, A Asai, S Kim, JD Nyandwi, S Khanuja, A Kantharuban, ...
arXiv preprint arXiv:2410.16153, 2024
22024
Better instruction-following through minimum bayes risk
I Wu, P Fernandes, A Bertsch, S Kim, S Pakazad, G Neubig
arXiv preprint arXiv:2410.02902, 2024
22024
MM-Eval: A Multilingual Meta-Evaluation Benchmark for LLM-as-a-Judge and Reward Models
G Son, D Yoon, J Suk, J Aula-Blasco, M Aslan, VT Kim, SB Islam, ...
arXiv preprint arXiv:2410.17578, 2024
12024
ÇöÀç ½Ã½ºÅÛÀÌ ÀÛµ¿µÇÁö ¾Ê½À´Ï´Ù. ³ªÁß¿¡ ´Ù½Ã ½ÃµµÇØ ÁÖ¼¼¿ä.
ÇмúÀÚ·á 1–20