Korean Language Understanding Evaluation
Korean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language understanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible to anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain unambiguous annotations for all datasets. Furthermore, we build an evaluation system and carefully choose evaluations metrics for every task, thus establishing fair comparison across Korean language models.
KLUE benchmark is composed of 8 tasks:
Variants: KLUE
This dataset is used in 1 benchmark:
No recent benchmark submissions available for this dataset.
No papers with results on this dataset found.