Artificial intelligence has traditionally advanced through automatic accuracy tests in tasks meant to approximate human knowledge.
Carefully crafted benchmark tests such as The General Language Understanding Evaluation benchmark (GLUE), the Massive Multitask Language Understanding data set (MMLU), and “Humanity’s Last Exam,” have used large arrays of questions to score how well a large language model knows about a lot of things.
However, those tests are
Article Source
https://www.zdnet.com/article/reasoning-ai-models-are-overwhelming-the-benchmark-tests-its-time-for-human-evaluation/