You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We propose a new test to measure a text model's multitask accuracy. The testcovers 57 tasks including elementary mathematics, US history, computer science,law, and more. To attain high accuracy on this test, models must possessextensive world knowledge and problem solving ability. We find that while mostrecent models have near random-chance accuracy, the very largest GPT-3 modelimproves over random chance by almost 20 percentage points on average. However,on every one of the 57 tasks, the best models still need substantialimprovements before they can reach expert-level accuracy. Models also havelopsided performance and frequently do not know when they are wrong. Worse,they still have near-random accuracy on some socially important subjects suchas morality and law. By comprehensively evaluating the breadth and depth of amodel's academic and professional understanding, our test can be used toanalyze models across many tasks and to identify important shortcomings.
AkihikoWatanabe
changed the title
a
Measuring Massive Multitask Language Understanding, Dan Hendrycks+, N/A, arXiv'20
Jul 26, 2023
AkihikoWatanabe
changed the title
Measuring Massive Multitask Language Understanding, Dan Hendrycks+, N/A, arXiv'20
Measuring Massive Multitask Language Understanding, Dan Hendrycks+, N/A, ICLR'21
Jul 26, 2023
URL
Affiliations
Abstract
Translation (by gpt-3.5-turbo)
Summary (by gpt-3.5-turbo)
The text was updated successfully, but these errors were encountered: