The Race to Make AI Smaller, Smarter
The New York Times
Oliver Whang, May 30, 2023
The BabyLM Challenge, organized by computer scientists at institutions including Johns Hopkins University and Switzerland's ETH Zurich, is aimed at creating more accessible, intuitive language models, in stark contrast to the race for ever-larger language models undertaken by big tech companies. The goal is to produce a mini-language model using datasets less than one-ten-thousandth the size used by most advanced large language models. As part of the challenge, researchers have been tasked with training language models on about 100 million words, with the winning model to be chosen based on the effectiveness of their generation and understanding of the nuances of language.
No comments:
Post a Comment