/* ---- Google Analytics Code Below */

Sunday, August 29, 2021

What have Language Models Learned?


What Have Language Models Learned?

Large language models are making it possible for computers to write stories, program a website and turn captions into images.

One of the first of these models, BERT, is trained by taking sentences, splitting them into individual words, randomly hiding some of them, and predicting what the hidden words are. After doing this millions of times, BERT has “read” enough Shakespeare to predict how this phrase usually ends:

No comments: