BERT and GPT-2 are fairly old models...
Science
Studies, research findings, and interesting tidbits from the ever-expanding scientific world.
Subcommunities on Beehaw:
Be sure to also check out these other Fediverse science communities:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
The first preprint was submitted 7 apr 2022. It's quite common that a scientific paper in a peer reviewed journal takes that long to be published, particularly if the reviewers ask for corrections (the final version here is the third version).
Not mentioning that research leading to an article needs time, and writing a scientific paper needs time too.
Good point. It just seems odd that the Columbia article calls them "current language models," whereas the coauthor of the paper is quoted as only calling them "the best models [the authors of the paper] have studied."
Well, yes. AI models don't extract meaning. They parrot statistically likely responses based on words used. They had to research that?