When was GPT-3 published by OpenAI?
GPT-3 was published on the 28th of May 2020. A group of thirty-one engineers and researchers at OpenAI released an arXiv preprint describing the third-generation language model on that date.
Short answers, pulled from the story.
GPT-3 was published on the 28th of May 2020. A group of thirty-one engineers and researchers at OpenAI released an arXiv preprint describing the third-generation language model on that date.
GPT-3 contains 175 billion parameters with a context window of 2048 tokens. Each parameter required 16-bit precision consuming 350GB of storage space alone while allowing the system to process substantial blocks of text before losing track of earlier information.
Sixty percent of the weighted pre-training dataset came from Common Crawl containing 410 billion byte-pair-encoded tokens. The remaining data included WebText2, Books1, Books2, and Wikipedia which together provided hundreds of billions of words scraped from sixty million domains over twelve years.
Microsoft announced on the 22nd of September 2020 that it had licensed GPT-3 exclusively for its own products and services. This multi-billion dollar investment permits OpenAI to offer a public-facing API while only Microsoft gains access to the underlying source code.
Noam Chomsky expressed skepticism stating that GPT-3 works just as well for impossible languages as for actual languages. He argued the system is refuted by normal scientific criteria if intended as a true language model because it tells us nothing about language or cognition generally.