32.3068, Fun Fact: What is GPT-3 all about?

The LINGUIST List linguist at listserv.linguistlist.org
Wed Sep 29 19:31:32 UTC 2021


LINGUIST List: Vol-32-3068. Wed Sep 29 2021. ISSN: 1069 - 4875.

Subject: 32.3068, Fun Fact: What is GPT-3 all about?

Moderator: Malgorzata E. Cavar (linguist at linguistlist.org)
Student Moderator: Jeremy Coburn, Lauren Perkins
Managing Editor: Becca Morris
Team: Helen Aristar-Dry, Everett Green, Sarah Robinson, Nils Hjortnaes, Joshua Sims, Billy Dickson
Jobs: jobs at linguistlist.org | Conferences: callconf at linguistlist.org | Pubs: pubs at linguistlist.org

Homepage: http://linguistlist.org

Please support the LL editors and operation with a donation at:
           https://funddrive.linguistlist.org/donate/

Editor for this issue: Joshua Sims <joshua at linguistlist.org>
================================================================


Date: Wed, 29 Sep 2021 15:29:23
From: LINGUIST List [linguist at linguistlist.org]
Subject: Fun Fact: What is GPT-3 all about?

 Hello fellow linguists! Billy here—today’s fun fact is about GPT-3! My goal is to pull back the curtain on this mystery (or at least give it a slight tug).

What is this GPT-3? GPT-3 stands for third generation Generative Pre-trained Transformer. I won’t blame you if at first the name leads you to envision a Chomskyan Optimus Prime visiting the local gym.

Let’s break this down.

Generative here means that some text is inputted, and the model will generate some output. The output is based on probability predictions for how likely a sequence of words is to appear. For example, “I like eating apples” is more probable than “I like eating cars” (unless perhaps we are talking about Megatron).

Pre-trained means that this model has been exposed previously to large amounts of text (almost 45 terabytes!) in order to refine its calculations (i.e. adjusting weights within the neural network). This model is unsupervised, which means that it looks at large amounts of data and forms its own representation of patterns without human intervention.

Transformer is the neural-network architecture of the model. This applies special techniques (such as attention and self-attention) to optimize these calculations.

But what about linguistics?

A linguist understanding the rich complexity of language may not be satisfied with representing such intricacy with linear algebra. Hybrid approaches (coined as neuro-symbolic) use these neural network architectures along with traditional symbolic approaches (human written rules) to incorporate deeper semantic reasoning over language.

What do you think? What do you make of GPT-3? How can the complexity of language be captured in a computer? We would love to hear your opinions, no matter how technical your background!

Thanks for reading,

Billy

Linguist List Student Moderator

PS: If you would to like to dive deeper into the technical aspects of GPT-3, I highly recommend checking out Jay Alammar’s detailed explanation here: 
https://jalammar.github.io/how-gpt3-works-visualizations-animations/



------------------------------------------------------------------------------

***************************    LINGUIST List Support    ***************************
 The 2020 Fund Drive is under way! Please visit https://funddrive.linguistlist.org
  to find out how to donate and check how your university, country or discipline
     ranks in the fund drive challenges. Or go directly to the donation site:
                   https://crowdfunding.iu.edu/the-linguist-list

                        Let's make this a short fund drive!
                Please feel free to share the link to our campaign:
                    https://funddrive.linguistlist.org/donate/
 


----------------------------------------------------------
LINGUIST List: Vol-32-3068	
----------------------------------------------------------






More information about the LINGUIST mailing list