You are viewing a single comment's thread from:

RE: AI-Summaries Weekly Report #1

in #aisummaries23 days ago

Part 3/8:

An LLM functions as a mathematical entity that predicts the subsequent word for any text given. Rather than delivering a single definitive word, these models generate probabilistic predictions for all potential next words. Building a chatbot involves inputting a scripted interaction alongside user input while prompting the model to compute the next word iteratively. This method produces outputs that reflect a more natural conversation style, especially when it randomly selects from less likely options.

The Training Process

To create an LLM, massive datasets—most often sourced from the internet—are processed. For instance, the training dataset for GPT-3 would take over 2,600 years for an average human to read continuously. Modern models train on exponentially more data.