From e2b93a9bbd5ab015c0a1c9c714f9c11a8881de30 Mon Sep 17 00:00:00 2001 From: osmarks Date: Thu, 5 Sep 2024 20:05:25 +0000 Subject: [PATCH] =?UTF-8?q?Edit=20=E2=80=98large=5Flanguage=5Fmodel?= =?UTF-8?q?=E2=80=99?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- large_language_model.myco | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/large_language_model.myco b/large_language_model.myco index 59f564a..fb73502 100644 --- a/large_language_model.myco +++ b/large_language_model.myco @@ -1,4 +1,4 @@ -A large language model is a [[neural net]] model of [[language]] which is [[large]], usually in the sense of parameter count or total [[compute]], making them [[good]] at text prediction by [[scaling laws]]. Usually these are [[autoregressive]] and pretrained on general text data with a next token prediction [[loss function]], though this is not necessarily required. The largest large LLMs known are around 2 trillion parameters, though the smallest LLM is not known. +A large language model is a [[neural net]] model of [[language]] which is [[large]], usually in the sense of parameter count or total [[compute]], making it [[good]] at text prediction by [[scaling laws]]. Usually these are [[autoregressive]] and pretrained on general text data with a next token prediction [[loss function]], though this is not necessarily required. The largest large LLMs known are around 2 trillion parameters, though the smallest LLM is not known. == History