Contents
Neural network trained on large text corpora to generate and understand language
This article hasn't been written yet
This is a stub — a placeholder for an article that is referenced by other articles but hasn't been fully written. Contribute this article
A large language model (LLM) is a neural network with billions of parameters trained on massive text datasets to predict and generate natural language. Modern LLMs are based on the transformer architecture introduced in 2017. Notable examples include the GPT series (OpenAI), Claude (Anthropic), Gemini (Google), and LLaMA (Meta).
A large language model (LLM) is a neural network with billions of parameters trained on massive text datasets to predict and generate natural language. Modern LLMs are based on the transformer architecture introduced in 2017. Notable examples include the GPT series (OpenAI), Claude (Anthropic), Gemini (Google), and LLaMA (Meta).