enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. List of large language models - Wikipedia

    en.wikipedia.org/wiki/List_of_large_language_models

    A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. This page lists notable large language models.

  3. Large language model - Wikipedia

    en.wikipedia.org/wiki/Large_language_model

    A large language model (LLM) is a type of machine learning model designed for natural language processing tasks such as language generation. LLMs are language models with many parameters, and are trained with self-supervised learning on a vast amount of text. The largest and most capable LLMs are generative pretrained transformers (GPTs).

  4. BERT (language model) - Wikipedia

    en.wikipedia.org/wiki/BERT_(language_model)

    It is notable for its dramatic improvement over previous state-of-the-art models, and as an early example of a large language model. As of 2020, BERT is a ubiquitous baseline in natural language processing (NLP) experiments. [3] BERT is trained by masked token prediction and next sentence prediction.

  5. T5 (language model) - Wikipedia

    en.wikipedia.org/wiki/T5_(language_model)

    T5 (Text-to-Text Transfer Transformer) is a series of large language models developed by Google AI introduced in 2019. [1] [2] Like the original Transformer model, [3] T5 models are encoder-decoder Transformers, where the encoder processes the input text, and the decoder generates the output text.

  6. Llama (language model) - Wikipedia

    en.wikipedia.org/wiki/Llama_(language_model)

    Llama (Large Language Model Meta AI, formerly stylized as LLaMA) is a family of large language models (LLMs) released by Meta AI starting in February 2023. [2] [3] The latest version is Llama 3.3, released in December 2024. [4] Llama models are trained at different parameter sizes, ranging between 1B and 405B. [5]

  7. MMLU - Wikipedia

    en.wikipedia.org/wiki/MMLU

    The MMLU consists of about 16,000 multiple-choice questions spanning 57 academic subjects including mathematics, philosophy, law, and medicine. It is one of the most commonly used benchmarks for comparing the capabilities of large language models, with over 100 million downloads as of July 2024. [1] [2]

  8. Chinchilla (language model) - Wikipedia

    en.wikipedia.org/wiki/Chinchilla_(language_model)

    It is named "chinchilla" because it is a further development over a previous model family named Gopher. Both model families were trained in order to investigate the scaling laws of large language models. [2] It claimed to outperform GPT-3. It considerably simplifies downstream utilization because it requires much less computer power for ...

  9. GPT-J - Wikipedia

    en.wikipedia.org/wiki/GPT-J

    GPT-J or GPT-J-6B is an open-source large language model (LLM) developed by EleutherAI in 2021. [1] As the name suggests, it is a generative pre-trained transformer model designed to produce human-like text that continues from a prompt.