Coté

Coté

Key generative AI terms with concise definitions

I can never remember all of these, so here’s a list of key generative AI terms with concise definitions:

Key Generative AI Terms

Core AI Concepts

  • Inference – The process of an AI model generating output based on input.
  • Training – Teaching an AI model by adjusting its weights using large datasets.
  • LLM (Large Language Model) – A neural network trained on vast text data to generate human-like language responses.
  • Token – A unit of text (word, subword, or character) that a model processes.
  • Weights – Numerical values that determine how much influence different inputs have on a model’s output.
  • Open Weights – AI models whose weight parameters are publicly available for use, modification, and inspection.

Training & Optimization Techniques

  • Distillation – Compressing a large AI model into a smaller one while retaining most of its capabilities.
  • Fine-tuning – Retraining a model on a specific dataset to specialize it for certain tasks.
  • Pretraining – The initial training phase where a model learns general language patterns before fine-tuning.
  • RLHF (Reinforcement Learning from Human Feedback) – A technique where human reviewers rate model responses to refine behavior.

Model Behavior & Characteristics

  • Transformer – The neural network architecture behind modern LLMs, enabling efficient text generation.
  • Self-Attention – A mechanism that helps models understand relationships between words in a sentence.
  • Prompt Engineering – Crafting input text to guide an AI model’s response effectively.
  • Hallucination – When an AI generates incorrect or nonsensical information that sounds plausible.
  • Bias – Systematic errors in an AI model’s outputs due to imbalances in its training data.

Learning & Generalization

  • Zero-shot Learning – When an AI performs a task it wasn’t explicitly trained for by leveraging general knowledge.
  • Few-shot Learning – When an AI adapts to a task using just a few examples provided in the input.
  • Multimodal AI – AI models that process and generate multiple types of data (e.g., text, images, audio).

Examples of Open-Weight Models

  • Meta’s LLaMA models – Research-friendly open-weight LLMs.
  • Mistral & Mixtral – Efficient open-weight models.
  • Falcon & Bloom – Community-driven, open-weight AI models.

From ChatGPT.

@cote@hachyderm.io, @cote@cote.io, @cote, https://proven.lol/a60da7, @cote@social.lol