Foundation Model

Foundation Models are a form of General Intelligence, where the models can generate output from the human instruction. It uses Self-Supervised Learning to learn pattern in the input. It is different from other ML architectures in that way.

It can learn the next word probability from the given input and previously generated output. It can generate sharper image or even new image, given a blurry image or a prompt to generate image.

They are called foundation model because they are the foundation of different ML models for different tasks, i.e., BERT model can be used for multiple downstream tasks, even they are trained for those.

Use Cases:

  1. Language Processing
  2. Visual Comprehension
  3. Code Generation

Examples:

  1. BERT
  2. GPT
  3. BLOOM
  4. Cohere

References


Related Notes