Skip to content

Latest commit

 

History

History
25 lines (15 loc) · 731 Bytes

section_transformer_llms.md

File metadata and controls

25 lines (15 loc) · 731 Bytes

Transformer

  • New model architecture introduced in Attention is All You Need
  • Understands relationships between words in a document
  • Applies attention weight between any inputs ("self-attention")

­ transformer

­ Source: Attention is All You Need


Large Language Models (LLMs)

  • Transformer trained with huge amounts of data.
  • Neural Network with millions or billions of weights.

Play time

Back to the Notebook.