Skip to content

Information Theory


title: Introduction to Information Theory

Introduction

Information theory quantifies uncertainty and information content. Concepts such as entropy and cross-entropy are deeply entwined with language modeling and loss functions used to train neural networks.

Knowledge Points

  • Entropy
  • Cross-Entropy
  • Kullback-Leibler (KL) Divergence
  • Why information theory matters for language modeling
  • Manual entropy calculation exercise
  • Implementing cross-entropy loss