Software Engineering KB

Home

❯

09 Machine Learning and AI

❯

01 Deep Learning

❯

01 Concept

❯

Layer Normalization

Layer Normalization

Feb 10, 20261 min read

  • deep-learning
  • normalization
  • layer-norm

Layer Normalization

← Back to Neural Network Fundamentals

Normalizes across the feature dimension for each individual example (rather than across the batch). Preferred in Transformers and RNNs because it does not depend on batch size and works well with variable-length sequences.

Related

  • Batch Normalization (normalizes across batch instead)
  • Transformers (standard normalization for Transformers)

deep-learning normalization layer-norm


Graph View

  • Layer Normalization
  • Related

Backlinks

  • Neural Network Fundamentals
  • Batch Normalization
  • Multi-Head Attention

Created with Quartz v4.5.2 © 2026

  • GitHub