from Hacker News

Brain Floating Point (Bfloat16)

by ml_basics on 5/28/23, 6:27 PM with 1 comments

  • by ml_basics on 5/28/23, 6:31 PM

    bfloat16 is probably familiar only to ML practitioners; it's a reduced precision floating point format designed for ML models. I was surprised to learn that the "b" stands for "brain", as in the team at google that developed it along with many other advances in machine learning.