next BIG future

next BIG future

Share this post

next BIG future
next BIG future
Starting the Era of 1-bit LLMs – With Microsoft Research
Copy link
Facebook
Email
Notes
More

Starting the Era of 1-bit LLMs – With Microsoft Research

by Brian Wang

NextBigFuture's avatar
NextBigFuture
Mar 28, 2024
∙ Paid

Share this post

next BIG future
next BIG future
Starting the Era of 1-bit LLMs – With Microsoft Research
Copy link
Facebook
Email
Notes
More
Share

BitNet is paving the way for a new era of 1-bit Large Language Models (LLMs). In this work, we introduce a 1-bit LLM variant, namely BitNet b1.58, in which every single parameter (or weight) of the LLM is ternary {-1, 0, 1}.</a> It matches the full-precision (i.e., FP16 or BF16) Transformer LLM with the same model size and training tokens in terms of bo…

Keep reading with a 7-day free trial

Subscribe to next BIG future to keep reading this post and get 7 days of free access to the full post archives.

Already a paid subscriber? Sign in
© 2025 Nextbigfuture
Privacy ∙ Terms ∙ Collection notice
Start writingGet the app
Substack is the home for great culture

Share

Copy link
Facebook
Email
Notes
More