註冊並分享邀請連結,可獲得影片播放與邀請獎勵。

Alexander Whedon
@alex_whedon
Building better algorithms. Co-Founder at @subquadratic
加入 November 2024
59 正在關注    24.5K 粉絲
Introducing SubQ - a major breakthrough in LLM intelligence. It is the first model built on a fully sub-quadratic sparse-attention architecture (SSA), And the first frontier model with a 12 million token context window which is: - 52x faster than FlashAttention at 1MM tokens - Less than 5% the cost of Opus Transformer-based LLMs waste compute by processing every possible relationship between words (standard attention). Only a small fraction actually matter. @subquadratic finds and focuses only on the ones that do. That's nearly 1,000x less compute and a new way for LLMs to scale.
顯示更多
0
1.5K
23.1K
2.9K
轉發到社區