@fabrice it's still like 5-10 exabytes of extra bandwidth, and it's not going to be diff-friendly. They developed a new dedicated binary patcher to shrink Chrome updates (Zucchini) fairly recently so it can't be too cheap for them.
Discussion
@gabrielesvelto it's also curious that a 4GB model that can run on consumer hardware is useful enough to force onto every user of Chrome, and somehow the AI companies expect everyone to pay them for these services.
@gundersen indeed. But there is nothing rational about this market. It's all C-suite FOMO, vibes, smoke and mirrors.
@gabrielesvelto I think you are absolutely correct!
@gabrielesvelto This download is going to happen every time they update the blob isn't it?
@pwloftus yes. It's basically a bunch of sparse matrixes, that get recomputed every time you re-train the model. It's very unlikely that it will be amenable to binary patching or incremental updates.
@gabrielesvelto I bet it's actually cheap for Google. They have network caches (or used to at least for YT) very close to the ISPs, and this kind of payload is very cache friendly.