btop reports some gpu, network and disk information that I don't think shows up in htop, feels a bit more comprehensive maybe? Both are fine, but I too use btop, it's nice.
Random trivia: I think btop has been rewritten like 3-5 times now? It's sort of an inside joke to the point that someone suggested another rewrite from C++ to Rust ( https://github.com/aristocratos/btop/issues/5 ). I guess the guy just likes writing system monitoring console apps.
It's not uncommon on sensitive stories like this for the government to loop-in journalists ahead of time so they can pull together background and research with an agreed-upon embargo until some point in the future.
This wasn't the US government telling the newspaper they couldn't report on a story they had uncovered from their own investigation.
There's quantization which basically compresses the model to use a smaller data type for each weight. Reduces memory requirements by half or even more.
There's also airllm which loads a part of the model into RAM, runs those calculations, unloads that part, loads the next part, etc... It's a nice option but the performance of all that loading/unloading is never going to be great, especially on a huge model like llama 405b
Then there are some neat projects to distribute models across multiple computers like exo and petals. They're more targeted at a p2p-style random collection of computers. I've run petals in a small cluster and it works reasonably well.
When they’re not recording your desktop in an unencrypted database for AI, boot-looping your computer with bad patches or showing ads in your start menu, they’re disabling your account for calling family to see if they’re still alive. Damn.
Taking ollama for instance, either the whole model runs in vram and compute is done on the gpu, or it runs in system ram and compute is done on the cpu. Running models on CPU is horribly slow. You won’t want to do it for large models
LM studio and others allow you to run part of the model on GPU and part on CPU, splitting memory requirements but still pretty slow.
Even the smaller 7B parameter models run pretty slow in CPU and the huge models are orders of magnitude slower
So technically more system ram will let you run some larger models but you will quickly figure out you just don’t want to do it.
Boeing made $76B in revenue in 2023. This is slightly more than 1 day's revenue for them ($210M / day) or a bit more than 10 days profit for them ($21M / day). They will keep doing what they're doing, but increase their spending on a PR campaign to improve their public image.
The recent comments are from the announcement of the ladybird browser project which is forked from some browser code from Serenity OS, I guess people are digging into who wrote the code.
Not arguing that the new comments on the PR are good/bad or anything, just a bit of context.
Also, the few points others are talking about needing others, there's a group-finder and I'd say most people running those raids in group finder groups don't talk at all, so you can just pretend they're NPCs if you want.
btop reports some gpu, network and disk information that I don't think shows up in htop, feels a bit more comprehensive maybe? Both are fine, but I too use btop, it's nice.
Random trivia: I think btop has been rewritten like 3-5 times now? It's sort of an inside joke to the point that someone suggested another rewrite from C++ to Rust ( https://github.com/aristocratos/btop/issues/5 ). I guess the guy just likes writing system monitoring console apps.