Skip Navigation

InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)P
Posts
0
Comments
172
Joined
3 yr. ago

  • Easiest shorting money I ever made.

  • It's not uncommon on sensitive stories like this for the government to loop-in journalists ahead of time so they can pull together background and research with an agreed-upon embargo until some point in the future.

    This wasn't the US government telling the newspaper they couldn't report on a story they had uncovered from their own investigation.

  • I guess this solves part of the mystery about why the French rioted when they raised the retirement age last year

  • There's quantization which basically compresses the model to use a smaller data type for each weight. Reduces memory requirements by half or even more.

    There's also airllm which loads a part of the model into RAM, runs those calculations, unloads that part, loads the next part, etc... It's a nice option but the performance of all that loading/unloading is never going to be great, especially on a huge model like llama 405b

    Then there are some neat projects to distribute models across multiple computers like exo and petals. They're more targeted at a p2p-style random collection of computers. I've run petals in a small cluster and it works reasonably well.

  • MAWP - Archer

  • When they’re not recording your desktop in an unencrypted database for AI, boot-looping your computer with bad patches or showing ads in your start menu, they’re disabling your account for calling family to see if they’re still alive. Damn.

  • Taking ollama for instance, either the whole model runs in vram and compute is done on the gpu, or it runs in system ram and compute is done on the cpu. Running models on CPU is horribly slow. You won’t want to do it for large models

    LM studio and others allow you to run part of the model on GPU and part on CPU, splitting memory requirements but still pretty slow.

    Even the smaller 7B parameter models run pretty slow in CPU and the huge models are orders of magnitude slower

    So technically more system ram will let you run some larger models but you will quickly figure out you just don’t want to do it.

  • Boeing made $76B in revenue in 2023. This is slightly more than 1 day's revenue for them ($210M / day) or a bit more than 10 days profit for them ($21M / day). They will keep doing what they're doing, but increase their spending on a PR campaign to improve their public image.

  • Respect, but...

  • FWIW they didn't merge it, they closed the PR without merging, link to line that still exists on master.

    The recent comments are from the announcement of the ladybird browser project which is forked from some browser code from Serenity OS, I guess people are digging into who wrote the code.

    Not arguing that the new comments on the PR are good/bad or anything, just a bit of context.

  • I’ve been tempted to try and install plasma mobile on a tablet.

  • Why no arch install?

  • Also, the few points others are talking about needing others, there's a group-finder and I'd say most people running those raids in group finder groups don't talk at all, so you can just pretend they're NPCs if you want.

  • Been 100% linux for like 6-9 months now, these stories make me thankful for finally making the switch.

    I've tried to make the switch 3-4 times in the past and was stopped by 2 main things:

    • Drivers / Laptops were tough to get set up
    • Gaming

    The experience was so much better this time and I really have no regrets. I don't imagine I'll ever run Windows again outside of a VM

  • Elon “Nick Cannon” Musk

  • Seems like arch gets KDE into stable within a couple days of release generally. Or there’s the kde-unstable repo that already has it

  • How does one train a cat to make waffles? Asking for a friend