Skip Navigation

Posts
0
Comments
413
Joined
3 yr. ago

Basically a deer with a human face. Despite probably being some sort of magical nature spirit, his interests are primarily in technology and politics and science fiction.

Spent many years on Reddit and is now exploring new vistas in social media.

  • Those recent failures only come across as cracks for people who see AI as magic in the first place. What they're really cracks in is people's misperceptions about what AI can do.

    Recent AI advances are still amazing and world-changing. People have been spoiled by science fiction, though, and are disappointed that it's not the person-in-a-robot-body kind of AI that they imagined they were being promised. Turns out we don't need to jump straight to that level to still get dramatic changes to society and the economy out of it.

    I get strong "everything is amazing and nobody is happy" vibes from this sort of thing.

  • It's the "make some people non-white" kludge that's the specific problem being discussed here.

    The training data skewing white is a different problem, but IMO not as big of one. The solution is simple, as I've discovered over many months of using local image generators. Let the user specify what exactly they want.

  • It's not the training data that's the problem here.

  • The term "AI" has a much broader meaning and use than the sci-fi "thinking machine" that people are interpeting it as. The term has been in use by scientists for many decades already and these generative image programs and LLMs definitely fit within it.

    You are likely thinking of AGI, or artificial general intelligence. We don't have those yet, but these things aren't intended to be AGI so that's to be expected.

  • And even if it was Google, these companies aren't magic. Once there's a proof of concept out there that something like this can be done other companies will dump resources into catching up with it. Cue the famous "we have no moat" memo.

  • Well, I hope my answer clarifies it. You can't prevent LLMs from being trained on your public posts.

  • We're sick of closed walled-garden monoliths like Reddit! Let's move to an open federated protocol where anyone can participate and the APIs can't be locked down!

    ...wait, not like that!

    Yeah. This is what you signed up for when you joined the Fediverse, the ActivityPub protocol broadcasts your content to any other servers that ask for it. And just generally, that's how the Internet works. You're putting up a public billboard and expecting to be able to control who gets to look at it. That's not going to work. Even robots.txt is just a gentleman's agreement, it's not enforceable.

    If you really want to prevent AI from training on your content with any degree of certainty you're probably looking for a private forum of some kind that's run by someone you trust.

  • You'd need to gut the car completely and rebuild it, it would be more work than starting from scratch.

  • Negative examples are just as useful to train on as positive ones.

  • I've lost track, is AI a good thing today or a bad thing?

  • It's currently 2024, so we're still okay. :)

  • The Eugenics War ran from 1992 to 1996, so I think we're probably okay.

  • They're rolling it out gradually, as is customary for routine updates.

    I'm not sure why this is worthy of a headline, frankly. This is how Microsoft typically does these things. I guess it's the "...with AI involved somehow!" Bit in the title that makes it interesting? I expect that's going to get old fairly quickly.

  • Sora's capabilities aren't really relevant to the competition if OpenAI isn't allowing it to be used, though. All it does is let the actual competitors know what's possible if they try, which can make it easier to get investment.

  • Indeed, the level of obsession some people have with Elon Musk is kind of ridiculous.

  • Writing code to do math is different from actually doing the math. I can easily write "x = 8982.2 / 98984", but ask me what value x actually has and I'll need to do a lot more work and quite probably get it wrong.

    This is why one of the common improvements for LLM execution frameworks these days is to give them access to external tools. Essentially, give it access to a calculator.

  • Exactly. Article looked fine to me, if it was AI-written then it did a good job.

  • It's not exactly training, but Google just recently previewed a LLM with a million-token context that can do effectively the same thing. One of the tests they did was to put a dictionary for a very obscure language (only 200 speakers worldwide) into the context, knowing that nothing about that language was in its original training data, and the LLM was able to translate it fluently.

    OpenAI has already said they’re not making that publicly available for now

    This just means that OpenAI is voluntarily ceding the field to more ambitious companies.

  • How dare they provide a useful tool like this, those bastards.

  • Why do you think so, and why does it matter?