Skip Navigation

Jack Riddle[Any/All]

@ jackr @lemmy.dbzer0.com

Posts
2
Comments
202
Joined
10 mo. ago

  • what text are you reading that has a 0% error rate?

    as I said, the text has a 0% error rate about the contents of the text, which is what the LLM is summarising, and to which it adds it's own error rate. Then you read that and add your error rate.

    the question is can we make a system that has an error rate that is close to or lower than a person's

    can we???

    could you read and summarize 75 novels with a 0% error rate?

    why… would I want that? I read novels because I like reading novels? I also think that on summaries LLMs are especially bad, since there is no distinction between "important" and "unimportant" in the architecture. The point of a summary is to only get the important points, so it clashes.

    provide a page reference to all of the passages written in iambic pentameter?

    no LLM can do this. LLMs are notoriously bad at doing any analysis of this kind of style element because of their architecture. why would you pick this example

    Meanwhile an LLM could produce a summary, with citations generated and tracked by non-AI systems, with an error rate comparable to a human (assuming the human was given a few months to work on the problem) in seconds.

    I still have not seen any evidence for this, and it still does not adress the point that the summary would be pretty much unreadable

  • The study of this in academia

    you are linking to an arxiv preprint. I do not know these researchers. there is nothing that indicates to me that this source is any more credible than a blog post.

    has found that LLM hallucination rate can be dropped to almost nothing

    where? It doesn't seem to be in this preprint, which is mostly a history of RAG and mentions hallucinations only as a problem affecting certain types of RAG more than other types. It makes some relative claims about accuracy that suggest including irrelevant data might make models more accurate. It doesn't mention anything about “hallucination rate being dropped to almost nothing”.

    (less than a human)

    you know what has a 0% hallucination rate about the contents of a text? the text

    You can see in the images I posted that it both answered the question and also correctly cited the source which was the entire point of contention.

    this is anecdotal evidence, and also not the only point of contention. Another point was, for example, that ai text is horrible to read. I don't think RAG(or any other tacked-on tool they've been trying for the past few years) fixes that.

  • see, the problem is that I am not going to be reading that text because I know it is unreliable and ai text makes my eyes glaze over, so I will be clicking on all those links until I find something that is reliable. On a search engine I can just click through every link or refine my search with something like site:reddit.com site:wikipedia.org or format:pdf or something similar. With a chatbot, I need to write out the entire question, look at the four or so links it provided and then reprompt it if it doesn't contain what I'm looking for. I also get a limited amount of searches per day because I am not paying for a chatbot subscription. This is completely pointless to me.

  • good basilisk save us all. so they built a script for their chatbot that allows it to purchase more chatbots? seems like a great use of money. Also, what's with the insanely placed emdashes? did conway write this for him or has his brain been rotted so much that he writes like an LLM? large parts seem human-written at least…

  • the screenshot is very very clearly LLM generated right? This is so insanely stupid

  • hate it when you are working on a major featue for the next release but tim keeps continvoucly morging

  • so I fail to see why I should be using an LLM at all then. If I am going to the webpages anyway, why shouldn't I just use startpage/searx/yacy/whatever?

  • so we are using the "regular search which has always given you garbage" and taking that garbage automatically to get summarised by the hallucinator and we are supposed to trust the output somehow?

  • unfortunately, julia has been adding "agentic code" to their codebase for a while now.

  • bluemonday1984@awful.systems update: quokka reached out to me and apparently you had been banned on another instance for report abuse and that ban had synchronised to quokk.au. You should be unbanned now which means the next stubsack should federate again.

    E: I do not know how tags work E2: why does that format to a mailto link?

  • OT: I tried switching away from my instance to quokk.au because it becomes harder and harder to justify being on a pro-ai instance, but it seems that your posts stopped federating recently.

    Which makes that a bit harder to do. Any clue why?

  • Locked

    feddit.org's Zionist bar problem: community ban(s) vote

    Jump
  • The bot shows the user flair of /0 users. The pirate icon means you said you had an interest in piracy during signup

    1. microblog site ✅
    2. funny image (meme) ✅

    my checks are all green boss

  • what about xxy? What about people with female sex characteristics but xy chromosomes? This seems stupid.

  • see, the problem with this is that no person on earth calls it "oat drink" or whatever. Oat milk is the accepted term. The etymology is only there to highlight how ridiculous the entire thing is.

  • just don't ask for gender? let the player either design a sprite or pick one and pick pronouns or just refer to the player with they/them. Sex is only a fact insofar as the collection of bodily features we commonly refer to as sex exist. the reference to it is still a social choice, and we could've just as easily picked another group of features, or none at all.

  • it does smell really good tho

  • yes.

  • Locked

    feddit.org's Zionist bar problem: community ban(s) vote

    Jump
  • no I mean it was a sort of instant reaction, the site does look cool and I will check it out later. It just took me by surprise

  • Locked

    feddit.org's Zionist bar problem: community ban(s) vote

    Jump
  • the bot replies with your user flair. This means you have mentioned you were an anarchist when signing up

  • SneerClub @awful.systems

    The solution to AI lying is… wikipedia but worse, apparently

    www.greaterwrong.com /posts/yFTMGKh9Muqpdtrmb/wikipedia-but-written-by-ais
  • SneerClub @awful.systems

    Found a pretty good blog post on our friends in the wild

    www.antipope.org /charlie/blog-static/2023/11/dont-create-the-torment-nexus.html