Homie been smokin’ them data science rocks, it seems.
Literally made an account on this instance just to let them know I think they’re fucking dense, but I decided they’re not even worth interacting with personally.
Huh? The whole point of this emerging scientific debate is that AI use might be proportionally unsafe, i. e. it might be a risk factor causing and/or exacerbating psychosis. Now sure this is still just a hypothesis and it’s too early to make definite epidemiological statements, but it’s just as wrong to blankly state that AI is “still just as safe”.
Alright, but the point is that the “X level of safety” AI is at might be a dangerous level in the first place. I don’t think anybody is arguing that AI got more dangerous as a psychosis risk factor over the past year or so, they’re arguing that AI was a risk factor to begin with, and with increased AI use more evidence of this turns up. So you saying that the inherent risk of AI hasn’t changed is kind of a moot point because that’s not what the debate is about.
Also notice that I clearly said it’s too early to tell one way or the other, so there’s no reason to malign me as uncritical.
I can agree with that. (As an aside, I think scientific findings are almost always exaggerated like this in popular journalism.)
I’d say the long and short of it is that we simply don’t (and can’t) know yet. But I think more research on possible links between AI and psychotic delusions is definitely useful, because I find the idea of a connection plausible.
Says the person calling people “fucks agreeing with this shit take”, and “brain-dead AI haters” and “less-critical readers” and just in this thread alone. Who knows what else I’d find in looking in your full posting history.
Not a very convincing act, even for a clank-fucker.
Removed by mod
Do you think statisticians aren’t well aware of this?
Removed by mod
If the statisticians involved in this case study are anywhere close to as unhinged as you are then it’s no wonder they got those results lol
Homie been smokin’ them data science rocks, it seems.
Literally made an account on this instance just to let them know I think they’re fucking dense, but I decided they’re not even worth interacting with personally.
Huh? The whole point of this emerging scientific debate is that AI use might be proportionally unsafe, i. e. it might be a risk factor causing and/or exacerbating psychosis. Now sure this is still just a hypothesis and it’s too early to make definite epidemiological statements, but it’s just as wrong to blankly state that AI is “still just as safe”.
Removed by mod
Alright, but the point is that the “X level of safety” AI is at might be a dangerous level in the first place. I don’t think anybody is arguing that AI got more dangerous as a psychosis risk factor over the past year or so, they’re arguing that AI was a risk factor to begin with, and with increased AI use more evidence of this turns up. So you saying that the inherent risk of AI hasn’t changed is kind of a moot point because that’s not what the debate is about.
Also notice that I clearly said it’s too early to tell one way or the other, so there’s no reason to malign me as uncritical.
Removed by mod
I can agree with that. (As an aside, I think scientific findings are almost always exaggerated like this in popular journalism.)
I’d say the long and short of it is that we simply don’t (and can’t) know yet. But I think more research on possible links between AI and psychotic delusions is definitely useful, because I find the idea of a connection plausible.
How do LLM interactions compare to… Kinder eggs or lawn darts in terms of safety?
Kinder eggs are incredibly safe. Lawn darts … less so.
Says the person calling people “fucks agreeing with this shit take”, and “brain-dead AI haters” and “less-critical readers” and just in this thread alone. Who knows what else I’d find in looking in your full posting history.
Not a very convincing act, even for a clank-fucker.
Removed by mod