he/him
Short answer is no, but you can buy a card holder that integrates into the case: https://shop.fairphone.com/shop/fairphone-6-card-holder-1895?category=5&color=188
I thought I'd miss not having contactless phone payments but actually the card holder thing is fine.
He was redirected to Sainsbury's, which apologised and offered him a £75 shopping voucher.
Get absolutely fucked Sainsbury's, what a joke. "We're sorry we called you a criminal and chucked you out of our shop, here's some fake toy money you can only spend at the same place we humiliated you in."
I know I'm a weirdo in this regard but I would voluntarily get a digital ID from the government. I can see how it would be beneficial in accessing public services but also private services too. I use the Government Gateway account for a wide range of things already including state benefits and self assessment, and it would be great if I could use the same account to prove my identity when signing up for bank accounts.
Right now, to prove my identity to my bank, I have to take a photo of my driving licence, and a video of my face. Then, in probably the most humiliating and degrading thing I have ever been asked to do by my phone, I have to SMILE AT THE CAMERA in order to trigger a selfie to be taken.
My driving licence, video and selfie-with-enforced-smile is then sent to a company that uses "AI" to determine my identity. Now, I know that they do not always use AI to do this, because my wife used to work for this company, and she spent all day looking at people's photos and driving licences to see if they matched. They were able to provide 24/7 service by having large offices in 3 different time zones, so that an 8-10 hour work day would span the full 24 hour day.
That entire process for verifying my ID with my bank is way more dystopian, alienating and humiliating than signing up for a digital ID with the government. So while I don't support mandatory ID, I do support a government issued ID that was widely recognised by banks, workplaces, estate agents, and other private institutions that are necessary for modern life.
This is a good summary. At this point I am too deeply invested in to NextCloud to switch to a different thing, as I've switched my whole family off OneDrive now and I just cba to go through that again. I can handle it being dogshit and I've got used to it's bugs - a form of stockholm syndrome. I suspect a lot of people are similar to me - we use NextCloud because it's the biggest name and has been around forever, not because it's what we want.
Anyway, performance is clearly a problem, and has been since I started on OwnCloud 10 years ago. I wish the devs would do something to improve it but again, having used it for 10 years, I know that they won't. When it finally blows up I'll move to something else I guess.
Until verification is complete, access to [social features] will be limited to friends only
don't threaten me with a good time
I'm willing to bet that, since their sister got a smart meter, everyone's bills have doubled.
I'm totally with you on not strapping batteries to my feet. That feels like a good rule to live by for anyone.
It doesn't actually give any examples of removed posts or screenshots of the reasons why? Surely it can't just be because a town name happens to contain "lsd" in the middle of it?
Autofill is total shit I agree, but I've been copying and pasting from my password manager for over a decade and it's been fine. I get that autofill would be much less friction, but I really don't mind copying and pasting.
It's funny because this means that Aldi has better quality food than Tesco. That's not something I would have thought before. Before, I would have just assumed Aldi was cheaper, but now I assume that Aldi is cheaper AND better.
I looked into spray foam insulation but not only were there lots of risks, but it was more expensive than traditional warm roof insulation with PIR boards or similar. I do think people should research what they put in their own homes as it wasn't hard to find information that ruled out spray foam insulation fairly quickly.
Having said that, there is clearly some sort of regulatory gap here as not being able to mortgage your home is a very serious consequence of a relatively small and seemingly innocuous home improvement decision.
I guess most won't bother to read the full post and will instead react negatively to the title.
Exactly, it talks about ads in one paragraph of a very long post, and it's mostly to talk about all the problems that an ad revenue model has for FOSS!
Honestly people need to RTFArticle. It's talking about the result of interviews with developers on how they would prefer to be compensated, not definitive plans for what is or is not going to be allowed in F-Droid in the future.
Sounds like you're in the UK, if so I'd recommend legit companies run by old nerds like Mythic Beasts: https://www.mythic-beasts.com/domains
Same here. Works great, incredibly cheap too.
I’ve followed Robert Miles’ YouTube channel for years and watched his old numberphile videos before that. He’s a great communicator and a genuinely thoughtful guy. I think he’s overly keen on anthropomorphising what AI is doing, partly because it makes it easier to communicate, but also because I think it suits the field of research he’s dedicated himself to. In this particular video, he ascribes a “theory of mind” based on the LLM’s response to a traditional and well-known theory of mind test. The test is included in the training data, and ChatGPT3.5 successfully recognises it and responds correctly. However, when the details of the test (i.e. specific names, items, etc.) are changed, but the form of the problem is the same, ChatGPT3.5 fails. ChatGPT 4, however, still succeeds – which Miles concludes means that ChatGPT 4 has a stronger theory of mind.
My view is that this is obviously wrong. I mean, just prima facie absurd. ChatGPT3.5 correctly recognises the problem as a classic psychology question, and responds with the standard psychology answer. Miles says that the test is found in the training data. So it’s in ChatGPT4’s training data, too. And ChatGPT 4’s LLM is good enough that, even if you change the nouns used in the problem, it is still able to recognise that the problem is the same one found in its training data. That does not in any way prove it has a theory of mind! It just proves that the problem is in its training set! If 3.5 doesn’t have a theory of mind because a small change can break the link between training set and test set, how can 4.0 have a theory of mind, if 4.0 is doing the same thing that 3.5 is doing, just with the link intact?
The most obvious problem is that the theory of mind test is designed for determining whether children have developed a theory of mind yet. That is, they test whether the development of the human brain has reached a stage that is common among other human brains, in which they can correctly understand that other people may have different internal mental states. We know that humans are, generally, capable of doing this, that this understanding is developed during childhood years, and that some children develop it sooner than others. So we have devised a test to distinguish between those children who have developed this capability and those children who have not yet.
It would be absurd to apply the same test to anything other than a human child. It would be like giving the LLM the “mirror test” for animal self-awareness. Clearly, since the LLM cannot recognise itself in a mirror, it is not self-aware. Is that a reasonable conclusion too? I won't go too hard on this, because it's a small part of a much wider point, and I'm sure if you pushed him on this, he would agree that LLMs don't actually have a theory of mind, they merely regurgitate the answer correctly (many animals can be similarly trained to pass theory of mind tests by rewarding them for pecking/tapping/barking etc at the right answer).
Indeed, Miles’ substantial point is that the “overton window” for AI Safety has shifted, bringing it into the mainstream of tech and political discourse. To that extent, it doesn’t matter whether ChatGPT has consciousness or not, or a theory of mind, as long as enough people in mainstream tech and political discourse believe it does for it to warrant greater attention on AI Safety. Miles further believes that AI Safety is important in its own right, so perhaps he doesn’t mind whether or not the overton window has shifted on the basis of AI's true capability or its imagined capability. He hints at, but doesn’t really explore, the ulterior motives for large tech companies to suggest that the tools they are developing are so powerful that they might destroy the world. (He doesn’t even say it as explicitly as I did just then, which I think is a failing.) But maybe that’s ok for him, as long as AI Safety research is being taken seriously.
I disagree. It would be better to base policy on things that are true, and if you have to believe that LLMs have a theory of mind in order to gain mainstream attention on AI Safety, then I think this will lead us to bad policymaking. It will miss the real harms that AI pose – facial recognition used to bar people from shops that have a disproportionately high error rate for black people, resumé scanners and other hiring tools that, again, disproportionately discriminate against black people and other minorities, non-consensual AI porn, etc etc. We may well need policies to regulate this stuff, but focus on hypothetical existential risk of AGI in the future, over the very real and present harms that AI is doing right now, is misguided and dangerous.
If policymakers actually understood the tech and the risks even to the extent that Miles's YouTube viewers did, maybe they'd come to the same conclusion that he does about the risk of AGI, and would be able to balance the imperative to act against all of the other things that the government should be prioritising. But, call me a sceptic, but I do not believe that politicians actually get any of this at all, and they just like being on stage with Elon Musk...
- JumpDeleted
AI Ruined My Year - YouTube
The summary is total rubbish and completely misrepresents what it's actually about. I'm not sure why anyone would bother including that poorly AI-generated summary, if they had already watched the video. Useless AI bullshit.
The video is actually about the movement of AI Safety over the past year from something of fringe academic interest or curiosity into the mainstream of tech discourse, and even into active government policy. He discusses the advancements in AI in the past year in the context of AI Safety, namely, that they are moving faster than expected and that this increases the urgency of AI Safety research.
I've followed Robert Miles' YouTube channel for years and watched his old numberphile videos before "GenAI" was really a thing. He's a great communicator and a genuinely thoughtful guy. I think he's overly keen on anthropomorphising what AI is doing, partly because it makes it easier to communicate, but also because I think it suits the field of research he's dedicated himself to. In this particular video, he ascribes a "theory of mind" based on the LLM's response to a traditional and well-known theory of mind test. The test is included in the training data, and ChatGPT3.5 successfully recognises it and responds correctly. However, when the details of the test (i.e. specific names, items, etc.) are changed, but the form of the problem is the same, ChatGPT3.5 fails. ChatGPT 4, however, still succeeds -- which Miles concludes means that ChatGPT 4 has a stronger theory of mind.
My view is that this is obviously wrong. I mean, just prima facie absurd. ChatGPT3.5 correctly recognises the problem as a classic psychology question, and responds with the standard psychology answer. Miles says that the test is found in the training data. So it's in ChatGPT4's training data, too. And ChatGPT 4's LLM is good enough that, even if you change the nouns used in the problem, it is still able to recognise that the problem is the same one found in its training data. That does not in any way prove it has a theory of mind! It just proves that the problem is in its training set! If 3.5 doesn't have a theory of mind because a small change can mess up its answer, how can 4.0 have a theory of mind, if 4.0 is doing the same thing that 3.5 is doing, just a bit better?
The most obvious problem is that the theory of mind test is designed for determining whether children have developed a theory of mind yet. That is, they test whether the development of the human brain has reached a stage that is common among other human brains, in which they can correctly understand that other people may have different internal mental states. We know that humans are, generally, capable of doing this, that this understanding is developed during childhood years, and that some children develop it sooner than others. So we have devised a test to distinguish between those children who have developed this capability and those children who have not.
It would be absurd to apply the same test to anything other than a human child. It would be like giving the LLM the "mirror test" for animal self-awareness. Clearly, since the LLM cannot recognise itself in a mirror, it is not self-aware. Is that a reasonable conclusion too? Or do we cherry-pick the existing tests to suit the LLM's capabilities?
Now, Miles' substantial point is that the "overton window" for AI Safety has shifted, bringing it into the mainstream of tech and political discourse. To that extent, it doesn't matter whether ChatGPT has consciousness or not, or a theory of mind, as long as enough people in mainstream tech and political discourse believe it does for it to warrant greater attention on AI Safety. Miles further believes that AI Safety is important in its own right, so perhaps he doesn't mind whether or not the overton window has shifted on the basis of true AI capability or imagined capability. He hints at, but doesn't really explore, the ulterior motives for large tech companies to suggest that the tools they are developing are so powerful that they might destroy the world. (He doesn't even say it as explicitly as I did just then, which I think is a failing.) But maybe that's ok for him, as long as AI Safety research is being taken seriously.
I disagree. It would be better to base policy on things that are true, and if you have to believe that LLMs have a theory of mind in order to gain mainstream attention on AI Safety, then I think this will lead us to bad policymaking. It will miss the real harms that AI pose -- facial recognition used to bar people from shops that have a disproportionately high error rate for black people, resumé scanners and other hiring tools that, again, disproportionately discriminate against black people and other minorities, non-consensual AI porn, etc etc. We may well need policies to regulate this stuff, but focus on hypothetical existential risk of AGI in the future, over the very real and present harms that AI is doing right now, is misguided and dangerous.
It's a pity, because if AI Safety had just stayed an academic curiosity (as Rob says it was for him), maybe we'd have the policy resources to tackle the real and present problems that AI is causing for people.
People who say there's no difference between Tories and Labour can get in the sea. Or do some national service, idk.
I don't name my servers anything special, but I do name my various Zigbee sensors in Home Assistant after Egyptian gods. Atum-Ra, Tefnut, Shu, etc. I've avoided the ones that also coincide with Stargate gods, as I thought that would be too exciting for me.
Honestly this is better than nothing and very welcome for me. Not often I say good things about Michael Gove but he's done a great job on this and the cladding fiasco.
The real thing that cripples me as a leaseholder though is the service charges, which have doubled since I bought the place. The whole thing is a total con.
I found cleverkeys to be terrible at swipe/gesture typing. It would fail at even simple, common, seemingly unambiguous words. Heliboard + gboard swipe lib was orders of magnitude better.
I still use cleverkeys for termux though as the key gestures are brilliant.