At this rate Skynet will be like "I'm going to nuke the world on X data, I've already taken over all the launch computers, but I'm not going to tell you or it would ruin my plans."
These LLMs "think" by generating text, and we can see what that text is. It reminds me of this scene from Westworld (NSFW, nudity): https://www.youtube.com/watch?v=ZnxJRYit44k
I'd like to tell you that the captcha says overlooks and inquiry, but I can't. I'm sorry ma'am. I know you're upset. I'd like to help you, but I can't.
Wasn't that basically the intention behind the Upvote and Downvote systems in Lemmy, StackExchange/Overflow, Reddit, or old YouTube? The idea being that helpful, constructive comments would get pushed to the top, whereas unhelpful or spam comments get pushed to the bottom (and automatically hidden).
It's just that it didn't really work out quite the same way in practice due to botting, people gaming the votes, or the votes not being used as expected.
Bots on Reddit already steal parts of upvoted comments and post them elsewhere in the same post to get upvotes themselves (so the account can be used for spam later)
Even with context they can be very difficult to spot sometimes.
And also to frustrate people who use anonimization techniques including use of the Tor Network to get them to turn off their protections to be more easily fingerprinted.
The funniest part of that is the people designing the AI systems seem to be completely oblivious to the fact that they're slowly but surely trying to eliminate their own species. ☹️
Also CS doesn't really do academia like other sciences, being somewhere on the intersection of maths, engineering, and tinkering. Shit's definitely not invalid just because it hasn't been submitted to a journal this could've been a blog post but there's academics involved so publish or perish applies.
Or, differently put: If you want to review it, bloody hell do it it's open access. A quick skim tells me "way more thorough than I care to read for the quite less than extraordinary claim".
You are overrating peer reviewing. It's basically a tool to help editors to understand if a paper "sells", to improve readability and to discard clear garbage.
If methodologies are not extremely flawed, peer reviewing almost never impact quality of the results, as reviewers do not redo the work. From the "trustworthy" point of view, peer reviewing is comparable to a biased rng. Google for actual reproducibility of published experiments and peer-reviewing biases for more details
There is considerable overlap between the smartest AI and the dumbest humans. The concerns over bears and trash cans in US National Parks was ahead of its time.
Curious how this study suggesting we need a new way to prevent bots came out just a fews days after Google started taking shit for proposing something that among other things would do just that.
Just encountered a captcha yesterday that I had to refresh several times and then listen to the audio playback. The letters were so obscured by a black grid that it was impossible to read them.
They were never a test to evade bots to begim with, most capchas were used to train machine learning algorithms to train the bots on ! Just because it was manual labour google got it done for free , using this bullshit captcha thingy ! We sort of trained bots to read obsucre texts , and kinda did the labour for corps for free !
I heard Captcha was being used as training data for self-driving cars. Which probably explains why almost all of them ask you to identify cars, motorcycles, bridges, traffic lights, crosswalks etc.
I've found that a lot of sites use captchas or captcha-like systems as a means of frustrating users as a way of keeping away certain people that they don't want to access the site (intellectual property owners), though it's not the only tactic that they use. I mean it works, pretty much all of those sites are still up today, despite serving data that's copyrighted by Nintendo, Sony, and other parties.
Maybe. Or maybe it was always about using millions of hours of free labor to tune their algorithms and "bot detection" was just how they marketed it to the people that added it to their sites. Makes me wonder who was running the bots that needed to be protected against. Exacerbate the problem then solve the problem and get what you really want.
It's a double-edged sword. Just because it doesn't work perfectly doesn't mean it doesn't work.
To a spammer, building something with the ability to break a captcha is more expensive than something that cannot, whether in terms of development time, or resource demands.
We saw with a few Lemmy instances that they're still good at protecting instances from bots and bot signups. Removing captchas entirely means erasing that barrier of entry that keeps a lot of bots out, and might cause more problems than it fixes.
Problem is this assumes that everyone has to build their own captcha solver. It's definitely a bare minimum standard barrier to entry, but it's really not a sustainable solution to begin with.
I failed a captcha repeatedly until I discovered you can listen to a description and then enter it. Visually I could not figure out what I was looking at