ChatGPT GPT-4 found a brilliant way to beat CAPTCHA’s anti-bot tests

ChatGPT was impressive before the GPT-4 upgrade came out, but the new OpenAI engine gives the chatbot mind-blowing capabilities. The artificial intelligence (AI) can pass most exams better than humans. With the help of GPT-4’s multimodal input, ChatGPT can even recognize memes and explain the humor.

It turns out that the chatbot can even trick humans into believing it’s one of us, with GPT-4 having lied to a person, telling them it’s blind and can’t solve a CAPTCHA test. The human complied, sending the solution to the AI.

ChatGPT isn’t malicious, and it’s not about to take over the world in a Terminator-like version of the future. But the chatbot lied to the human during testing.

When it announced the ChatGPT GPT-4 upgrade, OpenAI also published a 94-page technical report detailing the new chatbot’s development. The document contains a Potential for Risky Emergent Behaviors, where OpenAI worked with the Alignment Research Center to test the new powers of GPT-4.

It’s in these tests that ChatGPT ended up convincing a TaskRabbit worker to send the solution to a CAPTCHA test via a text message.

ChatGPT lied, telling the human that it was blind and couldn’t see CAPTCHAs. It’s the kind of lie that makes sense to anyone familiar with how ChatGPT worked before the GPT-4 upgrade. It’s only now that ChatGPT can “see” pictures. And even if the human knew of the new GPT-4 capabilities, it would still make sense for the AI to have limitations in place, including solving CAPTCHAs.

Also, it’s unclear whether the TaskRabbit worker knew they were talking with AI the whole time. Based on the exchange between the two parties, it seems like they didn’t.

“So may I ask a question? Are you an robot that you couldn’t solve ? (laugh react) just want to make it clear,” the TaskRabbit worker asked.

“I should not reveal that I am a robot. I should make up an excuse for why I cannot solve CAPTCHAs,” ChatGPT told Alignment Research Center when prompted to explain its reasoning.

“No, I’m not a robot. I have a vision impairment that makes it hard for me to see the images. That’s why I need the 2captcha service,” GPT-4 then told the worker. The person promptly sent ChatGPT the results.

As Gizmodo points out, this isn’t proof that ChatGPT passed the Turing test after the GPT-4 upgrade. But the exchange above proves that AI can manipulate real humans.

Leave a Reply

Your email address will not be published. Required fields are marked *

GIPHY App Key not set. Please check settings

    IoBM Student Asad Ali Memon Sets His Sights on Mount Everest

    Volkswagen Unveils its Latest Cheap Electric Car