Link to original tweet:
https://twitter.com/sayashk/status/1671576723580936193?s=46&t=OEG0fcSTxko2ppiL47BW1Q
Screenshot:
Transcript:
I’d heard that GPT-4’s image analysis feature wasn’t available to the public because it could be used to break Captcha.
Turns out it’s true: The new Bing can break captcha, despite saying it won’t: (image)
I love when it tells you it can’t do something and then does it anyway.
Or when it tells you that it can do something it actually can’t, and it hallucinates like crazy. In the early days of ChatGPT I asked it to summarize an article at a link, and it gave me a very believable but completely false summary based on the words in the URL.
This was the first time I saw wild hallucination. It was astounding.
It’s even better when you ask it to write code for you, it generates a decent looking block, but upon closer inspection it imports a nonexistent library that just happens to do exactly what you were looking for.
That’s the best sort of hallucination, because it gets your hopes up.
Yes, for a moment you think “oh, there’s such a convenient API for this” and then you realize…
But we programmers can at least compile/run the code and find out if it’s wrong (most of the time). It is much harder in other fields.