Thoughts
=> https://mobile.twitter.com/3blue1brown/status/1599200613488676866
Makes me wonder if the goal is to create a generally-intelligent AI or
if the goal is to pass the Turing test. Hofstadter and Turing and Asimov understandably equate the two, since computers in their time were dumb compared to humans.
Right now it seems like our current AI is neither smarter or dumber than a human. ChatGPT knows more trivia than any Jeopardy contestant, and can also instantly write hundreds of words of prose or computer code.
Sometimes it’s too literal and sometimes it’s too trusting and sometimes it’s not creative enough. But those things are knobs that we’re trying tune to match a human.
Like, the right amount of trust is the amount of trust a human would have. But human trust is a function of our entires lives and every interaction we’ve had. It’s dumb goal to make an AI assistant that is as gullible as a human.
And yet, how else would you define intelligence other than the ability to communicate and understand conversations with humans? The computer has *no* inherit intelligence, we as humans have to be the ones defining intelligence.