New blog post: When Visual Chatbot is wrong, it just keeps digging itself a hole.
Visual Chatbot is so much fun.
@janellecshane I'm fascinated that it just... lies. Even if that's a bit of anthropomorphism, there's no intent to deceive. It's easy to catch it in contradictions but the fact that it doesn't just fold and give up feels different somehow.
I'm way out of touch with how these actually work, I only played with making a NN like 8 years ago, but here's a weird question - if you were to use adversarial techniques on this, would you be making a better liar, a better description of photos, or both?
@floatoverblow I wonder if the lying is because it was trained to copy human responses to questions - and the humans very rarely folded in confusion. What the bot doesn't know is this is because the humans were very rarely confused. It copied the confidence but not the reason.
@janellecshane It reminds me of my 4 year old niece. She'll never say she doesn't know about a question, she'll figure out something plausible and say that instead. Makes teaching her to read very difficult, but entertaining - she'll often just read what she thinks things _probably_ say about the pictures.
@janellecshane I read the blog post but, after playing with the chatbot a little myself, are you sure it's not just badly made (or, as you suggest, answering randomly)? I tested by just typing random letter strings, to which it always replied "yes."
In one interaction it appeared to be gaslighting me, though, so maybe it really is just a lying liar that lies.