I kinda want to poke at their latest models. It is funny to me that people think LLMs can pass the turing test now. Oh really? Ask them any question where there is a common interpretation and a slightly less common one and then see if they can find it without spelling it out exactly. Two fun examples with grok 3 from the past day. Give it the Man crossing a river with a wolf a goat and a cabbage, but don't make any restrictions on what the boat is capable of. Over fit much? Read the question I asked, not the one you assumed it to be! Granted most humans would screw up too, but at least with a human you can tell them to go back and read the question carefully and they will get it. Today I asked it to make the following statement bold. "I can probably do it, with enough luck and no real showstoppers." It added <b> tags. And when I told it that still wasn't bold. It made it all caps, then increased the font size, then made it flash.

Replies (1)

Yeah, it's funny that way. Search engine and fast meme modification are all in really getting out of it. And I guess the occasional coding help, but for really trivial stuff ("how to join data frames with blah blah") type stuff. Oh, and I guess I was successful in getting to to write a script to scrape a bunch of datasets online for me (something that would have taken me days to figure out on my own). Im zeroing in on: should use it way more for code than i do, early initial search (but nothing too deep as it just spews junk, when you know the subject enough to see it), occasional and inoffensive memes