@TheQuQu after using it for close to a month I am both amazed at how much it can do but also convinced that it's not even close to general intelligence.
@Moon Since it's just statistical token prediction (albeit incredibly good token prediction), it has no internal mental state whatsoever, so it can only reason "on paper". That's why folks get better answers if they ask for it to explain its reasoning before answering.
The funniest bit was when people tried to play 20 questions with it. Because it can't keep secrets, it'd just answer "yes" and "no" randomly, and then when you made a guess it would pretend it was thinking of that thing all along if it matched prior answers.
@TheQuQu one advance of gpt over markov chains is it has a limited memory. I have been reading about ways that people are trying to overcome the limited amount of space it has. for example I was trying to do stories with it and after so many messages it forgets details. people are trying to improve that.
@Moon That's usually a parlor trick, they give it all sorts of APIs it can invoke. For instance, because LLMs suck at math it'll recognize if there's a math question and outsource it to a Wolfram Alpha-style service. It probably has a clock and timer API endpoint it can hit.
In the early days before they locked it down tight, someone got it to expose a list of its API endpoints by leaking the preprompt.
@Moon Local open models are doing better than OpenAI is at that now, there's finetunes of Llama 2 that can have a 16k token context, which I believe is bigger than chatGPT supports. But even this memory is just the entire token history. If you want "hidden state" you can teach it to output tokens that denote a block of text to hide from the user, but I don't know of any that do that yet.
@Moon@TheQuQu There's always the option of chaining LLMs and autoprompting to overcome base limitations of token prediction to overcome the limitations of the basic token prediction design. You can create multiple sets of system prompts and even different models/finetunes that are specialized on different aspects of cognition, and have them work out an answer to a prompt together while sharing a database they can query.