Submitted by cancolak t3_119d8ls in singularity
In this article, Stephen Wolfram (known for Wolfram|Alpha, Mathematica, etc.) discusses the inner workings of ChatGPT. It's an in-depth look at what goes on under the hood of an LLM and one of the best explanations of how neural networks work. It's a great read for anyone who wishes to actually understand this amazing piece of technology.
My main takeaways from it were:
- Some aspects of neural network design are well understood, and their structure is fairly straightforward mechanically. However, it is almost impossible to get a human understanding of what the machine is doing inside each particular step. In that sense, they are indeed black boxes
- Contrary to popular belief, neural networks don't represent the ultimate next step forward in computing. They obviously are limited by their size and the data available, but beyond that, they tend to perform badly at computationally irreducible tasks. He makes the point that most of nature can be boiled down to computationally irreducible processes, making neural nets an unlikely candidate for generating previously unavailable knowledge of reality. Luckily for us, computers are fairly good at computationally irreducible tasks (think multiplying very large numbers or running complex programs in parallel, etc.) so we can count on their continued aid
- Humans tend to think of natural human tasks such as thinking and speaking as very complicated processes, however, the success of ChatGPT at speaking may indicate otherwise. Since neural networks are good at computationally reducible tasks, the fact that they ended up becoming very good at natural language might suggest that thought & speech aren't particularly difficult, at least computationally. Furthermore, this might suggest that there could be some fairly simple rules yet uncovered which underline language patterns
This analysis by a very smart guy who's worked with neural networks for 43 years has reaffirmed my belief that there exists no easily viable path from an LLM to a conscious machine. That is if we DO NOT define consciousness to be the ability to conjure language-based thoughts. ChatGPT already proved that it can do that. If we define consciousness to be the entirety of human experience, with all of awareness and sense-perception and all the other hard-to-explain stuff bundled in (a lot of which are presumably shared by other forms of life and brought about by evolution over eons), then it's highly unlikely that a neural net gets there. That is because natural processes, at least according to Wolfram, are computationally irreducible.
RiotNrrd2001 t1_j9mddet wrote
I imagine at some point LLMs will be paired with tools that can handle the things they themselves are poor at. Instead of remembering that 3 + 4 = 8 the way it has to today, it will outsource such operations to a calculator which will tell it that the answer is actually 7. That ChatGPT can't do that today and still does as well as it does is actually pretty impressive, but... occasionally you still get an 8 where you really want a solidly dependable 7.
These are the early days. There is still some work to be done.