I’d like to know some perspectives about the following question that’s bugging me for a while now.
If there is such a thing as a state of mind, then why do we only capture a single state of the mind and expect it to do intelligent things? For example, we expect it will generalize its inputs?
It is easy to imagine that an input can be exclusive to another input(s) with regards to its place in memory. However, we seem to ignore this and believe we can generalize these inputs by using massive amount of parameters - deep learning. Probability theory can help tell the probability of an input or pattern to occur but it is not representative to a dynamically changing state of mind.
Just a thought.