| complyue
March 7 |
robf:
Happy to contribute to a separate thread on any of these themes if you want to create one.
Iâm also very interested to hear more of your thoughts regarding language, causes & effects, and etc., already felt a great deal of inspiration from your posts in this thread!
Glad youâre finding something promising in it.
I have some concrete experiments Iâd like to try. There is a lot of commonality with aspects of HTM. As I say, I corresponded in (the predecessor to?) this forum a lot between 2014-2016. At that time the whole sequentiality aspect was an even rarer fit. That was before the whole transformer explosion, and the only work on any kind of sequence in AI was HTM, and some stuff with LSTM for language translation.
The experiments I would like to try can be very simple. If someone wants to explore a simple implementation with me, that would be great.
Basically, what I want to try is to build a network of observed language sequences, and then give it a good âshakeâ, and see how the energy surfaces cluster. Rather as was described for community detection in the Brazilian paper. Only doing it for language sequences.
I already managed to implement such a language sequence network in an open source neurosimulator. And I was able to get it to oscillate. I was surprised how easy that was. I thought I would have to try some clever feedback mechanisms to get oscillations. But it turned out networks of language naturally feedback. Common words like âtheâ close the loop naturally, and any propagated excitation loops back around. The only thing I needed to do to get oscillation was to implement inhibition. And even that was simple. I just needed to inhibit randomly. Then I turned the inhibition up and down until I got a balance point, and the network oscillated nicely.
The next stage would be to try and find a hierarchy in the raster plot of those oscillations.
So far itâs really simple. Just a network of observed language sequences. Apply inhibition over it fairly sovereignly. And turn the inhibition up and down until it oscillates.
The next step is to look at a raster plot of spike firings in the neurosimulator, and try to see how we might squeeze some kind of hierarchical "communityâ structure in it.
robf:
So I think a more fundamental bias is just more easily seen in language.
But yes, I donât think language is fundamental at all. (Iâve actually spent a lifetime learning different languages to explore that idea for myself, subjectively!)
Maybe the semantics part can be fundamental? As the syntax on the surface and even rule-based grammar are not apparently fundamental.
The conclusion Iâve come to is that âthinkingâ is an assembly of perceptual impressions. Language is a dance which is part of that, and prompts thinking by its connections to it. But it is not fundamental to it.
Fundamental are observations, examples. What Kuhn called a âparadigmâ:
Kuhn SoSR, p.g. 192 (Postscript)
âWhen I speak of knowledge embedded in shared exemplars, I am not referring to a mode of knowing that is less systematic or less analyzable than knowledge embedded in rules, laws, or criteria of identification. Instead I have in mind a manner of knowing which is misconstrued if reconstructed in terms of rules that are first abstracted from exemplars and thereafter function in their stead.â
If you say something in two languages, the impression Iâve been able to form (from a non-native level) is that the sensation of thought beneath the two languages is much the same, and more a feeling of âbeing thereâ, than it is something connected to any one of the languages. Actual âthoughtâ is below it, and separate. The feeling of expressing an idea is much the same.
This is quite different to the hypothesis associated with Sapir-Whorf, that language completely governs our conceptual system (and which was the idea that first seduced me to the idea of learning another language.)
This feels to me like something very similar to some comments I read of Einsteinâs opinion about the same thing:
âThe ⌠elements in thought are ⌠images which can be ⌠reproduced and combinedâ
How Einstein Thought: Why âCombinatory Playâ Is the Secret of Genius
https://www.themarginalian.org/2013/08/14/how-einstein-thought-combinatorial-creativity/
So, âimagesâ and not words.
I have ârational thinkingâ compared to âdaily language usageâ in my mind, e.g. one usually say âthereâs a straight line sharply risingâ, but with formal math language, one can give y = 5x + 3
to state it more precisely. I suppose math as a language, is one (or even more) orders/levels higher than usual ânatural languagesâ? But semantically one can always answer that y
is 48
, when asked âwhat if x
is 9
â? Maybe less accurate and with more vague phrasing, but as well as the in-situ pragmatics demands.
The thing about maths is that it was shown to be incomplete. Formalizations are incomplete. I think the basic level is combinations of examples/perceptions. And those combinations can contradict. Maths excludes contradictions. But by doing so excludes entire areas of âtruthâ.
My signature example for that formal contradiction, giving rise to more mental descriptive power not less, is non-Euclidean geometry. That mathematicians sought for millenia to prove Euclidâs 5th(?) postulate that parallel lines never meet. Until⌠Gauss, decided to simply allow parallel lines to meet, and discovered it created for him an entirely new branch of mathematical truth, the geometry of spheres (instead of planes.) Parallel lines meeting or not meeting, is a contradiction. But it is not that one or the other is not âtrueâ, just that the two âtruthsâ apply to different aspects of the universe.
robf:
Great! Yes. Reservoir computing. Total agreement. Thatâs my current guess at an origin story for âcognitionâ (along the energy minima/maxima prediction enhancing grouping) lines Iâm describing. I imagine cognition started as a kind of âecho state machineâ over even very simple early nervous systems, imprinting events and consequences, causes and effects, as a kind of âecho mechanismâ. And then evolution simply enhanced that cause and effect echo state network, by evolving to enhance the âechoâ mechanism. Some kind of simple amplification of the âechoâ. Amplification by âstackingâ events which shared causes and effects. And that âenhancementâ is what we now call âconceptsâ. Proto-âconceptsâ might just be groupings of things which tend to share causes and effects. And found by seeking energy minima in networks of sequences of observations.
Along this line of thought, I would suggest our (humanâs) modeling of the world is not compact/precise description of the underlying (possibly chaotic) function at all, but a (good enough) approximation. We can apply Newtonianâs physics perfectly okay for daily living, even though we know itâs incorrect (compared to general relativity, and quantum mechanics in turn, and âŚ), we can still believe itâs universally true where it applies.
Our formal models, yes, I agree. They are âgood enoughâ approximations. But actual thought is itself a chaotic recombination of observations. And actually more powerful for that. We have been missing that power, because weâve been assuming the mechanism for generating formal models, is the same as the formal models themselves.
The archetypal formal description would be maths. And the thing about maths is that it was shown to be incomplete. Formalizations are incomplete. I think the basic level is combinations of examples/perceptions. And those combinations can contradict. Maths excludes contradictions. But by doing so excludes entire areas of âtruthâ.
My signature example for that formal contradiction, giving rise to more mental descriptive power not less, is non-Euclidean geometry. That mathematicians sought for millenia to prove Euclidâs 5th(?) postulate that parallel lines never meet. Until⌠Gauss, decided to simply allow parallel lines to meet, and discovered it created for him an entirely new branch of mathematical truth, the geometry of spheres (instead of planes.) Parallel lines meeting or not meeting, is a contradiction. But it is not that one or the other is not âtrueâ, just that the two âtruthsâ apply to different aspects of the universe.
We are capable to âlearnâ where the chaotic boundaries are, and enjoy smooth (simply approximatable) continuums between them. I suggest that there are âdeep-knowledgesâ which is about where/when/how some âshallow-knowledgeâ applies. Contemporary AIs seem to have learnt âshallow-knowledgesâ well, but not at all regarding deeper knowledges.
Fair enough with modeling âdeep-knowledgesâ. But I think it is modeling ânew-knowledgesâ where the current tech falls down.
âdeep-knowledgesâ you are talking about might be where the broader sensory data that neel_g talked about would start to become relevant. I donât disagree with him that broader data sources will be important. Those will be the âimages which can be ⌠reproduced and combinedâ which Einstein talked about. I just think that there is still an insight to be gained from the simpler language based system first, about how a system might find new structure all the time. Once we try that with the simple sequential system of language, we might apply it broadly across all perception.