Lexicon Induction From Temporal Structure

I get it! :smiley:

I recently discovered a bug in my implementation of HTM which could be affecting this scenario, so I will let @rhyolight comment on how well NuPIC handles a single repeating input.

In htm.js a single repeating input stabilizes on three sequential representations. I’m guessing that NuPIC probably bursts every other element in the sequence and eventually saturates the capacity of the columns (which wouldn’t be useful either)

While I understand the representation space a bit better, I’m still confused by the enwiki8 format. I don’t think I understand the problem well enough to recommend HTM for the job. As you said a couple times, as we feed in “one bit at a time”, it seems to me like a very inefficient use of the SP / TM algorithms. I don’t think it is going to perform very well on an encoding like this, because not enough semantics are being encoded in each time step. The full capacity of the SP / TM will never be used.

Moved from #htm-theory:neuroscience to #htm-theory

Does your algorithm attempt to (albeit perhaps with bugs) implement temporal learning, as illustrated by HTM School #12?

Yes. The bug may be a bit far into the weeds depending on how familiar you are with the algorithm, but I described it here

The point is to exploit the ability of HTM to abstract representations in a “hierarchy”, as shown in its iconic illustration thereof.

If HTM can abstract, from the trivial, two-“note”, data encoder, things like bytes, it may, then, be able to abstract, from bytes things like morphemes, punctuation, etc. at the next layer in the hierarchy, and then abstract words (what people normally think of as “the lexicon”) at the next layer of the hierarchy, then sentences – at which point “thought”, “semantics” etc. have become apparent in the SDR at that layer. These are the sorts of knowledge representation that can lead to better prediction, going back down the hierarchy to the bottom, of “the next bit” hence superior compression.

We think there is a whole lot more going on in one level of the hierarchy than we originally thought. None of our newest theory work includes hierarchy. So I could not recommend HTM at this point if you think hierarchy is the magic sauce. It may very well be, but we still have more to figure out in a cortical column before working on hierarchy again. For more info see H is for Hierarchy, but where is it?.

Thanks. That history is a good perspective. However, the link to Hawkins’s critical followup presentation wasn’t immediately obvious.

If I correctly understand from Hawkins’s talk the current strategy at Numenta:

Until the “function” of the macrocolumn is understood, “bottom up” (in terms of the neurophysiology), revisiting the hierarchy, “top down” (in terms of global function), is likely to be unsound. The, apparently minor, critique I would have of this is that even while driving research from the neurophysiology, there is, nevertheless, always, a range of plausible global functions under consideration. I mean this is, after all, what Numenta has basically stated about perception: That it is biased – primed by the plausible interpretations aka “expectation”. The reason I say “apparently minor” is that it is, indeed, a minor critique in that I’m sure Hawkins et al understand this is what they’re doing and, quite reasonably, expect other to as well. However, I say “apparently” because it is always good practice in science to be explicit about the hypotheses being tested – not individually or absolutely (ie: Not in the Popperian “falsifiable” sense) – but relatively: Rank ordering in terms of plausibility given the ground truth observations, which, in this case, is the neurophysiology.

This requires a review of the literature of macrocolumn function.

My particular interest about “lexicon induction” is that my review of that literature leads me to believe the most promising macrocolumn function involves a given lexicon specific to each macrocolumn. Once this lexicon is established, global phenomena naturally emerge such as syntax, grammar and semantics, in natural language.

The mystery is how the lexicon is established for each macrocolumn and how the lexicons are factored between macrocolumns.

Of course, I’m not suggesting that Numenta drive their work on the macrocolumn function from this hypothesis because that would be all too Popperian. No, I’m suggesting a more relativistic Plattian “strong inference” approach.

In this regard, I would also strongly suggest taking very seriously “Universal Measures of Intelligence” involving Kolmogorov Complexity when establishing the relative plausibility of unsupervised learning hypotheses. I don’t believe any of the current major efforts in AGI are doing this, although it has been known to be the correct approach since the early 1960s with the papers by Solomonoff, Kolmogorov, Chaitin et al. These were, coincident with Platt’s paper on strong inference.

PS: The top of Google’s search results page for strong inference and Popper comes up with an execrable paper titled “Fifty years of J. R. Platt’s strong inference.” I won’t dissect it but will simply quote the last, very Popperian, sentence: “It [strong inference – jab] is a message that can benefit anyone who is interested in tackling difficult problems – we must be bold enough to assume that one of our ideas is correct [emphasis – jab], and yet we must have the humility to abandon those ideas that don’t stand up to scrutiny.”

1 Like

Where you say “macrocolumn” I’m going to use the term “cortical column”, which is our current terminology (in contrast to the “mini-column”, which exists within layers of cortical columns).

Yes I think this is correct. Or at least one cortical column’s representation of a “ball” will not be the same as another’s. Every column has a unique encoding of reality.

I think they are pretty randomly initialized and they grow organically over time using established Hebbian learning rules.

Even though each cortical column represents objects with a completely different set of neurons, representations can still be communicated between columns because synaptic growth between columns links these ideas together.

Regarding your other comments, I can’t speak for Numenta’s research philosophy.

the “mini-column”, which exists within layers of cortical columns

Accepting the term “cortical column”, my impression of minicolumns is that they do not stack to form cortical columns but a cortical column is comprised of a large number of parallel minicolumns - each of which has all the layers of the cortical column.

Is my impression incorrect?

(There is clearly a terminology problem in the field of neuroanatomy as indicated by this quote from “The Neocortical Column” by DeFelipe et al: “the term Colum is used freely and promiscuously to refer to multiple, distinguishable entities, such as cellular or dendritic minicolumns or afferent macrocolumns, with respective diameters of <50 and 200–500 μm.”)

one cortical columns representation of a “ball” will not be the same as another’s

A “ball” meaning an n-dimensional sphere in an appropriately scaled space of the world’s properties, containing what I might call a “lexicon” that represents some property of the world – sensed and/or imputed, e.g. color in terms of a “lexicon” like (“red”, “green”, “yellow”, etc.) ?

I think they [the lexicon of each cortical column – jab] are pretty randomly initialized and they grow organically over time using established Hebbian learning rules.

If it is really that simple, I refer you back to my original, very simply stated but very difficult challenge here.

Hawkins is absolutely on target in placing time as the foundation. It was searching for prior work on neocortical modeling, so founded, that drew my attention to Numenta, and which elicited my challenge based on a serial bit stream to do a very simple lexical induction: the ‘byte’.

To quote Hawkins from the aforelinked video, at this time code:

“The trick to making progress is to really pick the right problem to solve. I can’t say that enough.”

There are a couple of things that HTM is currently lacking that make it ill suited for this particular challenge IMO.

Firstly, it lacks an ability to “self label” sequences in temporal memory. In practice, this means a robust temporal pooling strategy which is able to form stable representations of sequences which capture the proper semantics, and which can quickly adapt when switching from one sequence to another (attention will likely be a key component of the latter)

Secondly, it lacks stability in repeating sequences. Because this particular challenge involves only two unique inputs, it means that the system will encounter lots of cases where the same input is repeated many times in sequence. The system must still be able to encode context properly, and be able to distinguish those same repeating patterns if it exists within multiple higher-order sequences.

Thirdly, it lacks motor control. I suspect that even a human subject tasked with finding the concept of a “byte” from a stream of on / off inputs will not use pure sequence memory to do it – they will use motor control to manipulate the input and explore it (by physically scanning their eyes over a screen or sheet of printed paper, by replaying the input in their “mind’s eye”, etc.)

Note that these are all things that need to be addressed by HTM to properly model a single level in the hierarchy. I’m not sure what hierarchy itself would bring to the table without these more basic capabilities.

1 Like

Thirdly, it lacks motor control. I suspect that even a human subject tasked with finding the concept of a “byte” from a stream of on / off inputs will not use pure sequence memory to do it – they will use motor control to manipulate the input and explore it (by physically scanning their eyes over a screen or sheet of printed paper, by replaying the input in their “mind’s eye”, etc.)

Now you’re challenging me to do something I’ve been thinking about with regard to this question:

Actually synthesize music with these patterns to see if lexical induction emerges in humans from temporal input. I strongly suspect it does.

While you may be correct that humans – as entire organisms – utilize tempo-spatial input to the maximum extent possible, that detracts from the central importance of the challenge to Numenta’s model of the cortical column.

PS: I’m specifically and very carefully not addressing the hierarchy in this challenge, but rather am issuing a challenge to those who claim that it is important to figure out what a cortical column does as a prerequisite for hierarchy…

I don’t know whether mini-columns extend throughout all the layers in a cortical column or if some or most of them are localized within a layer.

Our terms have evolved over time. Perhaps this video will help with our current nomenclature:

Yes exactly. But some cortical columns will represent the ball in terms of somatic sensory input because they are processing that type of input. Other cortical columns processing visual input will represent visual aspects of the ball. In both cases, the cortical columns are mapping sensory features to locations in space.

Keep in mind that our cortexes did not evolve intelligence to process information at this level. Our brains process gargantuan amounts of parallel data, all inter-associated over time with our own actions and the world. It doesn’t surprise me when HTM is bad at solving these very low-level prediction tasks involving very small encodings. The encodings coming from all our senses are much, much larger than any of these we are even simulating.

True, and I am calling out the fact that there is a lot of work to be done in understanding and modeling the low-level capabilities that HTM theory needs to address before it will be suited for this challenge. I think we agree that the ability to extract concepts is a prerequisite for hierarchy. I am just being specific about a few of the capabilities (from my perspective) that are missing from HTM.

1 Like

Let’s be careful about the distinction between HTM theory and brain reality. We can argue forever about what evolution did. We can argue for a far less version of “forever” about HTM theory.

What is it in HTM theory that tells us to expect that it becomes worse at doing things as the number of SDRs a cortical column must deal with decreases?

HTM works because of the dynamics of large populations of neurons. Input into the system also involves large populations of neurons. If the input is sufficiently small, the SDR comparison properties we depend on to represent concepts breaks down. I just don’t think it will work.

Here is a good paper describing why HTM depends on these properties of SDRs requiring large populations of neurons.

We should also be careful about the distinction between the capabilities of HTM now in its current state, and the goal of HTM, which is to model intelligent systems by using the biology of the cortex as the point of reference. With that in mind, I think it makes perfect sense to reference brain realities in response to how HTM should/ will eventually react, and whether it is or ever will be suitable in a particular environment (in this case, small-scale isolated inputs with high density)

It is probably worth mentioning that Numenta’s current research is published at https://numenta.com/papers/ and latest is always at the top. We include code samples with most papers, including this last one.

I will try to make an effort to disambiguate between implemented theory vs conjecture. In any case, with speculation or not, it doesn’t seem to me that HTM will be competitive in this task.