AGI - what part does HTM/TBT play?

I guess you are talking about a hierarchy of reference range, that’s different from the hierarchy of composition. Both can be seen as clustering, but in the former the elements (minicolumns?) stay the same while the clusters may grow. In the later, the elements are lower-level clusters, not their elements. Which is a real hierarchy, while the former is not, higher levels are parallel to each other.

I recollect that he was talking about “unpredictable” as in not reducible to a simpler formula. The results are perfectly predictable by running the function. It’s easier to see if we replace prediction with compression, a more constructive term. Normally that means compression of representations, while he is talking about compression of computation, which hopefully compresses representations.

I am talking about both (also all higher) orders of compression. They can be combined if we multiply each by relative corresponding cost, in this case cost of memory / cost of operations, which are fluid and situation-specific.

Every goal is ultimately an instrument to some other goal, as soon as you broaden your perspective :). So, they are all instruments at some point, but more general / stable instruments are relative goals. And there is no more general instrument than prediction, the goal of pure curiosity, exploration, science.

I just defined you the purpose, several times over. It doesn’t get through because it sounds too abstract. Guess what, this whole subject is abstract, more than anything you can imagine.
Just try to name some other general purpose, and I will show you that it’s either reducible to prediction, or actually biologically specific.

I fail to see any relevance here.

Hi Boris, let me first of all thank you for taking the interest to respond and add your insights and experience, which really enriches such a forum, greatly. So please do not take any comments as critical, it is, as you mentioned, not easy to always communicate at the same level of abstraction. And some points here do enter the terrain of scientific philosophy. Let me start with the point on Kurt Gödel’s “Incompleteness Theorem”. This proof just reveals one important aspect about the reality of this entire universe. The fact that some “truths” about the universe cannot be mathematically proven, which also means they cannot be mathematically derived from any input data, has some strong implications. Not even if you had perfect information of the entire universe. This is a very profound discovery that seems to stand up to all mathematical scrutiny. It is almost as profound as “the big bang” to cosmology or “quantum mechanics” to physics. The implication here is some “real” things (like some real processes that take place) are not learnable, no matter how clever you are. This theorem of incompleteness is also strongly supported by Stephen Wolfram’s computational analysis and specifically his research resulting in the “Theory of Computational Irreducibility” which is basically a computational proof of Gödel’s theorem, for specific selected functions. And Wolfram’s complete work results in the “Theory of Computational Equivalence” which then concludes that our universe is full of many very simple functions (algorithms) which very surprisingly (for mathematicians) very quickly reach maximum complexity in their output, (this means, unpredictability by any mathematics). So both Gödel and Wolfram are revealing that we exist in a universe that is much less mathematically computable (even in theory) than any one had imagined in the early part of the 20th century and even at the end of the 20th century.

Therefore, this has implications for our definition of “intelligence” and how we can measure it. “Predictive Power” is still very important and it is possible, in many areas of reality (Stephen Wolfram calls these pockets of predictability, which could be big also). But the vast majority of computational space seems to clearly fall in the non computable arena, meaning not compressible by a single step. This means that many things can only be learned by doing every single step and memorizing or recording every single step. Computational equivalence (or irreducibility) means that you can only know the 3 trillionth digit in Pi by calculating every single digit one by one without skipping a single one. And the number of patterns in computational space that are simple and yet unpredictable is amazingly large. (Pi is no exception, it is a very common phenomena). So we have to let go of the idea that absolute intelligence can be measured only by predictive power. We have to start thinking more in the direction that any entity which we consider intelligent is trying to predict outcomes for a purpose, like survival. But in a universe in which most of reality is not predictable (not deterministic) we need to then add functions to make approximations and cope with uncertainties. A solid definition of intelligence must contain some metric of its ability to cope with uncertainty in a mathematically unpredictable universe. This “coping with unavoidable uncertainties” is what I call the purpose, like survival or like finding a path to a place or like getting to Mars alive. That is how I arrived yesterday to my conclusion (thanks to your comments) that “In the absolute absence of any given purpose (i.e. goals) intelligence does not have any meaning (any significance)”. The scaling of intelligence can only be specific relative to a purpose, i.e. how successfully it copes with uncertainty to fulfill a purpose.

Please bear in mind, this is a philosophical implication in the context of the absolute universe. Our pocket of reality within the universe (also our computational universe) has plenty of yet unexploited pockets of predictability. There is still a lot yet to be achieved, even if reality has set many limitations to our ideal space of computability. The field of quantum mechanics is also full of such examples of limitations. Which is why it resorts to so much statistics to make probabilistic predictions. It can only predict probabilities, not specific outcomes. But we learned to cope with that for our purposes.

Indeed this is very heavy stuff to digest, so do not feel compelled to comment, though would love any further insights on this if you have them, from your experience and research context. From an AI Engineering point of view, this reality may be a nuisance, but does not give us reason to stop searching for better paths to higher intelligence. And like Jeff states, we have our own human intelligence as a clear example of what this universe has made possible. And if we define “human intelligence” as our goal and purpose, then we have a set of metrics we could derive in order to measure our progress toward human level AGI.

Regarding:

This sounds like you have a lot of experience and deep understanding of the types of frameworks that can be applied at different levels. I follow your logic regarding the fact that we have different types of conceivable architectures (with lower level clusters vs. lower level elements). I guess to some degree, one can also say that the concept of elements can emerge from clusters. I personally am quite certain that hierarchical structures play a very important role in our brains. One open question is whether hierarchical processes coexist at multiple domains, perhaps even with overlapping. My intuition from lots of my reading on this topic in the context of neuroscience is that evolution has been able to produce multiple levels of hierarchical organization in our brain and CNS. Within the Numenta and HTM communities we already have several examples in the models. The HTM level and the TBT levels are basically nesting HTM within TBT and the TBT reference frames allow for very many overlapping formations. The question I would like to explore is how set-theory is built into the TBT paradigm and may allow for many levels of logical nesting. Do you have any insights in this area?

2 Likes

Could you elaborate a little, why this is an open question?

This is also very intriguing … nesting/hierarchy/overlapping seem to me to exist univerally in the universe of things. Would be greatly educational if you may list a few specific/real design issues.

Somewhat heavy going, but the points you make are simple enough. I don’t buy the idea of maths or Incompleteness having much to contribute, but the end result is the same. IMO it’s enough to say that in the real world the past is gone, the present is uncertain and the future is somewhat predictable. There is no “perfect information of the entire universe” other than the universe itself. QM ensures that this is so.

As Einstein put it: reality is an illusion, albeit a very persistent one. The role of intelligence is to construct a reality that is good enough to understand and take action that leads to a desired outcome in some small part of a real world.

Organic intelligence gets sensory input and and takes action in a physical world, but an AGI naturally receives digital inputs and takes digital action. The question I keep coming back to is: what task(s) should we set an AGI so that we know when we have one, or are making progress towards one?

A crow can drop stones into a narrow flask to raise the water level so it can drink. What can an AGI do?

I believe the contrary … Wolfram’s “Computational Irreducibility” and Godel’s Incompleteness have a critical contribution: They tell us not to expect successfully building an AGI that can perfectly predict the future. Wolfram showed simplistic algorithms can produce chaotic (aka unpredictable) outputs, Godel’s incompleteness theorem showed there exist truths that cannot be logically/rigorously deduced (aka proven).

To me, these are analogous to: understanding the laws of thermodynamics actually stopped futile scientific endeavors of building perpetual motion machines. It is significant.

We of course should not expect building AGIs to “predict the future”, in a general sense.

Perfectly intelligent and rational short term behavior could lead to long term disaster, as wars or financial market crashes have demonstrated again and again – what AGI is there to tell us how many “terms” in the sense of micro/short/medium/long/super long to consider, before we can conclude what strategy is “intelligent and rational”? If we believe Wolfram and/or Godel, we accept “no AGI can do that, period.”

As soon as we figure out an algorithm to accomplish a task – once we understand the algorithm – we consider that computation, or automation, and stop considering that as “intelligence”.

So, in my mind, you raised a great question that is actually unanswerable, since it’s like chasing after a moving target.

And that’s not wrong or bizzare – it’s as simple as the defition of “infinity”: for any distance one has reached, infinity lies farther away from it. I believe that is the rigorous mathematical definition of infinity. The definition of AGI cannot be much different … for any clever computer algorithm humanly designed/created, AGI should do better than that.

1 Like

But that’s my point. Maths may tell us not to expect perfection, but gives us no clue how close we might get or how useful a near miss might be. Maths tells us the Traveling Salesman problem has no perfect solution in any reasonable time, but no matter, near enough is usually good enough.

We of course should not expect building AGIs to “predict the future”, in a general sense.

But an AGI will do precisely that, as indeed do we and all intelligent animals. We predict the future of our model of reality all the time. Not perfectly of course, but well enough to the purpose at hand. The proof? Animals survive.

As soon as we figure out an algorithm to accomplish a task – once we understand the algorithm – we consider that computation, or automation, and stop considering that as “intelligence”.

Then you do miss the point. We already know how to turn algorithms into computer code. What we don’t have is the algorithm that allows a crow to solve a problem with pebbles and water without being explicitly programmed to do so. We don’t know the algorithm for how to “figure out an algorithm.”

And that’s not wrong or bizzare – it’s as simple as the defition of “infinity”: for any distance one has reached, infinity lies farther away from it. I believe that is the rigorous mathematical definition of infinity. The definition of AGI cannot be much different … for any clever computer algorithm humanly designed/created, AGI should do better than that.

[This is off topic but no, that isn’t the definition of infinity. See https://en.wikipedia.org/wiki/Infinity.]

The crow (corvids in general) example is excellent and apparently the only other animal to demonstrate this type of problem solving is an orangutan. What is significant, to my mind, is the fact that this is sort of a tricky problem and when observed in a bird it appears almost miraculous until you think about it. First, the birds only do this in a laboratory setting, not the wild. The researchers argue that the reason they do not do it in the wild is because they don’t have to—and as Aesop put it “necessity is the mother of invention.” In the experiments done by a group at Cambridge, the task was not to drink water but to get at a worm floating on the surface.

Ignoring the orangutan, note that the bird is constrained by its beak. It can’t tilt the bottle to get the worm closer to the neck or knock it over such that the worm flows out. I wonder if they tried this; i.e., not fixing the tube to a base, would the bird continue to use stones when a simpler solution existed? No matter, the bird is fully aware of how far its beak reaches and this is beginning to sound like the finger and the cup thing. So the bird knows that it cannot reach the worm and cannot knock the tube over.

Now I must point out that this problem is an example of tool use. Many bird species, not just corvids, are adept at using tools (33 bird families). One example is the use of a suitable stick to get at a grub that is burrowed into a tree. In the current example, the bird needs to decrease the distance between the end of its beak and a worm. It looks around, oh wow, the human has placed a pile of stones here. It takes a stone and drops it into the tube, sort of like the stick going into a hole. Wow, the worm is closer, let’s try another stone. Later, it will learn that a larger stone makes the process faster.

Now as human observers, this just appears incredible and we can’t imagine how we could build a machine to a level of intelligence to solve this same problem. If we could give a machine our level of intellect, it would have no problem solving this, just like we do. That is not what is happening here with the crow. The problem is already set up to be solved. Remember the ‘problem solving steps’ back in Mrs. Periwinkles 6th grade math class? That’s what we have here, but the bird did not do the set-up, evolution did. I will argue that this is why a bird with a very small brain can do what appears to require a lot more processing oomph.

1 Like

Uhhh… Appreciate your appreciation, but I think I already addressed your points. I think the problem is that you, and most mathematicians, think math-first, while I think problem-definition-in-terms-of-real-input- first. Real, unprocessed input is dense array of sensory samples, such as video. Which is almost always plenty compressible. That’s vs. whatever functions you may use to process it.
Big disconnect here.

Math has nothing to do with this specific universe. It’s just a body of computational short-cuts, to be used on whatever inputs that fit their argument format.

Yeah, what’s your goal and purpose? You are not defining anything here.

This is very simple, you don’t need “concepts” or a set theory to understand it.
Cluster is a set of elements, they are defined relative to each other, no need for larger context. You can think of cluster as active synapses in SDR or the whole neuron. Range hierarchy expands the number of potential elements in the cluster, while compositional hierarchy uses lower-level clusters (such as outputs of lower neurons) as potential elements in higher-level clusters.

If you really believe this then we have no common ground. Problem solving, learning territory, exploiting novel food sources, curiosity, theory of mind, tool use etc in animals are commonplace. This is intelligence operating in the physical world and it has high survival value.

Here in Australia we have 4 large smart birds: ravens/crows, magpies, currawongs and cockatoos at home in a suburban environment which they exploit effectively, but obviously never evolved to that role. Watch them as we do and it’s quickly obvious the kinds of strategies they apply. Dogs and cats, possums and rats have all learned to live with people: learned, not evolved.

Scientists study this behaviour in the lab for a good and proper reason: good science. Controlled conditions, repeatable results. Animals solve lab problems and suburban problems and in-the-wild problems using exactly the same intelligence and for the same reasons.

And BTW depending how you measure, a corvid has a brain size comparable to a chimp.

I have a robot that has learned to live with me. It recognizes me, says my name (very cute I might add when it says ‘Professor’ using its robotic voice). It has mapped its play space, manipulates toys and other objects placed in the pen, and automatically returns to its charger as needed. So, as far as Corvid intellect, we are there now. AGI? Hardly.

There are two ‘levels’, for lack of a better term, to AGI. The first is chimp AGI, which we do not have, and the second is AGI mediated by Consciousness, which we are a good deal off from.

1 Like

I don’t think that it is controversial to say that the brain has a variety of structures that interact to form general intelligence. As has been pointed out elsewhere in this forum, several types of memory exist, and while the cortex seems to have much the same structure over the entire sheet, many types of processing nodes can be found in the subcortical structures.

It seems to me that the subcortical structures orient the processing fabric of the cortex to access memory to learn and to solve problems.

What we call consciousness is the subjective experience of having the subcortex push and prod the cortex to experience the world via “attention” and chains of selective recall. In William James " The Principles of Psychology" in 1918 he describes this chaining of conscious attention. See figures 41, 42, and 43. Please consider the James text as describing the evolution of the contents of the “global workspace” model; in particular, the “Dehaene–Changeux model” implementation.

This direction of attention to search solution spaces by navigating higher dimension manifolds of information is a big part of how General Intelligence works and you will have to emulate what the subcortex does if you have any hope of creating AGI.

I am working to understand what mechanisms the subcortex uses to perform these tasks. Yes, the big cortex (HTM/TBT) does the heavy lifting but the subcortex does the driving.

2 Likes

I have a robot that has learned to live with me. It recognizes me, says my name (very cute I might add when it says ‘Professor’ using its robotic voice). It has mapped its play space, manipulates toys and other objects placed in the pen, and automatically returns to its charger as needed. So, as far as Corvid intellect, we are there now. AGI? Hardly.

You robot executes a series of algorithms that have been laboriously hard-coded into its program. So does a self-driving car. These algorithms are brittle: small deviations from expected conditions cause them to fail catastrophically. Your robot runs out of charge unable to bypass a small obstacle. The car can avoid other cars but runs over a bicycle. Animals do far better: survival depends on it.

There are two ‘levels’, for lack of a better term, to AGI. The first is chimp AGI, which we do not have, and the second is AGI mediated by Consciousness, which we are a good deal off from.

Too simplistic. There are many levels of intelligence above your robot and below human that animals display and that we cannot reproduce. Any of these qualifies as AGI.given our current state of knowledge.

The challenge (I repeat) is to describe a task (or a series of tasks) to encapsulate that difference.

This is how it is going to go down. Someone is going to ‘do it’ and that that time, and only at that time, we’ll know which of us was correct. Until then, the banter, which I quite enjoy BTW, will continue.

1 Like

It must be a surprise that some (or should I say many) animals make stupid decisions instead of intelligent ones, and go into extinction.

Survivorship bias … those that survived seem to have made intelligent decisions… Talk about shoot first then draw the target :slight_smile:

Every animal that has gone extinct managed to survive and reproduce for thousands or millions of generations before the environment changed enough that it could no longer adapt. They made no ‘poor decisions’, they just didn’t out-compete some other animals (maybe us), or their environment.

Every extinct species was a survivor first. There is good reason to believe that intelligence was the critical advantage for mammals and corvids (at least), to survive and to favour further evolution of those traits, even if some of them later did not.

It would be nice to make the reasoning more serious, or rigorous, or scientific.

Beside mammals, insects survive, worms survive, bacteria/viruses/fungi survive… what makes critical the contribution of “intelligence” to survial, or not?

1 Like

It’s a reasonable question, but I did choose my words carefully. There is no hard science to justify why one evolutionary path is taken instead of another, but there is logic to it.

Organisms will always survive as long as they can find their preferred environment (greater fitness out-competes them). But if it changes, the survival options are:

  1. Evolution: new genes selected by greater fitness, timescale of 1000+ generations
  2. Adaptation: existing genes selected by greater fitness, timescale of 10+ generations
  3. Intelligence: new learned behaviours confer greater fitness, timescale 1 generation or less.

The ravens and magpies that visit us have not evolved this behaviour, they have learned it because they can. A few other birds visit because the environment happens to suit them but the vast majority of birds do not. They are not fit for this environment and are unable to change.

The reasonable/evolved animal chooses an environment to suit its abilities. The unreasonable/intelligent animal learns new behaviours to suit the environment. This acquisition of new learned behaviours is the bit we really don’t know how to do. The Prof’s robot doesn’t either.

…and this, of course, is the $64,000 question. What compounds it is displayed in the rhetoric of this thread. Different animals with different size and structure brains and widely different environments to include laboratory.

Intelligence and languge led to human consciousness. We alone can ‘know’ what each of us is thinking via our superior intercommunication skills. We can mentally simulate time sequences (mental time travel) as well as put ourselves and our group into various scenarios that we can run at lightning speed to evlauate potential outcomes. No other creature on earth has that capability. Couple this to our tool-making ability, our limb dexterity and our mobility and the planet is alterable by us at a speed that evolution cannot touch.

Then there are our machines.

*quietly dropping link to video of a fella experimenting with artificial evolution via 32bit genes, which has links to a repo with his code*

1 Like

I see your perspective. Thanks for the discussion :slight_smile: