I think the fovea-motion stream mentioned here deserves a bit of detail.
Why this matters - because it might feed some experiential meat to HTM theory in order to move further. One that is conceptually simple enough for average software engineers to follow with lesser effort than previous experiments on HTM reinforcement learning.
Here are the core points of this proposal:
- We assume a visual robot is exposed to a picture.
- Unlike other vision-modelling strategies (CNN-s, Transformers, whatever) where the model swallows in the whole picture in order to āmake senseā of it, this one has two important properties which it uses to navigate the picture it is exposed to:
- it can only see a foveic patch of the picture - which means two things:
- the patch image is provided by a low resolution camera.
- this camera is movable in three directions - x, y, z. horizontal, vertical and zoom. What is important is as the camera zooms out it will provide a blurry low res image, so in order to āseeā more detail of an area it needs to close into that respective area
- the second property is it is able to move the camera. Motions of the camera are encoded as a set of three scalars representing either the new absolute position of the camera or a relative one to its current position. (I think the relative motion encoding is more fruitful in the beginning)
- So an (H)TM learning the āworldā is to make a good prediction of the next step on an arbitrary State->Action->State->Action ⦠stream, where āStateā is an SDR encoding what the camera sees, and āActionā is an SDR encoding a camera movement.
Well the above has a few problems:
- if camera xyz movements are random, that might hinder the TMās ability to learn.
- if they are not random, that might hinder TMās ability to predict a state following a new, out of training samples motion.
- most important, that strategy might not be as good at teaching the model what motion should I do next, simply because, in principle, any motion is possible hence valid.
For now Iāll leave these questions here, and follow up soon.