HTM-SP paper

Hello,

I have written two articles with the following titles:

  1. “Information-theoretic analysis of Hierarchical Temporal Memory-Spatial Pooler algorithm with a new upper bound for the standard information bottleneck method,” published in Frontiers in Computational Neuroscience in 2023.
  2. “Performance Comparison of Different HTM-Spatial Pooler Algorithms Based on Information-Theoretic Measures,” published in Neural Processing Letters in 2024.

I kindly invite members of the HTM Forum to visit and read my articles. Your feedback and insights would be greatly appreciated.

Thank you!

7 Likes

Here are the abstracts and DOI Links for the curious:

Information-theoretic analysis of Hierarchical Temporal Memory-Spatial Pooler algorithm with a new upper bound for the standard information bottleneck method

Shiva Sanati, Modjtaba Rouhani and Ghosheh Abed Hodtani (2023)
https://doi.org/10.3389/fncom.2023.1140782

ABSTRACT

Hierarchical Temporal Memory (HTM) is an unsupervised algorithm in machine learning. It models several fundamental neocortical computational principles. Spatial Pooler (SP) is one of the main components of the HTM, which continuously encodes streams of binary input from various layers and regions into sparse distributed representations. In this paper, the goal is to evaluate the sparsification in the SP algorithm from the perspective of information theory by the information bottleneck (IB), Cramer-Rao lower bound, and Fisher information matrix. This paper makes two main contributions. First, we introduce a new upper bound for the standard information bottleneck relation, which we refer to as modified-IB in this paper. This measure is used to evaluate the performance of the SP algorithm in different sparsity levels and various amounts of noise. The MNIST, Fashion-MNIST and NYC-Taxi datasets were fed to the SP algorithm separately. The SP algorithm with learning was found to be resistant to noise. Adding up to
40% noise to the input resulted in no discernible change in the output. Using the probabilistic mapping method and Hidden Markov Model, the sparse SP output representation was reconstructed in the input space. In the modified-IB relation, it is numerically calculated that a lower noise level and a higher sparsity level in the SP algorithm lead to a more effective reconstruction and SP with 2% sparsity produces the best results. Our second contribution is to prove mathematically that more sparsity leads to better performance of the SP algorithm. The data distribution was considered the Cauchy distribution, and the Cramer–Rao lower bound was analyzed to estimate SP’s output at different sparsity levels.


Performance Comparison of Different HTM-Spatial Pooler Algorithms Based on Information-Theoretic Measures

Shiva Sanati, Modjtaba Rouhani, Ghosheh Abed Hodtani (2024)
https://doi.org/10.1007/s11063-024-11546-8

ABSTRACT

Hierarchical temporal memory (HTM) is a promising unsupervised machine-learning algorithm that models key principles of neocortical computation. One of the main components of HTM is the spatial pooler (SP), which encodes binary input streams into sparse distributed representations (SDRs). In this paper, we propose an information-theoretic framework for the performance comparison of HTM-spatial pooler (SP) algorithms, specifically, for quantifying the similarities and differences between sparse distributed representations in SP algorithms. We evaluate SP’s standalone performance, as well as HTM’s overall performance. Our comparison of various SP algorithms using Renyi mutual information, Renyi divergence, and Henze–Penrose divergence measures reveals that the SP algorithm with learning and a logarithmic boosting function yields the most effective and useful data representation. Moreover, the most effective SP algorithm leads to superior HTM results. In addition, we utilize our proposed framework to compare HTM with other state-of-the-art sequential learning algorithms. We illustrate that HTM exhibits superior adaptability to pattern changes over time than long short term memory (LSTM), gated recurrent unit (GRU) and online sequential extreme learning machine (OS-ELM) algorithms. This superiority is evident from the lower Renyi divergence of HTM (0.23) compared to LSTM6000 (0.33), LSTM3000 (0.38), GRU (0.41), and OS-ELM (0.49). HTM also achieved the highest Renyi mutual information value of 0.79, outperforming LSTM6000 (0.73), LSTM3000 (0.71), GRU (0.68), and OS-ELM (0.62). These findings not only confirm the numerous advantages of HTM over other sequential learning algorithm, but also demonstrate the effectiveness of our proposed information-theoretic approach as a powerful framework for comparing and evaluating various learning algorithms.

3 Likes

Hello,

I’m glad that you succeded in using the htm.core framework for your reseach. I put a lot of work into that project.
And it’s nice to see more rigourous verification of the HTM algorithm and all of its variations.

I also came up with another form of homeostatic control, and I’d love to see your analysis of it. I wrote about my experiments in more depth here: https://discourse.numenta.org/t/synapse-competition/5957/15?u=dmac My conclusion was as follows:

Normally all of the synapses in the spatial pooler have a weight of either 0 or 1. Instead, you should divide the weight of each synapse by the total number of connected synapses to the postsynaptic cell.

So, instead of comparing cells based on the total number of synaptic inputs, this compares cells based on the fraction of their connected synapses that are active. The range of inputs to a cell is now in the range [0, 1]. Now cells can compete to activate on a level playing field even if they have different numbers of connected synapses, whereas previously cells with more synapses would have an advantage over cells with fewer synapses.

This is a technique that I’ve seen done by other similar NN models.

Sincerely

3 Likes