Respectfully: a meaningful thread cannot exceed a couple of dozen messages.
Now, to memorize a sequence as an SDR a net has to memorize a specific (optimized or not - like bigger or smaller) set of overlapping NOT substrings but subsequences.
Example: abcdefgh is perfectly encoded by abcd, cdef, efgh. Only original sequence will “excite” (hit) those substrings (simpler then subsequences). Given the “sensor” size is 4 and! {abcd} stays excited for 8 more ticks (cder for 4) - moving a sequence against the sensor (that is what sustained spike trains do).
Memorizing a sequence as a set of subsequences is effectively converts input from time domain into frequency domain - if we kind enough to memorize number of occurrences of particular subsequences.
That’s all: I’m experimenting with that conversion for a long time. Creating efficient stochastic machines like LLMs (sentiment, generation, classification) by collecting subsystems and their frequencies .
The key difference from Transformers is that all the nodes/entries of such hierarchical dictionaries associated with particular multimodal patterns. Transformers’ nodes are not. They stay stochastic machines forever. Hierarchical dictionaries as plastic layered networks with labelled nodes enable different kinds of nodes manipulations (thinking as you will). Explain lateral connections and being explainable themselves.
Easy to implement POC, hard to implement an efficient hierarchical dictionaries. Massive creation/deletions of nodes/connections asks for efficient RAM management. Conventional hashtables, trees or sorted arrays are not good enough. So, Python is not an answer, there is a need for something like Garbage Collector, so I do Java.
Even simple/POC implementation memorizes 100k-1m of integer sequences in 100m -10b frequencies space, effectively compares sets/vectors of up to 1m integer components (as opposed to hyperdimencional 20k binary vectors) , implements SDR-associative memory and very much neuro-feasible. Solves [stochastic] MLCS problem on multiple sources (not all MLCS, but many].
I can sure go on. Can explain how intuition/emotions can be modeled by those hierarchical dictionaries.
The question I’m trying to deal now is “what’s next?”
One (some) out-of-system individual(s) creates nice AGI theory and POC. Curiosity is satisfied. There are no single scenario where such a “group of dedicated guppies” can benefit from trying to make their discovery public. An individual vs system problem. The system (academia, investors, companies) will feel disgraced and fight back. Poor individuals
Anyways, use sets of subsequences, count frequencies, inflate when recur deflate with time. That’s easy.