In this example there are 3 spatial features per tilmestep (‘timeOfDayBits’, ‘weekendBits’, ‘consumptionBits’), and they are combined since the SP & TM ultimately see one vector per time step – however many spatial features that vector contains.
So if your sequences are composed of words, there’d be one encoding, SP vector and TM output per word (one after the other). The system will learn the transitions between words. There’s a big limitation to this though! Since there is no encoder for words in NuPIC, each different word would be treated as a distinct category by default – so the systems sees no similarity between words.
To my knowledge this issue has been most robustly handled by Cortical.io – which has basically created a method for generating SDRs from words and even groupings of words. You feed in a word, sentence, paragraph or whatever and get back an SDR – as you normally would from the SP. Then these SDRs are fed into the TM just like any other case.
If you’re processing words as you raw data type I’d highly recommend looking into Cortical.io and their work on this.