Hey all, I’ve been working on a project that uses HTM.java to identify EMG signals. I’m having a bit of trouble, however, figuring out how to best go about setting up my network. I’ll go ahead and explain in more detail what I’m looking to do, and what my questions are:
I recently purchased a Myo armband, and although it is quite fun, it is pretty limited in its functionality at this point in time. The idea is to slip the armband on and have it make your phone or desktop respond in some way when you perform certain gestures. Its main issue right now is that it only recognizes about 5 or 6 different hand poses (more if you factor in the armband’s gyroscope and accelerometer sensors, but I’m just concerned with the EMG data right now). The developers of the armband released an API that allows you to access a stream of EMG data from the armband via Bluetooth. I figured this would be a perfect application for an HTM - classifying and identifying a stream of spatial and temporal data!
The armband is essentially composed of 8 EMG sensors (they’re called pods). You just slide it up your forearm and wear it near your elbow. So I’m working with up to 8 streams of EMG readings from the band - because the API allows you to access each sensor’s EMG stream individually. This is where my main question comes in: What is the “best” way to set up a network to work with this data? Should I use all the data, or just some of it? How should I feed it into the network?
My current approach has been to set up a triple layer network (with these Parameters
):
// Using the beautiful fluent-style API...
network = Network.create("Network", p)
.add(Network.createRegion("Region 1")
.add(Network.createLayer("Layer 1", p)
.alterParameter(Parameters.KEY.AUTO_CLASSIFY, Boolean.TRUE)
.add(Anomaly.create())
.add(new TemporalMemory())
.add(new SpatialPooler()))
.add(Network.createLayer("Layer 2", p)
.add(new TemporalMemory())
.add(new SpatialPooler()))
.add(Network.createLayer("Layer 3", p)
.add(Sensor.create(ObservableSensor::create, SensorParams.create(
SensorParams.Keys::obs, new Object[] {"name", publisherSupplier}))))
.connect("Layer 1", "Layer 2")
.connect("Layer 2", "Layer 3"));
Here’s what the three header lines for the PublisherSupplier
look like:
Pod1,Pod2,Pod3,Pod4,Pod5,Pod6,Pod7,Pod8
float,float,float,float,float,float,float,float
B,B,B,B,B,B,B,B
Every time step I get a byte[8]
containing the most recent readings from the armband’s EMG stream. I then feed in the EMG data manually via a Publisher
. So, if I got this array:
{-1,-4,4,-7,-5,2,0,3}
I would feed each value into the network on a single line, like so:
manualPublisher.onNext("-1,-4,4,-7,-5,2,0,3");
Every time the network processes a line of input, I add that processing’s actual input, predictions, and anomaly score to a list. After I finish feeding in data, I graph the actual and predicted values for each pod on a line chart. So, here is the chart resulting from training a brand new network for the first time on the EMG readings created by closing and opening my fist:
You’ll notice the average anomaly score is 94% (that is calculated by summing the anomaly score for each line of input and dividing by the number of inputs).
After being trained on that same data 11 more times, the chart looks like this:
The anomaly score is down closer to 70% now. This is about as low as I can seem to get it. I only used data from one fist EMG sample in the above pictures for simplicity’s sake, but I’ll typically take about 10 samples and randomly train the network on them for up to 50 times. I even trained the network 100 times (only did that once, because it took quite a long time

I found a thesis paper (pdf download) online in which the authors did something similar to what I am looking to accomplish. If you’re able, I suggest perusing pages 49-68 of the document to see how they set up their network with NuPIC (the rest of it is mostly discussion of other neural networks and their EMG data capturing device, which appears to be roughly equivalent to one pod of the myo armband).
Here’s a schematic they used to illustrate their setup:
They go much more in depth in the document, but you can get the gist of it from that schematic. As far as I can tell, our networks are pretty similar, except for their category sensor (which as I understand it tells the classifier if it is looking at data from a index finger opening, or closing). I’d like to include a category sensor in my network setup, or something similar, because I’ll need to distinguish between (ideally) many different EMG data streams (e.g. opening and closing a fist, extension and contraction of the fingers, wrist movements, etc…). The only other option that I can see is to have an individual network train on each hand pose and see which network outputs the lowest anomaly score for a particular set of EMG samples. That is what I was going to do until I read their paper, but I think the category sensor approach would be far more efficient.
The problem is that I’m not entirely sure how to implement a category sensor in HTM.java. I thought that I might be able to just add a field to each input line identifying what pose the input is from. For example, an input line from an EMG sample of a fist closing would be:
manualPublisher.onNext("fist-close,-1,-4,4,-7,-5,2,0,3");
whereas an EMG sample of a fist opening would be:
manualPublisher.onNext("fist-open,-1,-4,4,-7,-5,2,0,3");
I have not yet tried out this approach, as I wanted to get some input from the community on my approach before going too much farther. But the main “issue” with this approach would be that the category identifier input would be processed by all stages of the network, rather than just the classifier. Granted, this might not matter much?
The authors of the thesis paper also appear to take a much different approach than I do when putting data into their network. It appears that they take 512 samples from their single EMG sensor and feed that into their network:
So they are essentially taking a section of the graph produced from their EMG sensor’s readings and feeding it into their network. Whereas I am taking a point from each of my 8 EMG sensors’ graphs at the same x-coord and feeding that into my network. If I were to take their approach, then it seems I would need to construct a network 8x as large as theirs - assuming I use the data from all 8 of the armband’s sensors.
I suppose that about sums it up. I’m sure I’ve left out an important detail or two, so please ask if anything doesn’t make sense and I’ll clarify. I’ve set up a GitHub repo for my project just now, so you can see the source. I’ll warn you, it isn’t the cleanest or most well documented code at this stage, as I’m still trying out a lot of different things… It’ll look better once I settle on one approach, which is why I’m posting this in the first place
Again, my main questions are:
- How should I set up my network?
- If it is a good approach, how should I best implement a category sensor?
- How should I go about feeding in the EMG data to my network?
Thanks in advance for any advice or input. I’m hoping some folks with more experience and knowledge in this area can help me out a bit!