Hey guys, can someone sanity check our model params here? We got these from NAB, before I make an issue in github, are these right?
{
"aggregationInfo":{
"seconds":0,
"fields":[
],
"months":0,
"days":0,
"years":0,
"hours":0,
"microseconds":0,
"weeks":0,
"minutes":0,
"milliseconds":0
},
"model":"HTMPrediction",
"version":1,
"predictAheadTime":null,
"modelParams":{
"sensorParams":{
"sensorAutoReset":null,
"encoders":{
"value":{
"name":"value",
"resolution":0.007692307692307693,
"seed":42,
"fieldname":"value",
"type":"RandomDistributedScalarEncoder"
},
"timestamp_dayOfWeek":null,
"timestamp_timeOfDay":{
"fieldname":"timestamp",
"timeOfDay":[
21,
9.49
],
"type":"DateEncoder",
"name":"timestamp"
},
"timestamp_weekend":null
},
"verbosity":0
},
"anomalyParams":{
"anomalyCacheRecords":null,
"autoDetectThreshold":null,
"autoDetectWaitRecords":5030
},
"spParams":{
"columnCount":2048,
"synPermInactiveDec":0.0005,
"spatialImp":"cpp",
"inputWidth":0,
"spVerbosity":0,
"synPermConnected":0.2,
"synPermActiveInc":0.003,
"potentialPct":0.8,
"numActiveColumnsPerInhArea":40,
"boostStrength":0.0,
"globalInhibition":1,
"seed":1956
},
"trainSPNetOnlyIfRequested":false,
"clParams":{
"alpha":0.035828933612158,
"verbosity":0,
"steps":"1",
"regionName":"SDRClassifierRegion"
},
"tmParams":{
"columnCount":2048,
"activationThreshold":13,
"pamLength":3,
"cellsPerColumn":32,
"permanenceDec":0.1,
"minThreshold":10,
"inputWidth":2048,
"maxSynapsesPerSegment":32,
"outputType":"normal",
"initialPerm":0.21,
"globalDecay":0.0,
"maxAge":0,
"newSynapseCount":20,
"maxSegmentsPerCell":128,
"permanenceInc":0.1,
"temporalImp":"cpp",
"seed":1960,
"verbosity":0
},
"tmEnable":true,
"clEnable":false,
"spEnable":true,
"inferenceType":"TemporalAnomaly"
}
}
Here is the error we are getting:
File "/opt/anomaly/anomalymodel.py", line 198, in process_sample
result = self.htm_model.run({"value": sample.value, "timestamp": sample.timestamp})
File "/usr/local/lib/python2.7/site-packages/nupic/frameworks/opf/htm_prediction_model.py", line 429, in run
self._sensorCompute(inputRecord)
File "/usr/local/lib/python2.7/site-packages/nupic/frameworks/opf/htm_prediction_model.py", line 514, in _sensorCompute
sensor.compute()
File "/usr/local/lib/python2.7/site-packages/nupic/engine/__init__.py", line 433, in compute
return self._region.compute()
File "/usr/local/lib/python2.7/site-packages/nupic/bindings/engine_internal.py", line 1499, in compute
return _engine_internal.Region_compute(self)
File "/usr/local/lib/python2.7/site-packages/nupic/bindings/regions/PyRegion.py", line 184, in guardedCompute
return self.compute(inputs, DictReadOnlyWrapper(outputs))
File "/usr/local/lib/python2.7/site-packages/nupic/regions/record_sensor.py", line 377, in compute
self.encoder.encodeIntoArray(data, outputs["dataOut"])
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/multi.py", line 100, in encodeIntoArray
encoder.encodeIntoArray(self._getInputValue(obj, name), output[offset:])
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 250, in encodeIntoArray
output[self.mapBucketIndexToNonZeroBits(bucketIdx)] = 1
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 233, in mapBucketIndexToNonZeroBits
self._createBucket(index)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
<... many copies of this line ...>
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 267, in _createBucket
self._createBucket(index+1)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 263, in _createBucket
index)
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 299, in _newRepresentation
not self._newRepresentationOK(newRepresentation, newIndex):
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 358, in _newRepresentationOK
if not self._overlapOK(i, newIndex, overlap=runningOverlap):
File "/usr/local/lib/python2.7/site-packages/nupic/encoders/random_distributed_scalar.py", line 402, in _overlapOK
if overlap <= self._maxOverlap:
AttributeError: 'RandomDistributedScalarEncoder' object has no attribute '_maxOverlap'