The idea behind the Python code is to assemble the raw algorithms together in such a way where there is no obscuring from infrastructure code. I can see what is going in and coming out of each algorithm, and If you "vet" the code, I can see once and for all what goes into what and what is expected out.
See, the problem is that there aren't any tests that show how things are connected. There are many great tests for the algorithms and then there are (some) integration tests for infrastructure code such as the NetworkEngine, but there is no documentation saying, "this" goes into this and this is what is expected as output/input for this algorithm.
I don't really want to argue whether the above is true or not - if somebody feels that there are plenty of examples of the above then please, point me to them (but please help me out with this request as well)
Summary: I would like to see everything pieced together in an officially sanctioned pseudo region or FauxRegion if you will and then I will do the same with the Java code (using the same data and I will see how long it takes the TM to produce good predictions, I will see when the SP stabilizes, I will see when the classifier reports the expected predictions and I will see the Anomaly code settle in to anomalies of 0.0.) After that I will introduce one or more anomalies in both the Python and Java versions and see exactly how they both handle them - and that should give me an idea of how the Java version is performing and perhaps some insight into whatever is the problem?
I hope that makes sense?
EDIT: Another point. I ask for this because I can reproduce the exact same thing in Java, so there is a one-to-one correspondence between the two and I can play with each and see how they perform. (i.e. I should see the exact same reaction (nearly) in the Java version as I do the Python).