I dont prefer overlap over jaccard, just thinking if there could be some advantage .
I’m thinking of a different task more along the lines of distinguishing/organizing many/millions vectors with variable sparsity say from 20/100_000 to 200/100_000 (not exactly SDR, may be not brain related ).
with pure overlap i think will be hard to find clumps of vectors … with Jaccard you use all the bits ! So you can say have targeted merge to compress/unionize selectively
As we know every bit has semantic meaning overlap only account for similarities, but if say A have the same overlap with B and C, but B has more bits than C … it would seem that A should be more similar to C, because it has less missing features !
Something like long term memory, where such variability is possible !