You can do like a transformer, and “learn” which contexts to attend to with “attention”. 10 billion dollars in investment from Microsoft to OpenAI says clearly it works quite well.
Or you can leave all the information in a network, and let the runtime context select what it wants to attend to.
To expand on that second solution…
I think the bigger problem than identifying context to generalize on, and even bigger than the entanglement of generalizations which Coecke etc. have focused on, is the fact that these generalizations actually act like expansions of the data, and there appears to be no limit to them. So any fixed set will always be incomplete, and the only way to completely capture them, is to find the ones relevant to a given situation at run time. I went into this a lot in the “How Your Brain Organizes Information” thread. This post might be a good summary of (the first part of) that:
I gave a lot of examples, from the history of linguistics, maths, and philosophy, to support the idea that meaning can’t be completely abstracted. That in fact “learning”, acts like an expansion of the data. But I thought the data that… @cezar_t presented was also good support - that doing more training over transformers, acts much like just using more data:
So that’s the first option. “Learn” the context to “attend” to. And capture the entangled, quantum, quality of generalizations, by having a black box, where no-one is sure what the structure is.
But I say it runs into the bigger problem, that generalizations expand. They are not only entangled as Coecke etc see, but they get forever bigger. The current solution to that seems to be to just try to have it as big as possible. (Which also means that only the biggest entities can get involved at all, and the little guy is reduced to begging, or paying, to get access to big company APIs.)
The real flaw with that is that no matter how big you make it, it will never be as “big” as human performance, and you’re forever chasing an asymptote:
The second option, one to deal with this expansion, the one I’m now focused on, Is to leave all the information in a network, leave it “embodied” in a set of data, essentially, and let the runtime context select what it wants to attend to. As I said earlier in this thread:
This explanation to @JarvisGoBrr might be a good summary: