[spam][dreaming of bikesheds, like usual] i'm trying to think on multimodality. new modes are really unfamiliar. vision vs audio vs text. i think there are models now, like the perceiver, that can pick things up pretty fast if there are known data labels. data labels aren't that hard to come up with when everything is in overlapping communities/homeostasis-communication systems. still, it's intersting to consider engaging data of unknown mode. a lot of community and hyper information is likely of unknown mode. i guess that's mostly a problem of having a way to find useful patterns in the data, and a way to use theoretical labels with it. maybe some kind of GAN-like system? if a system formed properties of communities, maybe it could learn to look in new data to identify where these properties were present by highly valuing patterns in patterns. looking for things like communication, part-bounds, repetitive scheduling, responding to nearness, categorisation of parts ...