It’s so amazingly simple and natural progress but also so powerful and profound of intermediate omnitoken, isnt it?
It’s just interwining multiple single modality/functionality neural networks in their middle layers by intermediate omnitoken with a parallel orthogonal data format, but may create so many possible variants and compositions with encoder and decoder of transformer or with other models or with possible future models, which finally will unlock so much potential and improvements possible.
This architecture of pre-main-post networks on intermediate omnitokens may need huge experiments on iterated training to optimize different configuration of networks including like different ways to combine encoder, yes, it will be definitely an exhausting but interesting job for you guys. haha.
By the way, those inference only or specific chip may need chip level support to this pre-main-post networks on intermediate omnitoken to adapt to different networks configuration and intermediate omnitoken, which may need to improve or change chip design too.
Be First to Comment