advstra
advstra t1_iw6lyol wrote
Reply to comment by sam__izdat in Relative representations enable zero-shot latent space communication by 51616
Yeah I got lost a bit there but I think that part is them trying to find a metaphor for what they were saying in the first half, before the "for example". I thought essentially they were suggesting Diffie Hellman key exchange can help with multimodal or otherwise incompatible training data, instead of tokenizers (or feature fusion), I'm not sure how they're suggesting to implement that though.
advstra t1_iw4in0h wrote
Reply to comment by TheLastVegan in Relative representations enable zero-shot latent space communication by 51616
People are making fun of you but this is exactly how CS papers sound (literally the first sentence of the abstract: Neural networks embed the geometric structure of a data manifold lying in a high-dimensional space into latent representations.). And from what I could understand more or less you actually weren't that far off?
advstra t1_iuftoy7 wrote
Reply to King, Me, Coloured pencils, 2022. by artbykabirhirani
That is amazing shading skills. The grid though? Wow.
advstra t1_iw6n49p wrote
Reply to comment by TheLastVegan in Relative representations enable zero-shot latent space communication by 51616
So in the paper from a quick skim read they're suggesting a new method for data representation (pairwise similarities), and you suggest adding style vectors (which is another representation method essentially as far as I know) can improve it for multimodal tasks? I think that makes sense, reminds me of contextual word embeddings if I didn't misunderstand anything.