I propose that we adopt the term "Large Self-Supervised Models (LSSMs)" as a replacement for "Foundation Models" and "LLMs". "LLMs" don't capture non-linguistic data and "Foundation Models" is too grandiose. Thoughts? @percyliang
@tdietterich @percyliang I find "ViT" pretty ok. So, instead of LLM, maybe LaT (Language Transformer) would have been a better term in hindsight. I think the L in LLM is too ambiguous. But yeah, Foundation Model is the worst and way to hype-y. Mark my words -- I won't use it in any professional articles
@tdietterich @percyliang For LSSM: it's not bad, but why not just saying "Transformer model" to refer to models with a transformer-like architecture and training procedure? I don't think it's necessary to introduce a new term.
@rasbt @percyliang Don’t you think there will be improvements beyond transformers? I prefer a purely functional description
@rasbt @tdietterich @percyliang What "ViT" expands to 🤔
@rasbt @tdietterich @percyliang Ends up, we go back to the root - attention. So I suggest LAM: Large Attention Model. 🤣
@rasbt @tdietterich @percyliang we're using "transformer model" in our classes.