I propose that we adopt the term "Large Self-Supervised Models (LSSMs)" as a replacement for "Foundation Models" and "LLMs". "LLMs" don't capture non-linguistic data and "Foundation Models" is too grandiose. Thoughts? @percyliang
@tdietterich @percyliang Ok but LLMs can still be used when they are trained on linguistic data. I would just eliminate the wording “Foundation Models”.
@frossi_t @percyliang Agreed. I wanted a term that could encompass models trained on multiple modalities: language + video + physical manipulation, etc.
@tdietterich @frossi_t @percyliang For image, language and video, VLMs feels nice to me. For more modalities I'm not sure, maybe perceiver had the right idea?
@tdietterich @frossi_t @percyliang Are these modalities and capabilities intended for "Foundation Models" though?
@tdietterich @frossi_t @percyliang Multitasked Multimedia Art Model