In this issue:
we discuss Self-Supervised Learning for Language;
we explore XLM-R, one of the most powerful SSL cross-lingual models ever built;
we cover Facebook’s fastText, a library for representation learning in language tasks.
💡 ML Concept of the Day: Self-Supervised Learning for Language
Continuing our series about self-supervised learning (SSL), we would like to cover its applications in language. Without a doubt, natural language processing (NLP) has been the domain in which SSL models have excelled the most. If you think about the entire wave of transformer models (see The Sequence’s series about Transformers), they can all be considered SSL methods in nature. As SSL evolves, many of the lessons from NLP methods have been extrapolated to other domains such as computer vision. However, SSL remains the most advanced area of SSL techniques.