🤘Edge#224: AlexaTM 20B is Amazon’s New Language Super Model Also Capable of Few-Shot Learning
The model is the largest seq2seq architecture capable of few-shot-learning
On Thursdays, we dive deep into one of the freshest research papers or technology frameworks that is worth your attention. Our goal is to keep you up to date with new developments in AI to complement the concepts we debate in other editions of our newsletter.
💥 What’s New in AI: AlexaTM 20B is Amazon’s New Language Super Model Also Capable of Few-Shot Learning
In the last few years, the progress in natural language understanding (NLU) really challenges human imagination. Some of the milestones achieved by models like OpenAI GPT-3 seemed unimaginable just a few years ago. Large AI labs like Microsoft Research, Google Brain, Alexa AI, DeepMind, and Meta AI are regularly pushing the boundaries of NLU research. One of the latest entrances in the language supermodel category came from Amazon’s Alexa AI labs with Alexa Teacher Models™ 20B, a large seq2seq model that set up new marks in few-shot learning.
As its name indicates,