Perceiver IO: a scalable, fully-attentional model that wo... | Perceiver IO: a scalable, fully-attentional model that wo...
Perceiver IO: a scalable, fully-attentional model that works on any modality
TLDR
We've added Perceiver IO to Transformers, the first Transformer-based neural network that works on all kinds of modalities (text, images, audio, video, point clouds,...) and combinations thereof. Take a look at the following Spaces to view some examples:

predicting optical flow between images
classifying images.
We also provide several notebooks.

Below, you can find a technical explanation of the model.