[spam] [personal] perceiver model notes

k gmkarl at gmail.com
Tue Jan 18 07:46:56 PST 2022


Draft summary of PerceiverForMaskedLM:

1. PerceiverTextPreprocessor:
  Inputs -> Embeddings
2A. PerceiverEncoder:
  Embeddings + Weights + Attention mask ->
PerdeiverAttention(is_cross_attention = True) -> Hidden states
2B. PerceiverEncoder:
  Hidden states + Attention mask -> layers of
PerceiverAttention(is_cross_attention = False) -> Hidden states
3. PerceiverBasicDecoder:
  Hidden states + Weights + Attention mask ->
PerceiverAttention(is_cross_attention = True) -> Decoded embeddings
  There's an additional Head mask that may be usable to alter
properties of the model on a per-layer basis.
4. PerceiverEmbeddingDecoder:
  Decoded embeddings -> Log probabilities

This is a helpful summary of perceiver models after looking through
the code to have an idea where to go to engage parts further.


More information about the cypherpunks mailing list