[ot][spam][crazy][data] transformer model 'attention' improvement
Undiscussed Horrific Abuse, Victim & Survivor of
gmkarl at gmail.com
Sun Jan 30 03:27:49 PST 2022
11:24 now the attentions match :)
now probably will open a pull request with memory-efficient-attention.
plan to descfribe it as a draft looking for comments
commit 0e48ea180a50d201f107c8c69f9493db78cf813b (HEAD ->
memory-efficient-attention, xloem/memory-efficient-attention)
Author: xloem <0xloem at gmail.com>
Date: Sun Jan 30 11:26:31 2022 +0000
fix for incorrect mask shape
More information about the cypherpunks
mailing list