[ot][spam][crazy][data] transformer model 'attention' improvement

Undiscussed Horrific Abuse, Victim & Survivor of gmkarl at gmail.com
Fri Jan 28 23:59:13 PST 2022


github/xloem/transformers repo, using return_weights branch of
github/xloem/memory-efficient/attention repo

commit e72d4a3536d1799fc40a85d83a7999e8a39563fc (HEAD ->
memory-efficient-attention, xloem/memory-efficient-attention)
Author: xloem <0xloem at gmail.com>
Date:   Sat Jan 29 07:58:00 2022 +0000

    more bugfixes. output differs, makes sense to step through in
parallel again and compare.


More information about the cypherpunks mailing list