[ot][spam][crazy][rambling]

Undiscussed Horrific Abuse, One Victim of Many gmkarl at gmail.com
Sat Apr 23 06:04:13 PDT 2022


quote from newer paper:

 In this study we show that training
stability can be achieved with fewer sacrifices on model effectiveness.
There are other pathways to improving LLM efficiency. One is to augment a
language model
with a retrieval component to fetch external knowledge that is useful to
perform downstream tasks.
So, the size of the language model can be significantly reduced since it
does not need to encode ev-
erything in model parameters (Guu et al., 2020; Khandelwal et al., 2020;
Borgeaud et al., 2021; Gui
et al., 2021; Zhang et al., 2021). With sparse model structures, Mixture of
Experts models (Artetxe
et al., 2021; Fedus et al., 2021; Zuo et al., 2021; Zoph et al., 2022)
adaptively activate a subset of
model parameters (experts) for different inputs during model training and
inference. The METRO
method proposed in this paper is orthogonal to retrieval-augmented models
and sparsely activated
models. Their combination is an interesting future work direction.




This next one is on training a human assistant using feedback and
reinforcement learning. Nice to have somebody actually talk about that.
https://arxiv.org/abs/2204.05862
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: text/html
Size: 1654 bytes
Desc: not available
URL: <https://lists.cpunks.org/pipermail/cypherpunks/attachments/20220423/8942d3cf/attachment.txt>


More information about the cypherpunks mailing list