2 Apr
2022
2 Apr
'22
12:51 p.m.
---- I think I can roughly summarise that if it is, it's not going to be a big return to learn it, since other model parts would train a different way. I'm now interested in brainstorming another approach. Something to replace the tokenizer. We could use the model's existing output to train the other approach, and then finetune it around the desired result. Time to move to the other thread. Overstayed this one a little.