2 Apr
2022
2 Apr
'22
11:56 a.m.
So, the place where using a little bit of hand-curated data might go here, at least at first, would be the tokenizer. Then we can see if we can use high-confidence logo code areas, to update low-confidence areas. We could plan to pass the confidence through other things, like a logo parser and a heuristic, to improve its accuracy. We need a detokenizer that can produces logo code. It's likely also helpful to unroll the generation code a little bit, so as to see the information that relates to confidence and access more of the places where loss is calculated and backpropagation performed, when training.