
> > > [so how did they get deepseek running on this thing? the page on it > > > has a link to their OS image ... > > > > https://www.eswincomputing.com/en/bocupload/2024/06/19/17187920991529ene8q.p... > > indicates that there are GPU drivers for normal python-based > > frameworks for this CPU ("Pytorch, Tensorflow, PaddlePaddle, ONNX, > > etc") > > i wonder if the lesson is that if you port a mainstream language model > to a small chip and then sell it, everybody will buy it.
OOPS
not today it seems
there's still some interest in image generation, but it mostly assumes that there's already a local way to do this.
i'm on windows right now, hopefully temporarily. i use wsl2 ubuntu :s :s : s
so maybe network servic--- [becau--
what if we used a super tiny model? maybe that's interesting or music synthesis or something !
pulling away from httptransformer could help sort out a little. it's definitely never been designed for diffusion models
[machine learning in general is designed in opposition to the things people like me try to do, the design choices are based around lots of infrastructure and minimal algorithmic resear
maybe let's try a tiny diffusion model or something the huggingface diffusers architecture seems more flexible than the transformers architecture, they [seem to kind of parameterize their pipelines to load submodels and wire them together, of course it also looks hardcoded into constructor classes,