[ot][spam][crazy][personal][crazy][crazy][crazy] crazy things

Undescribed Horrific Abuse, One Victim & Survivor of Many gmkarl at gmail.com
Mon Dec 26 16:15:36 PST 2022


1106                        "Using DeepSpeed Optimizer param name {}
as basic optimizer".format(
1107                            self.optimizer_name()))
(Pdb) up
> /home/user/.local/lib/python3.9/site-packages/deepspeed/runtime/engine.py(291)__init__()
-> self._configure_optimizer(optimizer, model_parameters)
(Pdb) list
286             # Configure optimizer and scheduler
287             self.optimizer = None
288             self.basic_optimizer = None
289             self.lr_scheduler = None
290             if model_parameters or optimizer:
291  ->             self._configure_optimizer(optimizer, model_parameters)
292                 self._configure_lr_scheduler(lr_scheduler)
293                 self._report_progress(0)
294             elif self.zero_optimization():
295                 # no optim selected but zero is enabled
296                 self.optimizer =
self._configure_zero_optimizer(optimizer=None)
(Pdb)

i'm learning to run SwissArmyTokenizer, which is a repurposeable
library made by chinese language model researchers. it can load
mainstream language models under a shared generalized transformer
architecture, which is a huge breath of fresh air after huggingface.
unfortunately it looks like they mostly only used it for their
research, the parts they use are the parts that have maintainedness.
By default it assumes you have 8 gpus and write access to a global
folder with path /data/qingsong .


More information about the cypherpunks mailing list