You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Logging accelerator and CPU memory during training models is a common scenario all practitioners run into. As discussed with @muellerzr over Slack, accelerate should have a utility for that (its location would be here).
Not sure if the torch profiler can be configured to give the information on that aggregation level. But even if it's possible, it should at least be documented, as many users are probably interested in this higher level view.
Logging accelerator and CPU memory during training models is a common scenario all practitioners run into. As discussed with @muellerzr over Slack,
accelerate
should have a utility for that (its location would be here).peft
has a good reference here:https://github.com/huggingface/peft/blob/ae55fdcc5c4830e0f9fb6e56f16555bafca392de/examples/oft_dreambooth/train_dreambooth.py#L421
Opening it up as a feature request in case anyone's interested in contributing this. This would be a massive help, IMO.
The text was updated successfully, but these errors were encountered: