You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
During evals, we currently use lightning.pytorch.callbacks.ModelCheckpoint to save the best model checkpoints during fit. While we only fit the decoder / head on the downstream tasks, the encoder usually remains frozen, so it's not necessary to include it into the saved checkpoints.
Furthermore, for big encoders (e.g. ViT-G), this becomes a major bottleneck in terms of runtime, taking up to 30 seconds to save the checkpoint, while the GPU remains idle:
The text was updated successfully, but these errors were encountered:
During evals, we currently use
lightning.pytorch.callbacks.ModelCheckpoint
to save the best model checkpoints duringfit
. While we only fit the decoder / head on the downstream tasks, the encoder usually remains frozen, so it's not necessary to include it into the saved checkpoints.Furthermore, for big encoders (e.g. ViT-G), this becomes a major bottleneck in terms of runtime, taking up to 30 seconds to save the checkpoint, while the GPU remains idle:
The text was updated successfully, but these errors were encountered: