watch()
less than a minute
function wandb.watch
wandb.watch(
models: 'torch.nn.Module | Sequence[torch.nn.Module]',
criterion: 'torch.F | None' = None,
log: "Literal['gradients', 'parameters', 'all'] | None" = 'gradients',
log_freq: 'int' = 1000,
idx: 'int | None' = None,
log_graph: 'bool' = False
) → None
Hook into given PyTorch model to monitor gradients and the model’s computational graph.
This function can track parameters, gradients, or both during training.
Args:
models: A single model or a sequence of models to be monitored.criterion: The loss function being optimized (optional).log: Specifies whether to log “gradients”, “parameters”, or “all”. Set to None to disable logging. (default=“gradients”).log_freq: Frequency (in batches) to log gradients and parameters. (default=1000)idx: Index used when tracking multiple models withwandb.watch. (default=None)log_graph: Whether to log the model’s computational graph. (default=False)
Raises:
ValueError: If wandb.init has not been called or if any of the models are not instances of torch.nn.Module.
Feedback
Was this page helpful?
Glad to hear it! If you have more to say, please let us know.
Sorry to hear that. Please tell us how we can improve.