Hi, currently in pytorch2.0, FSDP model do not support deepcopy, how can I copy a model param as ema and update it?
example, my fsdp model:
sharding_strategy=torch.distributed.fsdp.ShardingStrategy.SHARD_GRAD_OP
model = FSDP(model, sharding_strategy=sharding_strategy,
ignored_parameters = not_trainable, )
I can’t call deepcopy(model)
directly, how should I achive ema?