@@ -33,7 +33,7 @@ def train_one_epoch(model, optimizer, data_loader, device, epoch, print_freq, sc
targets = [{k: v.to(device) if isinstance(v, torch.Tensor) else v for k, v in t.items()} for t in targets]
with torch.cuda.amp.autocast(enabled=scaler is not None):
loss_dict = model(images, targets)
- print(f'loss_dict:{loss_dict}')
+ # print(f'loss_dict:{loss_dict}')
losses = sum(loss for loss in loss_dict.values())
# reduce losses over all GPUs for logging purposes
@@ -201,7 +201,7 @@ class MetricLogger:
def collate_fn(batch):
- print(f'batch:{len(batch)}')
+ # print(f'batch:{len(batch)}')
return tuple(zip(*batch))
def collate_fn_wirepoint(batch):