train——line_rcnn.py 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309
  1. # 2025/2/9
  2. import os
  3. import numpy as np
  4. import torch
  5. from models.config.config_tool import read_yaml
  6. from models.line_detect.dataset_LD import WirePointDataset
  7. from tools import utils
  8. from torch.utils.tensorboard import SummaryWriter
  9. import matplotlib as mpl
  10. from models.line_detect.line_net import linenet_resnet50_fpn
  11. from torchvision.utils import draw_bounding_boxes
  12. from models.wirenet.postprocess import postprocess
  13. from torchvision import transforms
  14. from PIL import Image
  15. from models.line_detect.postprocess import box_line_, show_
  16. import matplotlib.pyplot as plt
  17. device = torch.device('cuda' if torch.cuda.is_available() else 'cpu')
  18. def _loss(losses):
  19. total_loss = 0
  20. for i in losses.keys():
  21. if i != "loss_wirepoint":
  22. total_loss += losses[i]
  23. else:
  24. loss_labels = losses[i]["losses"]
  25. loss_labels_k = list(loss_labels[0].keys())
  26. for j, name in enumerate(loss_labels_k):
  27. loss = loss_labels[0][name].mean()
  28. total_loss += loss
  29. return total_loss
  30. cmap = plt.get_cmap("jet")
  31. norm = mpl.colors.Normalize(vmin=0.4, vmax=1.0)
  32. sm = plt.cm.ScalarMappable(cmap=cmap, norm=norm)
  33. sm.set_array([])
  34. def c(x):
  35. return sm.to_rgba(x)
  36. def imshow(im):
  37. plt.close()
  38. plt.tight_layout()
  39. plt.imshow(im)
  40. plt.colorbar(sm, fraction=0.046)
  41. plt.xlim([0, im.shape[0]])
  42. plt.ylim([im.shape[0], 0])
  43. def show_line(img, pred, epoch, writer):
  44. im = img.permute(1, 2, 0)
  45. writer.add_image("ori", im, epoch, dataformats="HWC")
  46. boxed_image = draw_bounding_boxes((img * 255).to(torch.uint8), pred[0]["boxes"],
  47. colors="yellow", width=1)
  48. writer.add_image("boxes", boxed_image.permute(1, 2, 0), epoch, dataformats="HWC")
  49. PLTOPTS = {"color": "#33FFFF", "s": 15, "edgecolors": "none", "zorder": 5}
  50. # print(f'pred[1]:{pred[1]}')
  51. H = pred[-1]['wires']
  52. lines = H["lines"][0].cpu().numpy() / 128 * im.shape[:2]
  53. scores = H["score"][0].cpu().numpy()
  54. for i in range(1, len(lines)):
  55. if (lines[i] == lines[0]).all():
  56. lines = lines[:i]
  57. scores = scores[:i]
  58. break
  59. # postprocess lines to remove overlapped lines
  60. diag = (im.shape[0] ** 2 + im.shape[1] ** 2) ** 0.5
  61. nlines, nscores = postprocess(lines, scores, diag * 0.01, 0, False)
  62. for i, t in enumerate([0.85]):
  63. plt.gca().set_axis_off()
  64. plt.subplots_adjust(top=1, bottom=0, right=1, left=0, hspace=0, wspace=0)
  65. plt.margins(0, 0)
  66. for (a, b), s in zip(nlines, nscores):
  67. if s < t:
  68. continue
  69. plt.plot([a[1], b[1]], [a[0], b[0]], c=c(s), linewidth=2, zorder=s)
  70. plt.scatter(a[1], a[0], **PLTOPTS)
  71. plt.scatter(b[1], b[0], **PLTOPTS)
  72. plt.gca().xaxis.set_major_locator(plt.NullLocator())
  73. plt.gca().yaxis.set_major_locator(plt.NullLocator())
  74. plt.imshow(im)
  75. plt.tight_layout()
  76. fig = plt.gcf()
  77. fig.canvas.draw()
  78. image_from_plot = np.frombuffer(fig.canvas.tostring_rgb(), dtype=np.uint8).reshape(
  79. fig.canvas.get_width_height()[::-1] + (3,))
  80. plt.close()
  81. img2 = transforms.ToTensor()(image_from_plot)
  82. writer.add_image("output", img2, epoch)
  83. def save_best_model(model, save_path, epoch, current_loss, best_loss, optimizer=None):
  84. os.makedirs(os.path.dirname(save_path), exist_ok=True)
  85. if current_loss < best_loss:
  86. checkpoint = {
  87. 'epoch': epoch,
  88. 'model_state_dict': model.state_dict(),
  89. 'loss': current_loss
  90. }
  91. if optimizer is not None:
  92. checkpoint['optimizer_state_dict'] = optimizer.state_dict()
  93. torch.save(checkpoint, save_path)
  94. print(f"Saved best model at epoch {epoch} with loss {current_loss:.4f}")
  95. return current_loss
  96. return best_loss
  97. def save_latest_model(model, save_path, epoch, optimizer=None):
  98. os.makedirs(os.path.dirname(save_path), exist_ok=True)
  99. checkpoint = {
  100. 'epoch': epoch,
  101. 'model_state_dict': model.state_dict(),
  102. }
  103. if optimizer is not None:
  104. checkpoint['optimizer_state_dict'] = optimizer.state_dict()
  105. torch.save(checkpoint, save_path)
  106. def load_best_model(model, optimizer, save_path, device):
  107. if os.path.exists(save_path):
  108. checkpoint = torch.load(save_path, map_location=device)
  109. model.load_state_dict(checkpoint['model_state_dict'])
  110. if optimizer is not None:
  111. optimizer.load_state_dict(checkpoint['optimizer_state_dict'])
  112. epoch = checkpoint['epoch']
  113. loss = checkpoint['loss']
  114. print(f"Loaded best model from {save_path} at epoch {epoch} with loss {loss:.4f}")
  115. else:
  116. print(f"No saved model found at {save_path}")
  117. return model, optimizer
  118. def predict(self, img, show_boxes=True, show_keypoint=True, show_line=True, save=False, save_path=None):
  119. self.load_weight('weights/best.pt')
  120. self.__model.eval()
  121. if isinstance(img, str):
  122. img = Image.open(img).convert("RGB")
  123. # 预处理图像
  124. img_tensor = self.transforms(img)
  125. with torch.no_grad():
  126. predictions = self.__model([img_tensor])
  127. # 后处理预测结果
  128. boxes = predictions[0]['boxes'].cpu().numpy()
  129. keypoints = predictions[0]['keypoints'].cpu().numpy()
  130. # 可视化预测结果
  131. if show_boxes or show_keypoint or show_line or save:
  132. fig, ax = plt.subplots(figsize=(10, 10))
  133. ax.imshow(np.array(img))
  134. if show_boxes:
  135. for box in boxes:
  136. x0, y0, x1, y1 = box
  137. ax.add_patch(plt.Rectangle((x0, y0), x1 - x0, y1 - y0, fill=False, edgecolor='yellow', linewidth=1))
  138. for (a, b) in keypoints:
  139. if show_keypoint:
  140. ax.scatter(a[0], a[1], c='c', s=2)
  141. ax.scatter(b[0], b[1], c='c', s=2)
  142. if show_line:
  143. ax.plot([a[0], b[0]], [a[1], b[1]], c='red', linewidth=1)
  144. if show_boxes or show_keypoint or show_line:
  145. plt.show()
  146. if save:
  147. fig.savefig(save_path)
  148. print(f"Prediction saved to {save_path}")
  149. plt.close(fig)
  150. if __name__ == '__main__':
  151. cfg = r'./config/wireframe.yaml'
  152. cfg = read_yaml(cfg)
  153. print(f'cfg:{cfg}')
  154. print(cfg['model']['n_dyn_negl'])
  155. # net = WirepointPredictor()
  156. dataset_train = WirePointDataset(dataset_path=cfg['io']['datadir'], dataset_type='train')
  157. train_sampler = torch.utils.data.RandomSampler(dataset_train)
  158. # test_sampler = torch.utils.data.SequentialSampler(dataset_test)
  159. train_batch_sampler = torch.utils.data.BatchSampler(train_sampler, batch_size=2, drop_last=True)
  160. train_collate_fn = utils.collate_fn_wirepoint
  161. data_loader_train = torch.utils.data.DataLoader(
  162. dataset_train, batch_sampler=train_batch_sampler, num_workers=8, collate_fn=train_collate_fn
  163. )
  164. dataset_val = WirePointDataset(dataset_path=cfg['io']['datadir'], dataset_type='val')
  165. val_sampler = torch.utils.data.RandomSampler(dataset_val)
  166. # test_sampler = torch.utils.data.SequentialSampler(dataset_test)
  167. val_batch_sampler = torch.utils.data.BatchSampler(val_sampler, batch_size=2, drop_last=True)
  168. val_collate_fn = utils.collate_fn_wirepoint
  169. data_loader_val = torch.utils.data.DataLoader(
  170. dataset_val, batch_sampler=val_batch_sampler, num_workers=8, collate_fn=val_collate_fn
  171. )
  172. model = linenet_resnet50_fpn().to(device)
  173. optimizer = torch.optim.Adam(model.parameters(), lr=cfg['optim']['lr'])
  174. writer = SummaryWriter(cfg['io']['logdir'])
  175. # 加载权重
  176. save_path = 'logs/pth/best_model.pth'
  177. model, optimizer = load_best_model(model, optimizer, save_path, device)
  178. logdir_with_pth = os.path.join(cfg['io']['logdir'], 'pth')
  179. os.makedirs(logdir_with_pth, exist_ok=True) # 创建目录(如果不存在)
  180. latest_model_path = os.path.join(logdir_with_pth, 'latest_model.pth')
  181. best_model_path = os.path.join(logdir_with_pth, 'best_model.pth')
  182. global_step = 0
  183. def move_to_device(data, device):
  184. if isinstance(data, (list, tuple)):
  185. return type(data)(move_to_device(item, device) for item in data)
  186. elif isinstance(data, dict):
  187. return {key: move_to_device(value, device) for key, value in data.items()}
  188. elif isinstance(data, torch.Tensor):
  189. return data.to(device)
  190. else:
  191. return data # 对于非张量类型的数据不做任何改变
  192. def writer_loss(writer, losses, epoch):
  193. try:
  194. for key, value in losses.items():
  195. if key == 'loss_wirepoint':
  196. for subdict in losses['loss_wirepoint']['losses']:
  197. for subkey, subvalue in subdict.items():
  198. writer.add_scalar(f'loss/{subkey}',
  199. subvalue.item() if hasattr(subvalue, 'item') else subvalue,
  200. epoch)
  201. elif isinstance(value, torch.Tensor):
  202. writer.add_scalar(f'loss/{key}', value.item(), epoch)
  203. except Exception as e:
  204. print(f"TensorBoard logging error: {e}")
  205. for epoch in range(cfg['optim']['max_epoch']):
  206. print(f"epoch:{epoch}")
  207. model.train()
  208. total_train_loss = 0.0
  209. for imgs, targets in data_loader_train:
  210. losses = model(move_to_device(imgs, device), move_to_device(targets, device))
  211. # print(losses)
  212. loss = _loss(losses)
  213. total_train_loss += loss.item()
  214. optimizer.zero_grad()
  215. loss.backward()
  216. optimizer.step()
  217. writer_loss(writer, losses, epoch)
  218. model.eval()
  219. with torch.no_grad():
  220. for batch_idx, (imgs, targets) in enumerate(data_loader_val):
  221. pred = model(move_to_device(imgs, device))
  222. pred_ = box_line_(pred) # 将box与line对应
  223. show_(imgs, pred_, epoch, writer)
  224. if batch_idx == 0:
  225. show_line(imgs[0], pred, epoch, writer)
  226. break
  227. avg_train_loss = total_train_loss / len(data_loader_train)
  228. writer.add_scalar('loss/train', avg_train_loss, epoch)
  229. best_loss = 10000
  230. save_latest_model(
  231. model,
  232. latest_model_path,
  233. epoch,
  234. optimizer
  235. )
  236. best_loss = save_best_model(
  237. model,
  238. best_model_path,
  239. epoch,
  240. avg_train_loss,
  241. best_loss,
  242. optimizer
  243. )