# 根据每个轮廓的类型, 找到对应的焊缝轨迹 paths = [] self.welding_pts_3d = [] for k, info in enumerate(contours_info): c_type = info["type"] shape, path, pts_3d, end_point_2d = self.FindTypePath(pm, info) if path is None: vision_log.error(f"no path in {k} contour") img_show = cv2.drawContours( img_show, [info["box"].astype(int)], 0, (0, 0, 0), 5) continue
时间: 2024-02-14 16:18:37 浏览: 23
这段代码的功能是根据每个轮廓的类型,找到对应的焊缝轨迹,并将轨迹信息保存到 paths 和 self.welding_pts_3d 变量中。
具体来说,首先定义一个空列表 paths 和空列表 self.welding_pts_3d,用于保存轨迹信息。然后遍历 contours_info 中的每个轮廓,取出轮廓类型 c_type 和轮廓的其他信息 info。
接着调用 self.FindTypePath 函数,传入 pm 和 info 作为参数。FindTypePath 函数根据轮廓的类型 c_type,找到对应的焊缝轨迹,并返回轨迹的形状 shape、轨迹的像素坐标 path、轨迹的三维坐标 pts_3d 和轨迹的末尾点的像素坐标 end_point_2d。
如果轨迹不存在(即 path 为 None),则在控制台输出错误信息,并在可视化结果 img_show 上绘制出该轮廓的外接矩形框。如果轨迹存在,则将轨迹信息保存到 paths 和 self.welding_pts_3d 中。最终返回可视化结果 img_show。
相关问题
class PairDataset(BaseDataset): def initialize(self, opt): self.opt = opt self.root = opt.dataroot self.dir_A = os.path.join(opt.dataroot, opt.phase + 'A') self.dir_B = os.path.join(opt.dataroot, opt.phase + 'B') self.A_paths = make_dataset(self.dir_A) self.B_paths = make_dataset(self.dir_B) self.A_paths = sorted(self.A_paths) self.B_paths = sorted(self.B_paths) self.A_size = len(self.A_paths) self.B_size = len(self.B_paths)
这段代码定义了一个名为`PairDataset`的类,该类继承自`BaseDataset`。该类用于处理配对数据集,其中包含两个文件夹A和B,分别存储了配对数据的两个部分。
在`initialize`方法中,首先将传入的参数`opt`保存在实例变量`self.opt`中。然后,使用`os.path.join`方法将数据根目录`opt.dataroot`与阶段名称`opt.phase`和后缀A或B连接起来,得到文件夹A和B的路径。
接下来,使用`make_dataset`函数获取文件夹A和B中的文件路径,并分别保存在`self.A_paths`和`self.B_paths`中。
为了保证数据的有序性,使用`sorted`函数对文件路径进行排序。
最后,通过获取`self.A_paths`和`self.B_paths`的长度,得到文件夹A和B中的数据数量,并分别保存在`self.A_size`和`self.B_size`中。
修改以下代码使其能够输出模型预测结果: def open_image(self): file_dialog = QFileDialog() file_paths, _ = file_dialog.getOpenFileNames(self, "选择图片", "", "Image Files (*.png *.jpg *.jpeg)") if file_paths: self.display_images(file_paths) def preprocess_images(self, image_paths): data_transform = transforms.Compose([ transforms.CenterCrop(150), transforms.ToTensor(), transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225]) ]) self.current_image_paths = [] images = [] for image_path in image_paths: image = Image.open(image_path) image = data_transform(image) image = torch.unsqueeze(image, dim=0) images.append(image) self.current_image_paths.append(image_path) return images def predict_images(self): if not self.current_image_paths: return for i, image_path in enumerate(self.current_image_paths): image = self.preprocess_image(image_path) output = self.model(image) predicted_class = self.class_dict[output.argmax().item()] self.result_labels[i].setText(f"Predicted Class: {predicted_class}") self.progress_bar.setValue((i+1)*20) def display_images(self, image_paths): for i, image_path in enumerate(image_paths): image = QImage(image_path) image = image.scaled(300, 300, Qt.KeepAspectRatio) if i == 0: self.image_label_1.setPixmap(QPixmap.fromImage(image)) elif i == 1: self.image_label_2.setPixmap(QPixmap.fromImage(image)) elif i == 2: self.image_label_3.setPixmap(QPixmap.fromImage(image)) elif i == 3: self.image_label_4.setPixmap(QPixmap.fromImage(image)) elif i == 4: self.image_label_5.setPixmap(QPixmap.fromImage(image))
请注意,代码中的 preprocess_images 函数和 predict_images 函数中的 self.preprocess_image(image_path) 需要修改为 self.preprocess_images([image_path])。修改后的代码如下:
```
def open_image(self):
file_dialog = QFileDialog()
file_paths, _ = file_dialog.getOpenFileNames(self, "选择图片", "", "Image Files (*.png *.jpg *.jpeg)")
if file_paths:
self.display_images(file_paths)
def preprocess_images(self, image_paths):
data_transform = transforms.Compose([
transforms.CenterCrop(150),
transforms.ToTensor(),
transforms.Normalize([0.485, 0.456, 0.406], [0.229, 0.224, 0.225])
])
self.current_image_paths = []
images = []
for image_path in image_paths:
image = Image.open(image_path)
image = data_transform(image)
image = torch.unsqueeze(image, dim=0)
images.append(image)
self.current_image_paths.append(image_path)
return images
def predict_images(self):
if not self.current_image_paths:
return
for i, image_path in enumerate(self.current_image_paths):
image = self.preprocess_images([image_path])
output = self.model(image[0])
predicted_class = self.class_dict[output.argmax().item()]
self.result_labels[i].setText(f"Predicted Class: {predicted_class}")
self.progress_bar.setValue((i+1)*20)
def display_images(self, image_paths):
for i, image_path in enumerate(image_paths):
image = QImage(image_path)
image = image.scaled(300, 300, Qt.KeepAspectRatio)
if i == 0:
self.image_label_1.setPixmap(QPixmap.fromImage(image))
elif i == 1:
self.image_label_2.setPixmap(QPixmap.fromImage(image))
elif i == 2:
self.image_label_3.setPixmap(QPixmap.fromImage(image))
elif i == 3:
self.image_label_4.setPixmap(QPixmap.fromImage(image))
elif i == 4:
self.image_label_5.setPixmap(QPixmap.fromImage(image))
```