Workpiece Locating

Before using this tutorial, you should have created a Mech-Vision solution using the “General Workpiece Recognition” case project in the “Robot Communication Setup” section.

In this tutorial, you will first learn the project workflow, and then deploy the project by adjusting the Step parameters to recognize the workpieces’ poses and output the vision result.

Video Tutorial: Workpiece Locating
In this project, every time the Mech-Vision is run, a vision point will be output.

Introduction to the Project Workflow

The following table describes each Step in the project workflow.

No. Phase Step Image Description

1

Capture images

Capture Images from Camera

project build understand step function 1

Connect to the camera and capture images

2

Recognize workpieces

3D Workpiece Recognition

project build understand step function 2

Use 3D matching algorithms to calculate the workpieces’ poses (as pick points)

3

Adjust poses

Adjust poses

project build understand step function 3

Transform the pick points from the camera reference frame to the robot reference frame

4

Output the vision result

Procedure Out

project build understand step function 4

Output the workpieces’ poses for the robot to pick

A pick point refers to a point on the workpiece on which the robot can pick the object.

Adjust Step Parameters

In this section, you will deploy the project by adjusting the parameters of each Step.

Capture Images from Camera

You should adjust the parameters of the “Capture Images from Camera” Step to connect to the camera.

  1. Select the “Capture Images from Camera” Step, and click Select camera on the Step parameters tab.

    project build click select camera
  2. In the pop-up window, click image on the right of the camera serial No. to connect the camera. After the camera is connected, the icon image will turn into image.

    image

    After the camera is connected, select the parameter group. Click Select from and select the displayed parameter group.

    image
  3. After the camera is connected and the parameter group is selected, the calibration parameter group, IP address, and ports of the camera will be obtained automatically. Just keep the default settings of the other parameters.

    image

Now, you have connected the software to the camera.

3D Workpiece Recognition

The “3D Workpiece Recognition” Step has integrated a 3D Workpiece Recognition visualized configurator, which provides point cloud preprocessing, model-based matching, and pose (pick point) calculation.

Select the “3D Workpiece Recognition” Step, and click Open an Editor on the Step parameters tab.

project build open 3d workpiece recognition visual configuration tool

The 3D Workpiece Recognition visualized configurator is shown below.

project build check tool interface

Then follow the procedure to recognize workpieces.

project build 3d workpiece recognition workflow

Select Workpiece

After entering the 3D Workpiece Recognition visualized configurator, you need to make the point cloud model for the workpieces to recognize.

  1. Open the Model Editor.

    At the upper right corner of the 3D Workpiece Recognition visualized configurator, click Select workpiece.

    project build click select workpiece

    In the pop-up Workpiece Library window, click Model Editor.

    project build click model editor

    The following figure shows the Model Editor.

    project build model editor interface
  2. Generate the point cloud model by capturing point clouds.

    1. Capture the depth map of the object.

      Click the Capture point cloud button in the start screen, and select Capture point cloud in the pop-up window.

      project build generate point cloud model 1

      Due to the complex surface features of the track link, it is recommended to create a surface point cloud model for this workpiece. Therefore, clear the Use edge point cloud checkbox. Then click the Capture object button to capture the depth map of the target object.

      project build generate point cloud model 2

      The captured depth map of the target object and background is shown as below.

      project build generate point cloud model 3
    2. Capture background.

      In the upper-right corner, click the Remove background button.

      project build generate point cloud model 4

      Remove the target object in the camera’s field of view first, and click Capture background again to capture a depth map of the background.

      project build generate point cloud model 5

      The depth map of the background is shown as below. Then, click the Next button in the upper-right corner.

      project build generate point cloud model 7
    3. Subtract the background.

      Click Remove background to generate a point cloud model of the target object, as shown below.

      project build generate point cloud model 8

      Click Finish to import the point cloud model of the target object to Matching Model and Pick Point Editor.

      project build generate point cloud model 9
  3. Edit the point cloud model.

    A generated point cloud model may not meet the actual requirement. In such case, you need to edit the model, including removing outliers and downsampling the point cloud.

    1. Remove unwanted points.

      Click the project build edit point cloud model select icon button, select unwanted points to remove, and then click the project build edit point cloud model delete icon button to remove selected points.

      As the figure above, selected points are unwanted points and can be removed by following this step.

      project build edit point cloud model delete point
    2. Downsample the point cloud.

      Point cloud downsampling aims to reduce the number of points in the point cloud model, thus improving model matching efficiency.

      Click the project build edit point cloud model down sample icon button and set the sampling interval in the pop-up window.

      project build edit point cloud model down sample

      In the figure below, the left image is a point cloud model before downsampling, and the right one is after downsampling with a sampling interval of 3 mm.

      project build edit point cloud model down sample result
  4. Add a pick point.

    Click the project build add pose icon button on the toolbar to add a pose as a pick point to the point cloud model of the workpiece.

    project build click add pose

    The following figure shows the added pick point.

    project build check pose
  5. Save the model and the pick point.

    Close the Matching Model and Pick Point Editor, and click Yes in the pop-up window.

    project build save model and pose
  6. Select this workpiece from the Workpiece Library.

    After closing the Matching Model and Pick Point Editor, select the saved point cloud model of the workpiece, and click OK.

    project build select workpiece

    Subsequently, the target workpiece to recognize is displayed in the upper-right corner of the 3D Workpiece Recognition visualized configurator.

    project build workpiece select result

Now, you have selected the workpiece. Click Next on the bottom of the 3D Workpiece Recognition visualized configurator.

project build click next step 1

Preprocessing

Preprocessing is used to set an effective region for recognition to exclude the point cloud of unnecessary parts and keep only the point cloud of the workpiece, thus improving recognition efficiency.

The following figure displays the Preprocessing interface.

project build preprocess interface
  1. Set the region for recognition.

    Click Settings.

    project build click set 3d roi

    In visualized interface, set the region for recognition (3D ROI). Press and hold the Ctrl key, select the vertices of the 3D ROI, and drag the 3D ROI to the proper size. The following figure displays the set 3D ROI.

    project build set 3d roi
  2. Save the region for recognition.

    Click Save and apply to save the region for recognition.

    project build click save and use

Now, you have finished the preprocessing procedure. Click Next at the bottom of the 3D Workpiece Recognition visualized configurator to enter the recognition procedure.

project build click next step 2

Recognize workpieces

In this procedure, you can adjust the 3D matching related parameters in a visualized manner, and output the workpieces’ poses.

The following figure displays the Recognition interface.

project build recognize workpiece interface
  1. Since the robot needs to pick and place one workpiece based on the returned vision result in this tutorial, change the Output count upper limit parameter to 1 for this project.

    project build set output number
  2. View the visualized output result.

    Click Run Step.

    project build click run step

    You can view the visualized output result in the visualized area. As the figure below, the pose of one workpiece is output.

    project build check recognize workpiece result
  3. Save the configuration.

    Click the Finish button at the bottom of the 3D Workpiece Recognition visualized configurator.

    project build click finish

    In the pop-up window, click Save.

    project build click save

    Now, you have recognized the workpiece and calculated its pick point.

Adjust poses

The pick points output by the “3D Workpiece Recognition” Step is in the camera reference frame. To facilitate robot picking, you need to adjust the workpieces’ poses to transform them from the camera reference frame to the robot reference frame.

  1. Open the pose editor.

    Select the “Adjust Poses” Step, and click Open pose editor on the Step Parameters tab.

    project build click open pose editor

    The following figure shows the pose editor.

    project build pose editor interface
  2. Adjust the reference frame.

    In the upper-left corner of the pose editor, set Effect to Camera_To_Robot.

    project build set transform type
  3. View the reference frame transformation result.

    At the bottom of the pose editor, click the Run button.

    project build click pose editor run

    You can see the transformed pick points in the Scene Viewer of the pose editor.

    project build transform pose
  4. Save the configuration.

    Close the pose editor, and click Save in the pop-up window.

    project build save pose editor set

Now, the reference frame of the pick points has been transformed.

Procedure Out

The “Procedure Out” Step sends the results of the current project to the backend service.

Up to now, you have deployed the Mech-Vision project.

We Value Your Privacy

We use cookies to provide you with the best possible experience on our website. By continuing to use the site, you acknowledge that you agree to the use of cookies. If you decline, a single cookie will be used to ensure you're not tracked or remembered when you visit this website.