Set the Pick Points by Jogging the Robot, and Generate the Point Cloud Model Based on the Acquired Point Cloud

You are currently viewing the documentation for the latest version (2.1.0). To access a different version, click the "Switch version" button located in the upper-right corner of the page.

■ If you are not sure which version of the product you are currently using, please feel free to contact Mech-Mind Technical Support.

In this workflow, you can add the pick point to the target object by robot teaching, and then acquire the point cloud to generate the point cloud model and create the target object. This workflow is suitable for scenarios requiring high picking accuracy or where robots or other devices have significant errors.

In the homepage of the target object editor, click Select under the Jog robot and get point cloud workflow, and set the target object name to enter the configuration process. The overall configuration process is shown in the figure below.

overview
  1. Teach the pick point: Add the pick point to the target object by the robot teaching method.

  2. Acquire the point cloud: Use the current project to acquire the point cloud. Then adjust the parameters and set the 3D ROI to generate a point cloud model.

  3. Edit model: Edit the generated point cloud model, including the configurations for the point cloud display and the calibration of the object center point, to ensure better performance of the matching task.

  4. Set collision model (optional): Generate the collision model for collision detection during path planning.

The following sections provide detailed instructions on the configuration.

Teach the Pick Point

  1. Place the target object within the camera’s field of view and ensure that the robot can pick the target object properly.

  2. Use the teach pendant to control the robot to accurately reach the expected pick point of the target object.

  3. It is recommended to set the robot TCP near the center of the point cloud model to reduce picking errors.

  4. Record and enter the robot flange pose and TCP (tool pose relative to the robot flange) at the pick point in the parameter configuration section on the right.

  5. Use the teach pendant to control the robot to move away from the pick point, ensuring that the position of the target object remains unchanged during the departure process.

  6. Collect the point cloud of the target object and create a point cloud model.

After adding the pick point by jogging the robot, all tools configured in Mech-Viz or the path planning tool will be enabled by default, and the picking configuration is not required.

Acquire Point Cloud

Set Project Information

Select the “Capture Images from Camera” Step in the current project to acquire the point cloud. Click Acquire point cloud, and then the result can be viewed in the visualization area.

Configure Point Cloud Display Settings

By configuring Point cloud display settings, you can view different types of point cloud models, i.e., the models of all point clouds, surface point cloud only, and edge point cloud only, for the current target object and edit the point cloud model according to actual requirements. You can also change the display color of the point cloud model and view relevant reference information, such as the pseudo-textured point cloud.

Preprocess Parameters

For detailed explanations of the parameters, refer to Preprocessing Parameters.

If the “3D Target Object Recognition” Step is used in the project, you can enable Use parameters of Step “3D Target Object Recognition”, and then the parameter values in the “3D Target Object Recognition” Step will be synchronized.

Set ROI and Background

By selecting a 3D ROI or capturing an image of the background, you can quickly remove irrelevant point clouds from the scene, enabling rapid extraction of the target object’ point cloud.

If you need to remove the background by capturing the image of the background, you must move the target object out of the camera’s view after acquiring the point cloud. Then click Capture and remove background, and the tool will automatically capture an image of the background and remove the point cloud of the background.

Now the point cloud acquisition has completed. You can click Next to start editing the generated point cloud model.

Edit Point Cloud Model

The generated point cloud model should be edited for better performance in the subsequent matching task.

Check Model Information

Configure Point Cloud Display Settings

By configuring Point cloud display settings, you can view different types of point cloud models, i.e., the models of all point clouds, surface point cloud only, and edge point cloud only, for the current target object and edit the point cloud model according to actual requirements. You can also change the display color of the point cloud model and view relevant reference information, such as that of the pseudo-textured point cloud.

Calibrate Object Center Point

The tool can automatically calculate the object center point. If you need to calibrate the object center point, select the calculating method under Calibrate center point by application, and click Start calculating.

Method Introduction Operation

Re-calculate by using original center point

The default calculation method. Calculate the object center point according to the original object center point and features of the target object.

Calibrate to center of symmetry

Calculate the object center point according to the object’s symmetry.
After calculating the symmetry of the current model, the object center point will be set to the center of symmetry. There may be deviations in the calculation. If deviations occur, please adjust manually.

Calibrate to center of feature

Calculate the object center point according to the selected Feature type and the set 3D ROI.

  1. Select the Feature type according to the geometric features of the object, and the tool will calculate the object center point according to the feature type.

  2. (Optional) Enable Use 3D ROI. After the 3D ROI is set for the target object, the tool will automatically calculate the object center point according to the features of the target object in the 3D ROI.

Configure Point Cloud Model

To better use the point cloud model in the subsequent matching task and enhance matching accuracy, the tool provides the following two options for configuring the point cloud model. You can enable the Configure point cloud model switch as needed.

Avoid False Matches

Once Avoid false matches is enabled, more matching attempts will be made based on the settings to obtain matching results with higher confidence. However, more matching attempts will lead to longer processing time.

Two methods to avoid false matches are offered, i.e., Auto-calculate unlikely poses and Configure symmetry manually. See the table below for more information.

Method Introduction Operation

Auto-calculate unlikely poses

Poses that may cause false matches will be calculated automatically. In subsequent matches, poses that successfully match these poses will be considered unqualified and filtered out.

  1. ClickCalculate unlikely poses to calculate potentially mismatched poses.

  2. If you think a pose in the list will not cause a false match, you can click the × sign to remove it from the list.

Configure symmetry manually

For rotationally symmetric target objects, configuring the rotational symmetry of the point cloud model can prevent the robot’s end tool from unnecessary rotations when it is holding the target object. This increases the success rate of path planning and reduces the time required for path planning, allowing the robot to move more smoothly and swiftly.

Select the symmetry axis by referring to Rotational Symmetry of Target Objects, and then set the Order of symmetry and Angle range.

When Auto-calculate unlikely poses is selected, for the option to take effect in the subsequent matching, you need to configure the relevant parameters in the used matching Steps. See the following for details.

  • If the “3D Matching” Step is used, go to Avoid False Matches  Adjust Poses, and select Filter out unlikely poses in the drop-down menu. This parameter will appear when the parameter tuning level is set to Advanced or Expert.

  • If the “3D Target Object Recognition” Step is used, go to Avoid false matches  Adjust poses in the “Target object selection and recognition” process, and select Filter out unlikely poses. This parameter will appear when the Advanced mode switch is turned on.

Set Weight Template

In the target object recognition process, setting a weight template can enhance the accuracy of matching results by emphasizing the selected features for recognition. Weight templates are typically used to distinguish the orientation of the target object, and the set weight template will affect the matching results. The procedures to set a weight template are as follows.

A weight template can only be set when the Point cloud display settings is set to Display surface point cloud only.

  1. Click Edit template.

  2. In the visualization area, hold and press the right mouse button to select a part of the features on the target object. The selected part, i.e., the weight template, will be assigned more weight in the matching process.

    By holding Shift and the right mouse button together, you can set multiple weighted areas in a single point cloud model.

    set weight template
  3. Click Apply to complete setting the weight template.

For the configured weight template to take effect in the subsequent matching, go to the “Model Settings” parameter of the “3D Matching” Step, and select the model with properly set weight template. Then, go to “Pose Filtering” and turn on the switch Consider Weight in Result Verification. The “Consider Weight in Result Verification” parameter will appear after the “Parameter Tuning Level” is set to Expert.

Make sure you have completed editing the model first.

If the accuracy drift auto-correction feature is not deployed, click the Save button to save the above target object configuration. Or click the Next button to set the collision model.

If the accuracy drift auto-correction feature is deployed, click the Next button to collect data for the drift correction.

Collect Data for Drift Correction

The pick points added by jogging the robot are accurate and reliable. Even when the accuracy drift occurs in the vision system, these pick points will not be affected. In this case, if the drift correction feature in the vision system is enabled, the accuracy of the pick points will be reduced.

If the auto-correction feature has already been deployed, you need to collect data for drift correction after editing the model. The calculated drift compensation will be used for reverse compensation to ensure that the pick points added by jogging the robot remain accurate and feasible.

  1. Auto-correct accuracy drift in EIH vision system or auto-correct accuracy drift in ETH vision system.

    • If more than one day has passed since the last drift correction, rerun the robot program to obtain new drift correction data.

    • If the last drift correction was performed within one day, you may skip the correction process and directly click the Start check button.

  2. Check the program running result.

    After running the robot program, click the Start check button to check the running result of the robot program.

If the check passes, it indicates that valid drift correction data has been generated after running the robot auto-correction program.

If the check fails, it indicates that the robot auto-correction program did not generate valid drift correction data. In this case, it is recommended to rerun the robot auto-correction program to capture the calibration sphere poses.

Once the drift correction data has been collected successfully, you can click the Save button to save the target object configurations. If you want to set the collision model, click Next.

Set Collision Model (Optional)

The collision model is a 3D virtual object used in collision detection for path planning. You can configure the following settings on the collision model according to the actual situation.

Select Method for Generating Collision Model

The tool automatically recommends the collision model generating method based on the current configuration workflow. The recommended method for this case is Use STL model to generate point cloud cube. This tool will generate the point cloud cube based on the imported STL model and conduct collision detection. The collision model generated in this method features high accuracy, while the collision detection speed is lower.

Select STL Model

Click Select STL model and then select the STL model used to generate the point cloud cube.

Configure Point Cloud Cubes

Point cloud cubes are the cubes centered at each point in the point cloud. If these cubes collide with other objects, the point cloud is considered to have collided with these objects.

To perform collision detection, the point cloud is filled with cubes with an edge length of 2 mm. When the target object is too small and its dimensions are close to the edge length of the point cloud cube, the accuracy of the collision model will decrease. When the target object is too large and the collision model includes too many point cloud cubes, the collision detection speed will slow down.

Align Models

Aligning the collision model with the point cloud model of the target object ensures effective collision detection. You can click Auto-align point cloud model and collision model or manually adjust the pose of the collision model to achieve the alignment with the point cloud model of the target object.

Configure Symmetry of Held Target Object

Refer to Rotational Symmetry of Target Objects to select the axis of symmetry, and then set the Order of symmetry and Angle range.

Now, the collision model settings are completed. Click Save to save the target object to Solution folder\resource\workobject_library. Then the target object can be used in subsequent matching Steps.

We Value Your Privacy

We use cookies to provide you with the best possible experience on our website. By continuing to use the site, you acknowledge that you agree to the use of cookies. If you decline, a single cookie will be used to ensure you're not tracked or remembered when you visit this website.