Model Training

You are currently viewing the documentation for version 2.5.0. To access documentation for other versions, click the "Switch Version" button located in the upper-right corner of the page.

■ To use the latest version, visit the Mech-Mind Download Center to download it.

■ If you're unsure about the version of the product you are using, please contact Mech-Mind Technical Support for assistance.

You can get started with model training upon labeling. On the parameter bar of Training, you can configure training parameters, train models, and view training information.

Training Parameters

Click Parameter Configuration to open the Training parameter configuration window.

Data Augmentation

The data for training the model needs to contain as much as possible all situations that may actually occur. If the site does not have the corresponding data-collecting conditions, you can adjust the Data Enhancement parameters to prepare data that can not be collected, thus enriching the training data. It must be ensured that the augmented image data should conform to the on-site situation. If there are no rotations on the site, then there is no need to adjust the parameter “Rotation”; otherwise, the model’s performance may be affected.

Hover the mouse cursor over train ask icon to view the adjustment effect of each parameter.
  • Lightness

    It refers to how much light is present in the image. When the on-site lighting changes greatly, by adjusting the brightness range, you can augment the data to have larger variations in brightness.

  • Contrast

    Contradiction in luminance or color. When the objects are not obviously distinct from the background, you can adjust the contrast to make the object features more obvious.

  • Translation

    Add the specified horizontal and vertical offsets to all pixel coordinates of the image. When the positions of on-site objects (such as bins and pallets) move in a large range, by adjusting the translation range, you can augment the data in terms of object positions in images.

  • Rotation

    Rotate an image by a certain angle around a certain point to form a new image. In general, keeping the default parameters can meet the requirements. When the object orientations vary greatly, by adjusting the rotation range, you can augment the image data to have larger variations in object orientations.

  • Zoom

    Shrink or enlarge an image by a certain scale. When object distances from the camera vary greatly, by adjusting the scale range parameter, you can augment the data to have larger variations in object proportions in the images.

  • Flip horizontally

    Flips the image 180° left to right. If the objects to be recognized have left-right symmetry, you can select the Flip horizontally check box.

    Unnamed image
  • Flip vertically

    Flips the image 180° upside down. If the objects to be recognized have vertical symmetry, you can select the Flip vertically check box.

    Unnamed image
  • Label expansion

    Only supported in the Defect Segmentation module. Expand the regions of defects selected in an image by a certain scale. For most scenarios, you do not need to check this option. If the regions of defects are too small, you can select the Label expansion checkbox to avoid influence on training when the defect regions are too small after image scaling.

Training Parameters

  • Input image size

    The pixel-wise height and width of the image input to the neural network for training. It is recommended to use the default setting, but if the objects or defect regions in the images are small, you need to increase the input image size. The larger the image size, the higher the model accuracy, but the lower the training speed.

  • Batch size

    The number of samples selected for each time of neural network training. It is recommended to use the default settings; if you need to increase the training speed, you can appropriately increase the batch size. If the batch size is set too large, memory usage will increase.

  • Model type

    Defect Segmentation

    Normal

    Generally, it is recommended to use Normal mode.

    Enhanced

    You can choose the Enhanced mode when the model effect is not as expected or the accuracy requirement is high. This mode will decrease the training speed.

    Instance Segmentation

    Normal (better with GPU deployment)

    Suitable for scenarios that require high accuracy. This mode may result in slower training speed.

    Lite (better with CPU deployment)

    Suitable for scenarios that require high training speed.

  • Eval. interval

    The number of epochs for each evaluation interval during model training. It is recommended to use the default setting. Increasing the Eval. interval can increase the training speed. The larger the parameter, the faster the training; the smaller the parameter, the slower the training, but a smaller value helps select the optimal model.

  • Epochs

    The total number of epochs of model training. It is recommended to use the default setting. If the features of objects to be recognized are complex, it is necessary to increase the number of training epochs appropriately to improve the model performance, but increasing the number of epochs will lead to longer training time.

    It is not true that the bigger the number of epochs, the better. When the total number of epochs is set to be large, the model will continue to be trained after the accuracy stabilizes, which will result in a longer training time and the risk of overfitting.
  • Learning rate

    The learning rate sets the step length for each iteration of optimization during neural network training. It is recommended to use the default setting. When the loss curve shows a slow convergence, you can appropriately increase the learning rate; if the accuracy fluctuates greatly, you can appropriately decrease the learning rate.

  • GPU ID

    Graphics card information of the model deployment device. If multiple GPUs are available on the model deployment device, the training can be performed on a specified GPU.

  • Model simplification

    This option is used to simplify the neural network structure. Unselected by default. When the training data is relatively simple, checking the option can improve the training and inference speeds.

  • Max num of training objects

    Unselected by default. This option is to limit the maximum number of objects that can be recognized during training. In general, setting this parameter can speed up inference. Under certain circumstances, setting this parameter can help reduce GPU usage.

Model Finetuning

When a model is put into use for some time, it might not cover certain scenarios. At this point, the model should be iterated. Usually, using more data to re-train the model can do the job, but it could reduce the overall recognition accuracy and might take a long time. Hence, Model Finetuning can be used to iterate the model while maintaining its accuracy and saving time.

This feature only works under the Developer Mode. You can enable the Developer Mode by clicking Settings  Options.

Steps:

  1. Collect images with poor recognition results and add them into the training and validation sets.

  2. Enable Model Finetuning in the window of Training parameter settings and lower the learning rate accordingly; the number of training epochs can be reduced to 50–80.

  3. Confirm the changes to the parameters and start the training.

For example, you can enable Model Finetuning and then select the path of Super Model in the Training parameter configuration window of the Instance Segmentation model to finetune the Super Model.

Training Center

You can view the training progress and memory usage in the Training Center.

If there is already a training task in progress, other training tasks will enter a waiting queue.

  • Click train center stop to terminate the training task in progress.

  • Click train center delete to remove the training task from the waiting queue.

  • Click train center change to stick the training task to the top of the waiting queue.

We Value Your Privacy

We use cookies to provide you with the best possible experience on our website. By continuing to use the site, you acknowledge that you agree to the use of cookies. If you decline, a single cookie will be used to ensure you're not tracked or remembered when you visit this website.