Main Content

Training Data for Object Detection and Semantic Segmentation

You can use a labeling app and Computer Vision Toolbox™ objects and functions to train algorithms from ground truth data. Use the labeling app to interactively label ground truth data in a video, image sequence, image collection, or custom data source. Then, use the labeled data to create training data to train an object detector or to train a semantic segmentation network.

This workflow applies to the Image Labeler and Video Labeler apps only. To create training data for the Ground Truth Labeler (Automated Driving Toolbox) app in Automated Driving Toolbox™, use the gatherLabelData (Automated Driving Toolbox) function.

  1. Load data for labeling

  2. Label data and select an automation algorithm: Create ROI and scene labels within the app. For more details, see:

    You can choose from one of the built-in algorithms or create your own custom algorithm to label objects in your data. To learn how to create your own automation algorithm, see Create Automation Algorithm for Labeling.

  3. Export labels: After labeling your data, you can export the labels to the workspace or save them to a file. The labels are exported as a groundTruth object. If your data source consists of multiple image collections, label the entire set of image collections to obtain an array of groundTruth objects. For details about sharing groundTruth objects, see Share and Store Labeled Ground Truth Data.

  4. Create training data: To create training data from the groundTruth object, use one of these functions:

    For objects created using a video file or custom data source, the objectDetectorTrainingData and pixelLabelTrainingData functions write images to disk for groundTruth. Sample the ground truth data by specifying a sampling factor. Sampling mitigates overtraining an object detector on similar samples.

  5. Train algorithm:

See Also

Apps

Functions

Objects

Related Topics