In this tutorial, you will analyze the hyperspectral images of plastic samples. Your goal is to learn how to use Breeze to create a model based on machine learning which can be used for classification of different types of plastic.
The data consist of two images containing samples of five different types of plastic.
Polyethylene terephthalate bottle (PET BOTTLE)
Polyethylene terephthalate sheet (PET SHEET)
Polyethylene terephthalate glycol (PET G)
Polyvinyl chloride (PVC)
The tutorial images contain samples of known plastic type which will be used as training data set and test set. The spectral acquisition was carried out using a Hyspex SWIR-384 camera, with a spectral range of 930 - 2500 nm. The plastic samples were placed on a moving stage and broadband halogen lamps were used as illumination sources. The data used in this tutorial is reduced using average binning, 2 times spatially and 4 times spectrally leaving the images with 192 pixels across the field of view, and 72 spectral bands. This was done to reduce the size of the data files for internet downloading.
In this tutorial you will learn how to:
Use “Manual” segmentation to use your mouse to select samples in an image that will be used as the training set
Use “Grid and Inset” segmentation to add additional data points that will be used for the training
Training of a Machine learning classification model
Classification of the images using your machine learning model
Add additional training data to your model and then re-train and apply it to your image
Do a real-time analysis workflow with object segmentation and classification and apply it using “Simulator camera”
Download tutorial image data
OPTIONAL Change to dark mode by pressing the “Switch to Dark mode”
Enter the Record” view by pressing the “Record” button
You will see the following view (if you already have a Study in Record press the “Add” button in the lower-left corner).
Select the “Tutorial” tab.
Select “Plastic Classification” in the “Name” drop-down menu.
Press “OK” to start downloading the image data.
After the tutorial data is downloaded you will see the following table:
A “Study” called “Plastic_Classification” has now been created. It includes one training image and one test image. You can click on a row in the table to see the preview image (pseudo-RGB) for each image.
Click on the “Open” button to open the study or double-click on the study on the left.
The Group level should look like this:
The image data in the study is organized into two groups called “Train” and “Test”.
Select the Train group in the left menu and press the “Open” button again.
Click on the “Pixel Explore” tab.
To do a quick analysis of the spectral variation in the image, a PCA model has been created based on all pixels in the image. Each point in the “Variance scatter” plot corresponds to a pixel in the image. The points in the scatter plot are clustered based on spectral similarity. The color of the points in the scatter plot are based on density (i.e. red = many points close to each other).
The “Max variance image” is colored by the variation in the 1st component of the PCA model (the X-axis in the scatter plot, t1), and visualizes the biggest spectral variation in the image. In this case, this is the difference between the sample objects and the background.
Hold down the left mouse button to do a selection of a cluster of points to see where these pixels are located in the image. Move the mouse around in the image to see the spectral profile for individual pixels or do a selection to see the average spectra for several pixels.
Select training data
The training data consist of five different types of plastic as mentioned earlier.
To acquire the data needed from each class to train the model we will manually select areas in Pixel Explore. Select the Selection tool “Rect” that will make rectangular areas.
Now we will select areas in the “Max variance image” corresponding to the five different plastic types and the background. To select more than one area in the image hold down “Ctrl” on your keyboard. To zoom, use the scroll wheel on your mouse. To pan the zoomed image hold down the scroll wheel and move the image. First, select areas from five samples and the background. After you have selected the different samples and the background, press “Add Sample(s)” and then press OK.
You will now be back in the Table view and you can see your six manual selected segmentations.
Now we will enter known class information for the training samples. First press “Add Variable or Id”.
Select “Category (Classification variable)”
Change the name to “Plastic type” and press “Add”.
A new column in the table view has appeared.
Right-click on a row in the new Plastic type column and under New class write “PET Bottle” and press Enter on your keyboard or press “Add”
You will now see that the class of the first manual sample is PET Bottle. Do the same procedure and add the classes, “PET Sheet”, “PET G”, “PVC”, “PC” and “Background”. (Please note that the colors used for each class might be slightly different in the version of Breeze you are using).
Create a Machine learning model and retrain it
You will now create a Machine learning (ML) model based on the training set. But before we go into the model wizard to create the model we need to apply a second layer of segmentation to get more training data for the ML model. In this tutorial, we will use the “Grid and inset” segmentation.
Go to the “Analyse Tree” tab.
Click on “Manual” and you will see a plus sign appear.
Click on the plus sign.
Select “Segmentation”, press the drop-down menu beside “Method” and locate “Grid and inset” and press “OK”
In the menu to the right, change the “3x3” under the parameter “Grid” to “6x6”
Press “Apply Changes”
If you change the Segmentation in the drop-down menu under the table to Grid and Inset you can see the 6x6 grid added under each manually selected area.
Press the “Model” button at the bottom right corner.
In “Model” press the “Add” button at the bottom left corner.
Select “Classification” and in the drop-down menu change from “PLS-DA” to “Machine Learning”, press “OK”. ( You can change the name on the model if you want)
In the first step of the Classification wizard, you can see that the “Grid and inset” segmentation is selected. and the “Plastic type” descriptor that you will use to build the model. Press “Next”.
In the next step of the wizard, you can select the samples that you want to include in the model. By default, the measurements from the “Train” group have been included since they have entered class information. Each sample that will be used for the training corresponds to one of the segments in the 6x6 grid on the plastic pieces as you can see in the image on the left.
By default, all wavelength bands are included and no pretreatment is added. The graph on the right is showing the average spectrum for each sample (i.e. each grid). Above this graph, there is an option to select different Pretreatments of the spectral data. All default settings here are OK.
In the last step, we will train the model. For this tutorial, we can keep both the Algorithm on “Auto” and the training time at 30 seconds.
When the model has been trained you will see some models pop up in the model wizard. The model with (“X”) is the model that is chosen from the Auto algorithm. In this case, a Neural Network model was chosen.
It is possible to manually select a model type, but in this case, we will use Auto modeling.
(Comment: we will later in this tutorial add new training data and then retrain the model using the Auto function. If the data has changed significantly, it could be that the function chooses a different model type)
Go to the “Classification” tab to see how well the trained model worked on the training set.
Go back to Record by pressing the “Record” button at the bottom left corner.
Go to the “Analyse Tree”.
Click on “Measurement” and press the plus sign.
Select “Descriptor” and “Classification of categories” and write an “Alias” name like “Classification model”
The classification model will now be added to the Analysis tree. In the menu the right side, set the “Classification Type” to “Pixel class majority”. The “Weights” should have value “1”.
Go to the Table view again and press “Apply changes”. The classification model will now be applied to your image. Select the “Measurement” segmentation level and click in the column for the classification to see the results on the image. You can also press the button to add a legend for the image showing the color coding for the classes. (If you view the image with the “blend” button not selected it might be easier to see the difference between the different colors.)
As you can see, in this case, two of the “PET Sheet” samples are wrongly classified as “PET Bottle”. To see if these can be correctly classified, we will add them to the training data.
Go to “Pixel Explore” and select an area in the upper sample that was wrongly classified.
The press “Add Sample” and select the “PET Sheet” class.
If you go to table you can see that this area has now been added to “Manual” segmentation.
Go to the “Analyse Tree” and click on the Classifcation model node. Then press the “Duplicate” button.
Select the copy of the model and press the “Level Down” button twice so that it is moved after the “Grid and Inset” node.
Go back to the “Table” view, press “Apply changes” and the select the “Grid and Inset” Segmentation level.
In the Table columns for the classification model you will see a small green dot indicating if the sample is included in the model training set. If you scroll down in the table to the samples from the new manual segmentation that we added, you will see that these are not included in the training set.
Select the rows for all these samples, right click and press “Train”.
As you can see the samples now have a green dot indicating they are in the training set.
Right-click again on the selected rows and press “Retrain”. The model will be retrained with this data now included in the training set.
Change the Segmentation level to “Measurement” to see how this retrained model performed. As you can see in this example, the two previously miss-classified objects are now correctly classified. But one of the “PET Bottle” samples is now miss-classified as “PET Sheet”.
Now let’s repeat the procedure that we just did, but this time let’s add training data from this miss-classified sample.
Go the “Pixel explore”, select an area from this sample, press “Add sample” and set the class to “PET Bottle”.
In “Table” set the “Segmentation” to “Grid and inset”, select the new samples in the table, right-click and press “Train”, and then “Retrain”.
Switch the “Segmentation” to “Measurement” to look at the results of the updated classification model. As you can see from the image, all the samples are now finally correctly classified.
Now let’s try and apply the model to the “Test” image. We will also do a real-time analysis workflow with object segmentation and classification and apply it using a “Simulator camera”
Simulate real-time prediction
Now that you have created a model which you are satisfied with, we can go into “Play” to simulate how the model would perform in real time.
Press the “Up” button three times so you come back to the start page.
Press the “Settings” button
Select “Hardware” in the left table.
Make sure the Selected camera is “Prediktera Simulator Camera”.
The type is “FileReader” and Source “Automatic - Selected Study group” (the Frame rate setting changes how fast the analysis will be done in frames or lines per second)
Then press “Connect”.
Press the “Up” button.
Press the “Play” button and select the Plastic classification study.
Press the “Add” button
Select the “New” tab and press “OK”.
Go the the “Analyse Tree” tab and click on the “Manual” segmentation node and then press “Delete”.
Click on the “Measurement” node, the plus sign after it. Then select “Segmentation” and “Model Expression”.
In the Expression field write
Class != Background. This means that we will segment out pixels that are not from the “Background” class.
Click on the Classification model node and then press the “Level down” button to move the model to the end of the tree.
The model classification is now ready for real-time predictions. To test this, click on “Analyse” and the same images used in training the model and the test set will be evaluated in a real-time scenario.
Uncheck the “Save image measurements and calculated descriptors” option and select “Parallel” measurement segmentation.
Click on “Next”.
Click on “Start”.
The data will now be analyzed in a real-time mode with automatic segmentation of the plastic objects.
To get a bigger view of the samples. You can turn on/off the settings for the blend, legend and target tracking setting for the visualization.
Good job! Watch the plastic classification roll by and receive a perfect classification score. Click “Finish” to stop the predictions.