SupervisedImageClassificationExample.cxxΒΆ

Example source code (SupervisedImageClassificationExample.cxx):

// In OTB, a generic streamed filter called \doxygen{otb}{ImageClassificationFilter}
// is available to classify any input multi-channel image according to an input
// classification model file. This filter is generic because it works with any
// classification model type (SVM, KNN, Artificial Neural Network,...) generated
// within the OTB generic Machine Learning framework based on OpenCV (\cite{opencv_library}).
// The input model file is smartly parsed according to its content in order to
// identify which learning method was used to generate it. Once the classification
// method and model are known, the input image can be classified. More details are
// given in subsections \ref{ssec:LearningFromSamples} and \ref{ssec:LearningFromImages}
// to generate a classification model either from samples or from images.
// In this example we will illustrate its use. We start by including the
// appropriate header files.

#include "otbMachineLearningModelFactory.h"
#include "otbImageClassificationFilter.h"

#include "otbVectorImage.h"
#include "otbImage.h"
#include "otbImageFileReader.h"
#include "otbImageFileWriter.h"

int main(int itkNotUsed(argc), char* argv[])
{
  const char* infname    = argv[1];
  const char* modelfname = argv[2];
  const char* outfname   = argv[3];

  // We will assume double precision input images and will also define
  // the type for the labeled pixels.

  const unsigned int Dimension = 2;
  using PixelType              = double;
  using LabeledPixelType       = unsigned short;
  // Our classifier is generic enough to be able to process images
  // with any number of bands. We read the input image as a
  // \doxygen{otb}{VectorImage}. The labeled image will be a scalar image.
  using ImageType        = otb::VectorImage<PixelType, Dimension>;
  using LabeledImageType = otb::Image<LabeledPixelType, Dimension>;

  // We can now define the type for the classifier filter, which is
  // templated over its input and output image types.
  using ClassificationFilterType = otb::ImageClassificationFilter<ImageType, LabeledImageType>;
  using ModelType                = ClassificationFilterType::ModelType;

  // Moreover, it is necessary to define a \doxygen{otb}{MachineLearningModelFactory}
  // which is templated over its input and output pixel types. This factory is used
  // to parse the input model file and to define which classification method to use.
  using MachineLearningModelFactoryType = otb::MachineLearningModelFactory<PixelType, LabeledPixelType>;

  // And finally, we define the reader and the writer. Since the images
  // to classify can be very big, we will use a streamed writer which
  // will trigger the streaming ability of the classifier.
  using ReaderType = otb::ImageFileReader<ImageType>;
  using WriterType = otb::ImageFileWriter<LabeledImageType>;

  // We instantiate the classifier and the reader objects and we set
  // the existing model obtained in a previous training step.
  ClassificationFilterType::Pointer filter = ClassificationFilterType::New();

  ReaderType::Pointer reader = ReaderType::New();
  reader->SetFileName(infname);

  // The input model file is parsed according to its content and the generated model
  // is then loaded within the \doxygen{otb}{ImageClassificationFilter}.
  ModelType::Pointer model;
  model = MachineLearningModelFactoryType::CreateMachineLearningModel(modelfname, MachineLearningModelFactoryType::ReadMode);
  model->Load(modelfname);

  filter->SetModel(model);

  // We plug the pipeline and
  // trigger its execution by updating the output of the writer.

  filter->SetInput(reader->GetOutput());

  WriterType::Pointer writer = WriterType::New();
  writer->SetInput(filter->GetOutput());
  writer->SetFileName(outfname);
  writer->Update();
  return EXIT_SUCCESS;
}