This sample demonstrates how to execute inference using Inference Engine Graph Builder API to build a network on example of the LeNet classifications network.
XML file is not required for network building now. Inference Engine Graph Builder API allows building of a network "on the fly" from source code. The sample uses one-channel
ubyte pictures as input.
Upon the start-up the sample reads command line parameters and builds a network using Graph Builder API and passed weights file. Then, the application loads built network and an image to the Inference Engine plugin.
When inference is done, the application outputs inference results to the standard output stream.
NOTE: This sample is implemented to support models with FP32 weights only.
NOTE: By default, Inference Engine samples and demos expect input with BGR channels order. If you trained your model to work with RGB order, you need to manually rearrange the default channels order in the sample or demo application or reconvert your model using the Model Optimizer tool with
--reverse_input_channelsargument specified. For more information about the argument, refer to When to Specify Input Shapes section of Converting a Model Using General Conversion Parameters.
Running the application with the
-h option yields the following usage message:
Running the application with empty list of options yields the usage message given above.
For example, to do inference of an ubyte image on a GPU run the following command:
By default the application outputs top-10 inference results for each infer request. In addition to this information it will provide throughput value measured in frames per seconds.