Table Of Contents
- Description
- How does this sample work?
- Preparing sample data
- Running the sample
- Additional resources
- License
- Changelog
- Known issues
This sample, sampleGoogleNet, demonstrates how to import a model trained with Caffe into TensorRT using GoogleNet as an example. Specifically, this sample builds a TensorRT engine from the saved Caffe model, sets input values to the engine, and runs it.
This sample constructs a network based on a saved Caffe model and network description. This sample comes with a pre-trained model called googlenet.caffemodel
located in the data/googlenet
directory. The model used by this sample was trained using ImageNet. For more information, see the BAIR/BVLC GitHub page. The sample reads two Caffe files to build the network:
googlenet.prototxt
- The prototxt file that contains the network design.googlenet.caffemodel
- The model file which contains the trained weights for the network.
For more information, see Importing A Caffe Model Using The C++ Parser API.
The sample then builds the TensorRT engine using the constructed network. See Building an Engine in C++ for more information on this. Finally, the sample runs the engine with the test input (all zeroes) and reports if the sample ran as expected.
In this sample, the following layers are used. For more information about these layers, see the TensorRT Developer Guide: Layers documentation.
Activation layer
The Activation layer implements element-wise activation functions. Specifically, this sample uses the Activation layer with the type kRELU
.
Concatenation layer The Concatenation layer links together multiple tensors of the same non-channel sizes along the channel dimension.
Convolution layer The Convolution layer computes a 2D (channel, height, and width) convolution, with or without bias.
FullyConnected layer The FullyConnected layer implements a matrix-vector product, with or without bias.
LRN layer The LRN layer implements cross-channel Local Response Normalization.
Pooling layer
The Pooling layer implements pooling within a channel. Supported pooling types are maximum
, average
and maximum-average blend
.
SoftMax layer The SoftMax layer applies the SoftMax function on the input tensor along an input dimension specified by the user.
- Download the sample data from TensorRT release tarball, if not already mounted under
/usr/src/tensorrt/data
(NVIDIA NGC containers) and set it to$TRT_DATADIR
.export TRT_DATADIR=/usr/src/tensorrt/data
-
Compile the sample by following build instructions in TensorRT README.
-
Run the sample to build and run a GPU inference engine for GoogleNet.
./sample_googlenet --datadir=<path_to_data_directory> --useDLACore=N
For example:
./sample_googlenet --datadir $TRT_DATADIR/googlenet
NOTE: By default, this sample assumes both googlenet.prototxt
and googlenet.caffemodel
files are located in either the data/samples/googlenet/
or data/googlenet/
directories. The default directory can be changed by supplying the --datadir=<new_path/>
path as a command line argument.
- Verify that the sample ran successfully. If the sample runs successfully you should see output similar to the following:
This output shows that the input to the sample is called
&&&& RUNNING TensorRT.sample_googlenet # ./sample_googlenet [I] Building and running a GPU inference engine for GoogleNet [I] [TRT] Detected 1 input and 1 output network tensors. [I] Ran ./sample_googlenet with: [I] Input(s): data [I] Output(s): prob &&&& PASSED TensorRT.sample_googlenet # ./sample_googlenet
data
, the output tensor is calledprob
and the sample ran successfully;PASSED
.
To see the full list of available options and their descriptions, use the -h
or --help
command line option.
The following resources provide a deeper understanding about GoogleNet:
GoogleNet
Documentation
For terms and conditions for use, reproduction, and distribution, see the TensorRT Software License Agreement documentation.
February 2019
This README.md
file was recreated, updated and reviewed.
There are no known issues in this sample.