英姿勃勃的饺子 · 【文章】本田电动摩托怒刷曼岛TT ...· 4 周前 · |
发呆的手套 · Create a Django App ...· 2 月前 · |
火星上的油条 · error Parsing error: ...· 8 月前 · |
安静的感冒药 · Université de Montréal· 11 月前 · |
python axis padding |
https://docs.nvidia.com/deeplearning/tensorrt/api/python_api/infer/Graph/Layers.html |
玩足球的杯子
1 月前 |
For more information, including examples, refer to the TensorRT Operator’s Reference documentation .
Note that layer weight properties may be represented as NumPy arrays or Weights objects depending on whether the underlying datatype is supported by NumPy. Explicitly construct a Weights object from the property if you want a consistent type:
conv_layer = network.add_convolution_nd(...)
bias_weights = trt.Weights(conv_layer.bias)
PaddingMode#
tensorrt.PaddingMode#
- Enumerates types of padding available in convolution, deconvolution and pooling layers.
Padding mode takes precedence if both padding_mode
and pre_padding
are set.
EXPLICIT* corresponds to explicit padding.
SAME* implicitly calculates padding such that the output dimensions are the same as the input dimensions. For convolution and pooling,
output dimensions are determined by ceil(input dimensions, stride).
CAFFE* corresponds to symmetric padding.
Members:
EXPLICIT_ROUND_DOWN : Use explicit padding, rounding the output size down
EXPLICIT_ROUND_UP : Use explicit padding, rounding the output size up
SAME_UPPER : Use SAME padding, with pre_padding
<= post_padding
SAME_LOWER : Use SAME padding, with pre_padding
>= post_padding
class tensorrt.ICastLayer#
A layer that represents the cast function.
This layer casts the element of a given input tensor to a specified data type and returns an output tensor of the same shape in the converted type.
Conversions between all types except FP8 is supported.
- Variables
to_type – DataType
The specified data type of the output tensor.
class tensorrt.IConvolutionLayer#
A convolution layer in an INetworkDefinition
.
This layer performs a correlation operation between 3 or 4 dimensional filter with a 4 or 5 dimensional tensor to produce another 4 or 5 dimensional tensor.
An optional bias argument is supported, which adds a per-channel constant to each value in the output.
- Variables
num_output_maps – int
The number of output maps for the convolution.
pre_padding – DimsHW
The pre-padding. The start of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
post_padding – DimsHW
The post-padding. The end of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
padding_mode – PaddingMode
The padding mode. Padding mode takes precedence if both IConvolutionLayer.padding_mode
and either IConvolutionLayer.pre_padding
or IConvolutionLayer.post_padding
are set.
num_groups – int
The number of groups for a convolution. The input tensor channels are divided into this many groups, and a convolution is executed for each group, using a filter per group. The results of the group convolutions are concatenated to form the output. Note When using groups in int8 mode, the size of the groups (i.e. the channel count divided by the group count) must be a multiple of 4 for both input and output. Default: 1.
kernel – Weights
The kernel weights for the convolution. The weights are specified as a contiguous array in GKCRS order, where G is the number of groups, K the number of output feature maps, C the number of input channels, and R and S are the height and width of the filter.
bias – Weights
The bias weights for the convolution. Bias is optional. To omit bias, set this to an empty Weights
object. The bias is applied per-channel, so the number of weights (if non-zero) must be equal to the number of output feature maps.
kernel_size_nd – Dims
The multi-dimension kernel size of the convolution.
stride_nd – Dims
The multi-dimension stride of the convolution. Default: (1, …, 1)
padding_nd – Dims
The multi-dimension padding of the convolution. The input will be zero-padded by this number of elements in each dimension. If the padding is asymmetric, this value corresponds to the pre-padding. Default: (0, …, 0)
dilation_nd – Dims
The multi-dimension dilation for the convolution. Default: (1, …, 1)
tensorrt.InterpolationMode#
Various modes of interpolation, used in resize and grid_sample layers.
Members:
NEAREST : 1D, 2D, and 3D nearest neighbor interpolation.
LINEAR : Supports linear, bilinear, trilinear interpolation.
CUBIC : Supports bicubic interpolation.
tensorrt.SampleMode#
Controls how ISliceLayer and IGridSample handles out of bounds coordinates
Members:
STRICT_BOUNDS : Fail with error when the coordinates are out of bounds.
WRAP : Coordinates wrap around periodically.
CLAMP : Out of bounds indices are clamped to bounds
FILL : Use fill input value when coordinates are out of bounds.
REFLECT : Coordinates reflect.
class tensorrt.IGridSampleLayer#
A grid sample layer in an INetworkDefinition
.
This layer uses an input tensor and a grid tensor to produce an interpolated output tensor.
The input and grid tensors must shape tensors of rank 4. The only supported SampleMode s are
trt.samplemode.CLAMP, trt.samplemode.FILL, and trt.samplemode.REFLECT.
- Variables
interpolation_mode – class:InterpolationMode The interpolation type to use. Defaults to LINEAR.
align_corners – class:bool the align mode to use. Defaults to False.
sample_mode – SampleMode
The sample mode to use. Defaults to FILL.
RELU : Rectified Linear activation
SIGMOID : Sigmoid activation
TANH : Hyperbolic Tangent activation
LEAKY_RELU : Leaky Relu activation: f(x) = x if x >= 0, f(x) = alpha * x if x < 0
ELU : Elu activation: f(x) = x if x >= 0, f(x) = alpha * (exp(x) - 1) if x < 0
SELU : Selu activation: f(x) = beta * x if x > 0, f(x) = beta * (alpha * exp(x) - alpha) if x <= 0
SOFTSIGN : Softsign activation: f(x) = x / (1 + abs(x))
SOFTPLUS : Softplus activation: f(x) = alpha * log(exp(beta * x) + 1)
CLIP : Clip activation: f(x) = max(alpha, min(beta, x))
HARD_SIGMOID : Hard sigmoid activation: f(x) = max(0, min(1, alpha * x + beta))
SCALED_TANH : Scaled Tanh activation: f(x) = alpha * tanh(beta * x)
THRESHOLDED_RELU : Thresholded Relu activation: f(x) = x if x > alpha, f(x) = 0 if x <= alpha
GELU_ERF : GELU erf activation: 0.5 * x * (1 + erf(sqrt(0.5) * x))
GELU_TANH : GELU tanh activation: 0.5 * x * (1 + tanh(sqrt(2/pi) * (0.044715F * pow(x, 3) + x)))
class tensorrt.IActivationLayer#
An Activation layer in an INetworkDefinition
. This layer applies a per-element activation function to its input. The output has the same shape as the input.
- Variables
type – ActivationType
The type of activation to be performed.
alpha – float
The alpha parameter that is used by some parametric activations (LEAKY_RELU, ELU, SELU, SOFTPLUS, CLIP, HARD_SIGMOID, SCALED_TANH). Other activations ignore this parameter.
beta – float
The beta parameter that is used by some parametric activations (SELU, SOFTPLUS, CLIP, HARD_SIGMOID, SCALED_TANH). Other activations ignore this parameter.
MAX : Maximum over elements
AVERAGE : Average over elements. If the tensor is padded, the count includes the padding
MAX_AVERAGE_BLEND : Blending between the max pooling and average pooling: (1-blendFactor)*maxPool + blendFactor*avgPool
class tensorrt.IPoolingLayer#
A Pooling layer in an INetworkDefinition
. The layer applies a reduction operation within a window over the input.
- Variables
type – PoolingType
The type of pooling to be performed.
pre_padding – DimsHW
The pre-padding. The start of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
post_padding – DimsHW
The post-padding. The end of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
padding_mode – PaddingMode
The padding mode. Padding mode takes precedence if both IPoolingLayer.padding_mode
and either IPoolingLayer.pre_padding
or IPoolingLayer.post_padding
are set.
blend_factor – float
The blending factor for the max_average_blend mode: \(max_average_blendPool = (1-blendFactor)*maxPool + blendFactor*avgPool\) . blend_factor
is a user value in [0,1] with the default value of 0.0. This value only applies for the PoolingType.MAX_AVERAGE_BLEND
mode.
average_count_excludes_padding – bool
Whether average pooling uses as a denominator the overlap area between the window and the unpadded input. If this is not set, the denominator is the overlap between the pooling window and the padded input. Default: True
window_size_nd – Dims
The multi-dimension window size for pooling.
stride_nd – Dims
The multi-dimension stride for pooling. Default: (1, …, 1)
padding_nd – Dims
The multi-dimension padding for pooling. Default: (0, …, 0)
class tensorrt.ILRNLayer#
A LRN layer in an INetworkDefinition
. The output size is the same as the input size.
- Variables
window_size – int
The LRN window size. The window size must be odd and in the range of [1, 15].
alpha – float
The LRN alpha value. The valid range is [-1e20, 1e20].
beta – float
The LRN beta value. The valid range is [0.01, 1e5f].
k – float
The LRN K value. The valid range is [1e-5, 1e10].
UNIFORM : Identical coefficients across all elements of the tensor.
CHANNEL : Per-channel coefficients. The channel dimension is assumed to be the third to last dimension.
ELEMENTWISE : Elementwise coefficients.
class tensorrt.IScaleLayer#
A Scale layer in an INetworkDefinition
.
This layer applies a per-element computation to its input:
\(output = (input * scale + shift) ^ {power}\)
The coefficients can be applied on a per-tensor, per-channel, or per-element basis.
If the number of weights is 0, then a default value is used for shift, power, and scale. The default shift is 0, the default power is 1, and the default scale is 1.
The output size is the same as the input size.
The input tensor for this layer is required to have a minimum of 3 dimensions.
- Variables
mode – ScaleMode
The scale mode.
shift – Weights
The shift value.
scale – Weights
The scale value.
power – Weights
The power value.
channel_axis – int
The channel axis.
class tensorrt.ISoftMaxLayer#
A Softmax layer in an INetworkDefinition
.
This layer applies a per-channel softmax to its input.
The output size is the same as the input size.
- Variables
axes – int
The axis along which softmax is computed. Currently, only one axis can be set.
The axis is specified by setting the bit corresponding to the axis to 1, as a bit mask.
For example, consider an NCHW tensor as input (three non-batch dimensions).
By default, softmax is performed on the axis which is the number of axes minus three. It is 0 if there are fewer than 3 non-batch axes. For example, if the input is NCHW, the default axis is C. If the input is NHW, then the default axis is H.
Bit 0 corresponds to the N dimension boolean.
Bit 1 corresponds to the C dimension boolean.
Bit 2 corresponds to the H dimension boolean.
Bit 3 corresponds to the W dimension boolean.
By default, softmax is performed on the axis which is the number of axes minus three. It is 0 if
there are fewer than 3 axes. For example, if the input is NCHW, the default axis is C. If the input
is NHW, then the default axis is N.
For example, to perform softmax on axis R of a NPQRCHW input, set bit 3.
The following constraints must be satisfied to execute this layer on DLA:
Axis must be one of the channel or spatial dimensions.
There are two classes of supported input sizes:
Non-axis, non-batch dimensions are all 1 and the axis dimension is at most 8192. This is the recommended case for using softmax since it is the most accurate.
At least one non-axis, non-batch dimension greater than 1 and the axis dimension is at most 1024. Note that in this case, there may be some approximation error as the axis dimension size approaches the upper bound. See the TensorRT Developer Guide for more details on the approximation error.
class tensorrt.IConcatenationLayer#
A concatenation layer in an INetworkDefinition
.
The output channel size is the sum of the channel sizes of the inputs.
The other output sizes are the same as the other input sizes, which must all match.
- Variables
axis – int
The axis along which concatenation occurs. The default axis is the number of tensor dimensions minus three, or zero if the tensor has fewer than three dimensions. For example, for a tensor with dimensions NCHW, it is C.
num_output_maps – int
The number of output feature maps for the deconvolution.
pre_padding – DimsHW
The pre-padding. The start of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
post_padding – DimsHW
The post-padding. The end of input will be zero-padded by this number of elements in the height and width directions. Default: (0, 0)
padding_mode – PaddingMode
The padding mode. Padding mode takes precedence if both IDeconvolutionLayer.padding_mode
and either IDeconvolutionLayer.pre_padding
or IDeconvolutionLayer.post_padding
are set.
num_groups – int
The number of groups for a deconvolution. The input tensor channels are divided into this many groups, and a deconvolution is executed for each group, using a filter per group. The results of the group convolutions are concatenated to form the output. Note When using groups in int8 mode, the size of the groups (i.e. the channel count divided by the group count) must be a multiple of 4 for both input and output. Default: 1
kernel – Weights
The kernel weights for the deconvolution. The weights are specified as a contiguous array in CKRS order, where C the number of input channels, K the number of output feature maps, and R and S are the height and width of the filter.
bias – Weights
The bias weights for the deconvolution. Bias is optional. To omit bias, set this to an empty Weights
object. The bias is applied per-feature-map, so the number of weights (if non-zero) must be equal to the number of output feature maps.
kernel_size_nd – Dims
The multi-dimension kernel size of the convolution.
stride_nd – Dims
The multi-dimension stride of the deconvolution. Default: (1, …, 1)
padding_nd – Dims
The multi-dimension padding of the deconvolution. The input will be zero-padded by this number of elements in each dimension. Padding is symmetric. Default: (0, …, 0)
tensorrt.ElementWiseOperation#
The binary operations that may be performed by an ElementWise layer.
Members:
SUM : Sum of the two elements
PROD : Product of the two elements
MAX : Max of the two elements
MIN : Min of the two elements
SUB : Subtract the second element from the first
DIV : Divide the first element by the second
POW : The first element to the power of the second element
FLOOR_DIV : Floor division of the first element by the second
AND : Logical AND of two elements
OR : Logical OR of two elements
XOR : Logical XOR of two elements
EQUAL : Check if two elements are equal
GREATER : Check if element in first tensor is greater than corresponding element in second tensor
LESS : Check if element in first tensor is less than corresponding element in second tensor
class tensorrt.IElementWiseLayer#
A elementwise layer in an INetworkDefinition
.
This layer applies a per-element binary operation between corresponding elements of two tensors.
The input dimensions of the two input tensors must be equal, and the output tensor is the same size as each input.
- Variables
op – ElementWiseOperation
The binary operation for the layer.
- Variables
axis – int
The non-batch dimension axis to gather on. The axis must be less than the number of non-batch dimensions in the data input.
num_elementwise_dims – int
The number of leading dimensions of indices tensor to be handled elementwise. For GatherMode.DEFAULT, it can be 0 or 1. For GatherMode::kND, it can be between 0 and one less than rank(data). For GatherMode::kELEMENT, it must be 0.
mode – GatherMode
The gather mode.
tensorrt.UnaryOperation#
The unary operations that may be performed by a Unary layer.
Members:
EXP : Exponentiation
LOG : Log (base e)
SQRT : Square root
RECIP : Reciprocal
ABS : Absolute value
NEG : Negation
SIN : Sine
COS : Cosine
TAN : Tangent
SINH : Hyperbolic sine
COSH : Hyperbolic cosine
ASIN : Inverse sine
ACOS : Inverse cosine
ATAN : Inverse tangent
ASINH : Inverse hyperbolic sine
ACOSH : Inverse hyperbolic cosine
ATANH : Inverse hyperbolic tangent
CEIL : Ceiling
FLOOR : Floor
ERF : Gauss error function
NOT : Not
SIGN : Sign. If input > 0, output 1; if input < 0, output -1; if input == 0, output 0.
ROUND : Round to nearest even for floating-point data type.
ISINF : Return true if the input value equals +/- infinity for floating-point data type.
ISNAN : Return true if the input value equals NaN for floating-point data type.
tensorrt.ReduceOperation#
The reduce operations that may be performed by a Reduce layer
Members:
SUM :
PROD :
MAX :
MIN :
AVG :
op – ReduceOperation
The reduce operation for the layer.
axes – int
The axes over which to reduce.
keep_dims – bool
Specifies whether or not to keep the reduced dimensions for the layer.
- Variables
pre_padding_nd – Dims
The padding that is applied at the start of the tensor. Negative padding results in trimming the edge by the specified amount. Only 2 dimensions currently supported.
post_padding_nd – Dims
The padding that is applied at the end of the tensor. Negative padding results in trimming the edge by the specified amount. Only 2 dimensions currently supported.
class tensorrt.IParametricReLULayer#
A parametric ReLU layer in an INetworkDefinition
.
This layer applies a parametric ReLU activation to an input tensor (first input), with slopes taken from a
slopes tensor (second input). This can be viewed as a leaky ReLU operation where the negative slope differs
from element to element (and can in fact be learned).
The slopes tensor must be unidirectional broadcastable to the input tensor: the rank of the two tensors must
be the same, and all dimensions of the slopes tensor must either equal the input tensor or be 1.
The output tensor has the same shape as the input tensor.
ISelectLayer#
class tensorrt.ISelectLayer#
A select layer in an INetworkDefinition
.
This layer implements an element-wise ternary conditional operation. Wherever condition
is True
, elements are taken from the first input, and wherever condition
is False
, elements are taken from the second input.
IShuffleLayer#
class tensorrt.Permutation(*args, **kwargs)#
The elements of the permutation. The permutation is applied as outputDimensionIndex = permutation[inputDimensionIndex], so to permute from CHW order to HWC order, the required permutation is [1, 2, 0], and to permute from HWC to CHW, the required permutation is [2, 0, 1].
It supports iteration and indexing and is implicitly convertible to/from Python iterables (like tuple
or list
). Therefore, you can use those classes in place of Permutation
.
Overloaded function.
__init__(self: tensorrt.tensorrt.Permutation) -> None
__init__(self: tensorrt.tensorrt.Permutation, arg0: List[int]) -> None
class tensorrt.IShuffleLayer#
A shuffle layer in an INetworkDefinition
.
This class shuffles data by applying in sequence: a transpose operation, a reshape operation and a second transpose operation. The dimension types of the output are those of the reshape dimension.
- Variables
first_transpose – Permutation
The permutation applied by the first transpose operation. Default: Identity Permutation
reshape_dims – Dims
The reshaped dimensions.
Two special values can be used as dimensions.
Value 0 copies the corresponding dimension from input. This special value can be used more than once in the dimensions. If number of reshape dimensions is less than input, 0s are resolved by aligning the most significant dimensions of input.
Value -1 infers that particular dimension by looking at input and rest of the reshape dimensions. Note that only a maximum of one dimension is permitted to be specified as -1.
The product of the new dimensions must be equal to the product of the old.
second_transpose – Permutation
The permutation applied by the second transpose operation. Default: Identity Permutation
zero_is_placeholder – bool
The meaning of 0 in reshape dimensions.
If true, then a 0 in the reshape dimensions denotes copying the corresponding
dimension from the first input tensor. If false, then a 0 in the reshape
dimensions denotes a zero-length dimension.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Sets the input tensor for the given index. The index must be 0 for a static shuffle layer.
A static shuffle layer is converted to a dynamic shuffle layer by calling set_input()
with an index 1.
A dynamic shuffle layer cannot be converted back to a static shuffle layer.
For a dynamic shuffle layer, the values 0 and 1 are valid.
The indices in the dynamic case are as follows:
class tensorrt.ISliceLayer#
A slice layer in an INetworkDefinition
.
The slice layer has two variants, static and dynamic.
Static slice specifies the start, size, and stride dimensions at layer creation time via Dims
and can use the get/set accessor functions of the ISliceLayer
.
Dynamic slice specifies one or more of start, size, stride, or axes as ITensor`s, by using :func:`ILayer.set_input
to add a second, third, fourth, or sixth input respectively.
The corresponding Dims
are used if an input is missing or null.
An application can determine if the ISliceLayer
has a dynamic output shape based on whether the size or axes input is present and non-null.
The slice layer selects for each dimension a start location from within the input tensor, and copies elements to the output tensor using the specified stride across the input tensor.
Start, size, and stride tensors must be 1-D integer-typed shape tensors if not specified via Dims
.
An example of using slice on a tensor:
input = {{0, 2, 4}, {1, 3, 5}}
start = {1, 0}
size = {1, 2}
stride = {1, 2}
output = {{1, 5}}
If axes is provided then starts, ends, and strides must have the same length as axes and specifies a subset of dimensions to slice. If axes is not provided, starts, ends, and strides
must be of the same length as the rank of the input tensor.
An example of using slice on a tensor with axes specified:
input = {{0, 2, 4}, {1, 3, 5}}
start = {1}
size = {2}
stride = {1}
axes = {1}
output = {{2, 4}, {3, 5}}
When the sampleMode is SampleMode.CLAMP
or SampleMode.REFLECT
, for each input dimension, if its size is 0 then the corresponding output dimension must be 0 too.
When the sampleMode is SampleMode.FILL
, the fifth input to the slice layer is used to determine the value to fill in out-of-bound indices. It is an error to specify the fifth input in any other sample mode.
A slice layer can produce a shape tensor if the following conditions are met:
start
, size
, and stride
are build time constants, either as static Dims
or as constant input tensors.
axes
, if provided, is a build time constant, either as static Dims
or as a constant input tensor.
The number of elements in the output tensor does not exceed 2 * Dims.MAX_DIMS
.
The input tensor is a shape tensor if the output is a shape tensor.
The following constraints must be satisfied to execute this layer on DLA:
* start
, size
, and stride
are build time constants, either as static Dims
or as constant input tensors.
* axes
, if provided, is a build time constant, either as static Dims
or as a constant input tensor.
* sampleMode is SampleMode.DEFAULT
, SampleMode.WRAP
, or SampleMode.FILL
.
* Strides are 1 for all dimensions.
* Slicing is not performed on the first dimension.
* The input tensor has four dimensions.
* For SliceMode.FILL
, the fill value input is a scalar output of an IConstantLayer
with value 0 that is not consumed by any other layer.
- Variables
start – Dims
The start offset.
shape – Dims
The output dimensions.
stride – Dims
The slicing stride.
mode – SampleMode
Controls how ISliceLayer
handles out of bounds coordinates.
axes – Dims
The axes that starts, sizes, and strides correspond to.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Sets the input tensor for the given index. The index must be 0 or 4 for a static slice layer.
A static slice layer is converted to a dynamic slice layer by calling set_input()
with an index between 1 and 3.
A dynamic slice layer cannot be converted back to a static slice layer.
The indices are as follows:
If this function is called with a value greater than 0, then num_inputs
changes
from 1 to index + 1.
- Parameters
index – The index of the input tensor.
tensor – The input tensor.
class tensorrt.IShapeLayer#
A shape layer in an INetworkDefinition
. Used for getting the shape of a tensor.
This class sets the output to a one-dimensional tensor with the dimensions of the input tensor.
For example, if the input is a four-dimensional tensor (of any type) with
dimensions [2,3,5,7], the output tensor is a one-dimensional int64
tensor
of length 4 containing the sequence 2, 3, 5, 7.
ITopKLayer#
tensorrt.TopKOperation#
The operations that may be performed by a TopK layer
Members:
MAX : Maximum of the elements
MIN : Minimum of the elements
op – TopKOperation
The operation for the layer.
k – TopKOperation
the k value for the layer. Currently only values up to 3840 are supported.
Use the set_input() method with index 1 to pass in dynamic k as a tensor.
axes – TopKOperation
The axes along which to reduce.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Sets the input tensor for the given index. The index must be 0 or 1 for a TopK layer.
The indices are as follows:
- A scalar Int32 tensor containing a positive value corresponding to the number
of top elements to retrieve. Values larger than 3840 will result in a runtime
error. If provided, this will override the static k value in calculations.
tensorrt.MatrixOperation#
The matrix operations that may be performed by a Matrix layer
Members:
NONE :
TRANSPOSE : Transpose each matrix
VECTOR : Treat operand as collection of vectors
class tensorrt.IMatrixMultiplyLayer#
A matrix multiply layer in an INetworkDefinition
.
Let A be op(getInput(0)) and B be op(getInput(1)) where
op(x) denotes the corresponding MatrixOperation.
When A and B are matrices or vectors, computes the inner product A * B:
matrix * matrix -> matrix
matrix * vector -> vector
vector * matrix -> vector
vector * vector -> scalar
Inputs of higher rank are treated as collections of matrices or vectors.
The output will be a corresponding collection of matrices, vectors, or scalars.
- Variables
op0 – MatrixOperation
How to treat the first input.
op1 – MatrixOperation
How to treat the second input.
class tensorrt.IRaggedSoftMaxLayer#
A ragged softmax layer in an INetworkDefinition
.
This layer takes a ZxS input tensor and an additional Zx1 bounds tensor holding the lengths of the Z sequences.
This layer computes a softmax across each of the Z sequences.
The output tensor is of the same size as the input tensor.
IIdentityLayer#
class tensorrt.IIdentityLayer#
A layer that represents the identity function.
If tensor precision is explicitly specified, it can be used to transform from one precision to another.
Other than conversions between the same type (float32
-> float32
for example), the only valid conversions are:
(float32
| float16
| int32
| bool
) -> (float32
| float16
| int32
| bool
)
(float32
| float16
) -> uint8
uint8
-> (float32
| float16
)
IConstantLayer#
class tensorrt.IConstantLayer#
A constant layer in an INetworkDefinition
.
Note: This layer does not support boolean and uint8 types.
- Variables
weights – Weights
The weights for the layer.
shape – Dims
The shape of the layer.
class tensorrt.IResizeLayer#
A resize layer in an INetworkDefinition
.
Resize layer can be used for resizing a N-D tensor.
Resize layer currently supports the following configurations:
InterpolationMode.NEAREST - resizes innermost m dimensions of N-D, where 0 < m <= min(3, N) and N > 0.
InterpolationMode.LINEAR - resizes innermost m dimensions of N-D, where 0 < m <= min(3, N) and N > 0.
InterpolationMode.CUBIC - resizes innermost 2 dimensions of N-D, N >= 2.
Default resize mode is InterpolationMode.NEAREST.
Resize layer provides two ways to resize tensor dimensions:
- Set output dimensions directly. It can be done for static as well as dynamic resize layer.
Static resize layer requires output dimensions to be known at build-time.
Dynamic resize layer requires output dimensions to be set as one of the input tensors.
If executing this layer on DLA, the following combinations of parameters are supported:
In NEAREST mode:
(ResizeCoordinateTransformation.ASYMMETRIC, ResizeSelector.FORMULA, ResizeRoundMode.FLOOR)
(ResizeCoordinateTransformation.HALF_PIXEL, ResizeSelector.FORMULA, ResizeRoundMode.HALF_DOWN)
(ResizeCoordinateTransformation.HALF_PIXEL, ResizeSelector.FORMULA, ResizeRoundMode.HALF_UP)
In LINEAR and CUBIC mode:
(ResizeCoordinateTransformation.HALF_PIXEL, ResizeSelector.FORMULA)
(ResizeCoordinateTransformation.HALF_PIXEL, ResizeSelector.UPPER)
shape – Dims
The output dimensions. Must to equal to input dimensions size.
scales – List[float]
List of resize scales.
If executing this layer on DLA, there are three restrictions:
1. len(scales)
has to be exactly 4.
2. The first two elements in scales need to be exactly 1 (for unchanged batch and channel dimensions).
3. The last two elements in scales, representing the scale values along height and width dimensions,
respectively, need to be integer values in the range of [1, 32] for NEAREST mode and [1, 4] for LINEAR.
Example of DLA-supported scales: [1, 1, 2, 2].
resize_mode – InterpolationMode
Resize mode can be Linear, Cubic or Nearest.
coordinate_transformation – ResizeCoordinateTransformationDoc
Supported resize coordinate transformation modes are ALIGN_CORNERS, ASYMMETRIC and HALF_PIXEL.
selector_for_single_pixel – ResizeSelector
Supported resize selector modes are FORMULA and UPPER.
nearest_rounding – ResizeRoundMode
Supported resize Round modes are HALF_UP, HALF_DOWN, FLOOR and CEIL.
exclude_outside – int
If set to 1, the weight of sampling locations outside the input tensor will be set to 0, and the weight will be renormalized so that their sum is 1.0.
cubic_coeff – float
coefficient ‘a’ used in cubic interpolation.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Sets the input tensor for the given index.
If index == 1 and num_inputs == 1, num_inputs changes to 2.
Once such additional input is set, resize layer works in dynamic mode.
When index == 1 and num_inputs == 1, the output dimensions are used from
the input tensor, overriding the dimensions supplied by shape.
- Parameters
index – The index of the input tensor.
tensor – The input tensor.
add_iterator(self: tensorrt.tensorrt.ILoop, tensor: tensorrt.tensorrt.ITensor, axis: int = 0, reverse: bool = False) → tensorrt.tensorrt.IIteratorLayer#
Return layer that subscripts tensor by loop iteration.
For reverse=false, this is equivalent to add_gather(tensor, I, 0) where I is a
scalar tensor containing the loop iteration number.
For reverse=true, this is equivalent to add_gather(tensor, M-1-I, 0) where M is the trip count
computed from TripLimits of kind COUNT
.
- Parameters
tensor – The tensor to iterate over.
axis – The axis along which to iterate.
reverse – Whether to iterate in the reverse direction.
- Returns
The IIteratorLayer
, or None
if it could not be created.
add_loop_output(self: tensorrt.tensorrt.ILoop, tensor: tensorrt.tensorrt.ITensor, kind: tensorrt.tensorrt.LoopOutput, axis: int = 0) → tensorrt.tensorrt.ILoopOutputLayer#
Make an output for this loop, based on the given tensor.
If kind
is CONCATENATE
or REVERSE
, a second input specifying the
concatenation dimension must be added via method ILoopOutputLayer.set_input()
.
- Parameters
kind – The kind of loop output. See LoopOutput
axis – The axis for concatenation (if using kind
of CONCATENATE
or REVERSE
).
- Returns
The added ILoopOutputLayer
, or None
if it could not be created.
add_recurrence(self: tensorrt.tensorrt.ILoop, initial_value: tensorrt.tensorrt.ITensor) → tensorrt.tensorrt.IRecurrenceLayer#
Create a recurrence layer for this loop with initial_value as its first input.
- Parameters
initial_value – The initial value of the recurrence layer.
- Returns
The added IRecurrenceLayer
, or None
if it could not be created.
add_trip_limit(self: tensorrt.tensorrt.ILoop, tensor: tensorrt.tensorrt.ITensor, kind: tensorrt.tensorrt.TripLimit) → tensorrt.tensorrt.ITripLimitLayer#
Add a trip-count limiter, based on the given tensor.
There may be at most one COUNT
and one WHILE
limiter for a loop.
When both trip limits exist, the loop exits when the
count is reached or condition is falsified.
It is an error to not add at least one trip limiter.
For WHILE
, the input tensor must be the output of a subgraph that contains
only layers that are not ITripLimitLayer
, IIteratorLayer
or ILoopOutputLayer
.
Any IRecurrenceLayer
s in the subgraph must belong to the same loop as the
ITripLimitLayer
. A trivial example of this rule is that the input to the WHILE
is the output of an IRecurrenceLayer
for the same loop.
- Parameters
tensor – The input tensor. Must be available before the loop starts.
kind – The kind of trip limit. See TripLimit
- Returns
The added ITripLimitLayer
, or None
if it could not be created.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Set the first or second input.
If index==1 and the number of inputs is one, the input is appended.
The first input specifies the initial output value, and must come from outside the loop.
The second input specifies the next output value, and must come from inside the loop.
The two inputs must have the same dimensions.
- Parameters
index – The index of the input to set.
tensor – The input tensor.
axis – The axis to iterate over
reverse – For reverse=false, the layer is equivalent to add_gather(tensor, I, 0) where I is a
scalar tensor containing the loop iteration number.
For reverse=true, the layer is equivalent to add_gather(tensor, M-1-I, 0) where M is the trip count
computed from TripLimits of kind COUNT
.
The default is reverse=false.
LAST_VALUE : Output value is value of tensor for last iteration.
CONCATENATE : Output value is concatenation of values of tensor for each iteration, in forward order.
REVERSE : Output value is concatenation of values of tensor for each iteration, in reverse order.
class tensorrt.ILoopOutputLayer#
An ILoopOutputLayer
is the sole way to get output from a loop.
The first input tensor must be defined inside the loop; the output tensor is outside the loop.
The second input tensor, if present, must be defined outside the loop.
If kind
is LAST_VALUE
, a single input must be provided.
If kind
is CONCATENATE
or REVERSE
, a second input must be provided.
The second input must be a scalar “shape tensor”, defined before the loop commences,
that specifies the concatenation length of the output.
The output tensor has j more dimensions than the input tensor, where
j == 0 if kind
is LAST_VALUE
j == 1 if kind
is CONCATENATE
or REVERSE
.
- Variables
axis – The contenation axis. Ignored if kind
is LAST_VALUE
.
For example, if the input tensor has dimensions [b,c,d],
and kind
is CONCATENATE
, the output has four dimensions.
Let a be the value of the second input.
axis=0 causes the output to have dimensions [a,b,c,d].
axis=1 causes the output to have dimensions [b,a,c,d].
axis=2 causes the output to have dimensions [b,c,a,d].
axis=3 causes the output to have dimensions [b,c,d,a].
Default is axis is 0.
kind – The kind of loop output. See LoopOutput
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Like ILayer.set_input()
, but additionally works if index==1, num_inputs`==1, in which case :attr:`num_inputs
changes to 2.
tensorrt.FillOperation#
The tensor fill operations that may performed by an Fill layer.
Members:
LINSPACE : Generate evenly spaced numbers over a specified interval
RANDOM_UNIFORM : Generate a tensor with random values drawn from a uniform distribution
RANDOM_NORMAL : Generate a tensor with random values drawn from a normal distribution
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
replace an input of this layer with a specific tensor.
class tensorrt.IQuantizeLayer#
A Quantize layer in an INetworkDefinition
.
This layer accepts a floating-point data input tensor, and uses the scale and zeroPt inputs to
quantize the data to an 8-bit signed integer according to:
\(output = clamp(round(input / scale) + zeroPt)\)
Rounding type is rounding-to-nearest ties-to-even (https://en.wikipedia.org/wiki/Rounding#Round_half_to_even).
Clamping is in the range [-128, 127].
The first input (index 0) is the tensor to be quantized.
The second (index 1) and third (index 2) are the scale and zero point respectively.
Each of scale and zeroPt must be either a scalar, or a 1D tensor.
The zeroPt tensor is optional, and if not set, will be assumed to be zero. Its data type must be
tensorrt.int8. zeroPt must only contain zero-valued coefficients, because only symmetric quantization is
supported.
The scale value must be either a scalar for per-tensor quantization, or a 1D tensor for per-axis
quantization. The size of the 1-D scale tensor must match the size of the quantization axis. The size of the
scale must match the size of the zeroPt.
The subgraph which terminates with the scale tensor must be a build-time constant. The same restrictions apply
to the zeroPt.
The output type, if constrained, must be constrained to tensorrt.int8 or tensorrt.fp8. The input type, if constrained, must be
constrained to tensorrt.float32, tensorrt.float16 or tensorrt.bfloat16.
The output size is the same as the input size.
IQuantizeLayer supports tensorrt.float32, tensorrt.float16 and tensorrt.bfloat16 precision and will default to tensorrt.float32 precision during instantiation.
IQuantizeLayer supports tensorrt.int8 and tensorrt.float8 output.
- Variables
axis – int
The axis along which quantization occurs. The quantization axis is in reference to the input tensor’s dimensions.
to_type – DataType
The specified data type of the output tensor. Must be tensorrt.int8 or tensorrt.float8.
class tensorrt.IDequantizeLayer#
A Dequantize layer in an INetworkDefinition
.
This layer accepts a signed 8-bit integer input tensor, and uses the configured scale and zeroPt inputs to
dequantize the input according to:
\(output = (input - zeroPt) * scale\)
The first input (index 0) is the tensor to be quantized.
The second (index 1) and third (index 2) are the scale and zero point respectively.
Each of scale and zeroPt must be either a scalar, or a 1D tensor.
The zeroPt tensor is optional, and if not set, will be assumed to be zero. Its data type must be
tensorrt.int8. zeroPt must only contain zero-valued coefficients, because only symmetric quantization is
supported.
The scale value must be either a scalar for per-tensor quantization, or a 1D tensor for per-axis
quantization. The size of the 1-D scale tensor must match the size of the quantization axis. The size of the
scale must match the size of the zeroPt.
The subgraph which terminates with the scale tensor must be a build-time constant. The same restrictions apply
to the zeroPt.
The output type, if constrained, must be constrained to tensorrt.int8 or tensorrt.fp8. The input type, if constrained, must be
constrained to tensorrt.float32, tensorrt.float16 or tensorrt.bfloat16.
The output size is the same as the input size.
IDequantizeLayer supports tensorrt.int8 and tensorrt.float8 precision and will default to tensorrt.int8 precision during instantiation.
IDequantizeLayer supports tensorrt.float32, tensorrt.float16 and tensorrt.bfloat16 output.
- Variables
axis – int
The axis along which dequantization occurs. The dequantization axis is in reference to the input tensor’s dimensions.
to_type – DataType
The specified data type of the output tensor. Must be tensorrt.float32 or tensorrt.float16.
class tensorrt.IScatterLayer#
A Scatter layer as in INetworkDefinition
.
:ivar axis: axis to scatter on when using Scatter Element mode (ignored in ND mode)
:ivar mode: ScatterMode
The operation mode of the scatter.
IIfConditional#
class tensorrt.IIfConditional#
Helper for constructing conditionally-executed subgraphs.
An If-conditional conditionally executes (lazy evaluation) part of the network according
to the following pseudo-code:
If condition is true Then:
output = trueSubgraph(trueInputs);
Else:
output = falseSubgraph(falseInputs);
Emit output
Condition is a 0D boolean tensor (representing a scalar).
trueSubgraph represents a network subgraph that is executed when condition is evaluated to True.
falseSubgraph represents a network subgraph that is executed when condition is evaluated to False.
The following constraints apply to If-conditionals:
- Both the trueSubgraph and falseSubgraph must be defined.
- The number of output tensors in both subgraphs is the same.
- The type and shape of each output tensor from the true/false subgraphs are the same, except that the shapes are allowed to differ if the condition is a build-time constant.
add_input(self: tensorrt.tensorrt.IIfConditional, input: tensorrt.tensorrt.ITensor) → tensorrt.tensorrt.IIfConditionalInputLayer#
Make an input for this if-conditional, based on the given tensor.
- Parameters
input – An input to the conditional that can be used by either or both of the conditional’s subgraphs.
add_output(self: tensorrt.tensorrt.IIfConditional, true_subgraph_output: tensorrt.tensorrt.ITensor, false_subgraph_output: tensorrt.tensorrt.ITensor) → tensorrt.tensorrt.IIfConditionalOutputLayer#
Make an output for this if-conditional, based on the given tensors.
Each output layer of the if-conditional represents a single output of either the true-subgraph or the
false-subgraph of the if-conditional, depending on which subgraph was executed.
- Parameters
true_subgraph_output – The output of the subgraph executed when this conditional’s condition input evaluates to true.
false_subgraph_output – The output of the subgraph executed when this conditional’s condition input evaluates to false.
- Returns
The IIfConditionalOutputLayer
, or None
if it could not be created.
set_condition(self: tensorrt.tensorrt.IIfConditional, condition: tensorrt.tensorrt.ITensor) → tensorrt.tensorrt.IConditionLayer#
Set the condition tensor for this If-Conditional construct.
The condition
tensor must be a 0D data tensor (scalar) with type bool
.
- Parameters
condition – The condition tensor that will determine which subgraph to execute.
- Returns
The IConditionLayer
, or None
if it could not be created.
class tensorrt.IEinsumLayer#
An Einsum layer in an INetworkDefinition
.
This layer implements a summation over the elements of the inputs along dimensions specified by the equation parameter, based on the Einstein summation convention.
The layer can have one or more inputs of rank >= 0. All the inputs must be of same data type. This layer supports all TensorRT data types except bool
.
There is one output tensor of the same type as the input tensors. The shape of output tensor is determined by the equation.
The equation specifies ASCII lower-case letters for each dimension in the inputs in the same order as the dimensions, separated by comma for each input.
The dimensions labeled with the same subscript must match or be broadcastable.
Repeated subscript labels in one input take the diagonal.
Repeating a label across multiple inputs means that those axes will be multiplied.
Omitting a label from the output means values along those axes will be summed.
In implicit mode, the indices which appear once in the expression will be part of the output in increasing alphabetical order.
In explicit mode, the output can be controlled by specifying output subscript labels by adding an arrow (‘->’) followed by subscripts for the output.
For example, “ij,jk->ik” is equivalent to “ij,jk”.
Ellipsis (‘…’) can be used in place of subscripts to broadcast the dimensions.
See the TensorRT Developer Guide for more details on equation syntax.
Many common operations can be expressed using the Einsum equation.
For example:
Matrix Transpose: ij->ji
Sum: ij->
Matrix-Matrix Multiplication: ik,kj->ij
Dot Product: i,i->
Matrix-Vector Multiplication: ik,k->i
Batch Matrix Multiplication: ijk,ikl->ijl
Batch Diagonal: …ii->…i
Note that TensorRT does not support ellipsis or diagonal operations.
- Variables
equation – str
The Einsum equation of the layer.
The equation is a comma-separated list of subscript labels, where each label refers to a dimension of the corresponding tensor.
class tensorrt.IAssertionLayer#
An assertion layer in an INetworkDefinition
.
This layer implements assertions. The input must be a boolean shape tensor. If any element of it is False
, a build-time or run-time error occurs. Asserting equality of input dimensions may help the optimizer.
- Variables
message – string
Message to print if the assertion fails.
class tensorrt.IOneHotLayer#
A OneHot layer in a network definition.
The OneHot layer has three input tensors: Indices, Values, and Depth, one output tensor,
Output, and an axis attribute.
:ivar indices: is an Int32 tensor that determines which locations in Output to set as on_value.
:ivar values: is a two-element (rank=1) tensor that consists of [off_value, on_value]
:ivar depth: is an Int32 shape tensor of rank 0, which contains the depth (number of classes) of the one-hot encoding.
The depth tensor must be a build-time constant, and its value should be positive.
:returns: a tensor with rank = rank(indices)+1, where the added dimension contains the one-hot encoding.
:param axis: specifies to which dimension of the output one-hot encoding is added.
The data types of Output shall be equal to the Values data type.
The output is computed by copying off_values to all output elements, then setting on_value on the indices
specified by the indices tensor.
when axis = 0:
output[indices[i, j, k], i, j, k] = on_value for all i, j, k and off_value otherwise.
when axis = -1:
output[i, j, k, indices[i, j, k]] = on_value for all i, j, k and off_value otherwise.
INonZeroLayer#
class tensorrt.INonZeroLayer#
A NonZero layer in an INetworkDefinition
.
Computes the indices of the input tensor where the value is non-zero. The returned indices are in row-major order.
The output shape is always {D, C}, where D is the number of dimensions of the input and C is the number of non-zero values.
INMSLayer#
class tensorrt.INMSLayer#
A non-maximum suppression layer in an INetworkDefinition
.
Boxes: The input boxes tensor to the layer.
This tensor contains the input bounding boxes. It is a linear tensor of type float32
or float16
.
It has shape [batchSize, numInputBoundingBoxes, numClasses, 4] if the boxes are per class, or
[batchSize, numInputBoundingBoxes, 4] if the same boxes are to be used for each class.
Scores: The input scores tensor to the layer.
This tensor contains the per-box scores. It is a linear tensor of the same type as the boxes tensor.
It has shape [batchSize, numInputBoundingBoxes, numClasses].
MaxOutputBoxesPerClass: The input maxOutputBoxesPerClass tensor to the layer.
This tensor contains the maximum number of output boxes per batch item per class.
It is a scalar (0D tensor) of type int32
.
IoUThreshold is the maximum IoU for selected boxes.
It is a scalar (0D tensor) of type float32
in the range [0.0, 1.0].
It is an optional input with default 0.0.
Use set_input()
to add this optional tensor.
ScoreThreshold is the value that a box score must exceed in order to be selected.
It is a scalar (0D tensor) of type float32
. It is an optional input with default 0.0.
Use set_input()
to add this optional tensor.
The SelectedIndices output tensor contains the indices of the selected boxes.
It is a linear tensor of type int32
. It has shape [NumOutputBoxes, 3].]
Each row contains a (batchIndex, classIndex, boxIndex) tuple.
The output boxes are sorted in order of increasing batchIndex and then in order of decreasing score within each batchIndex.
For each batchIndex, the ordering of output boxes with the same score is unspecified.
If MaxOutputBoxesPerClass is a constant input, the maximum number of output boxes is
batchSize * numClasses * min(numInputBoundingBoxes, MaxOutputBoxesPerClass).
Otherwise, the maximum number of output boxes is batchSize * numClasses * numInputBoundingBoxes.
The maximum number of output boxes is used to determine the upper-bound on allocated memory for this output tensor.
The NumOutputBoxes output tensor contains the number of output boxes in selectedIndices.
It is a scalar (0D tensor) of type int32
.
The NMS algorithm iterates through a set of bounding boxes and their confidence scores,
in decreasing order of score. Boxes are selected if their score is above a given threshold,
and their intersection-over-union (IoU) with previously selected boxes is less than or equal
to a given threshold.
This layer implements NMS per batch item and per class.
For each batch item, the ordering of candidate bounding boxes with the same score is unspecified.
- Variables
bounding_box_format – BoundingBoxFormat
The bounding box format used by the layer. Default is CORNER_PAIRS.
topk_box_limit – int
The maximum number of filtered boxes considered for selection. Default is 2000 for SM 5.3 and 6.2 devices, and 5000 otherwise. The TopK box limit must be less than or equal to {2000 for SM 5.3 and 6.2 devices, 5000 otherwise}.
set_input(self: tensorrt.tensorrt.ILayer, index: int, tensor: tensorrt.tensorrt.ITensor) → None#
Sets the input tensor for the given index.
The indices are as follows:
If this function is called for an index greater or equal to num_inputs
,
then afterwards num_inputs
returns index + 1, and any missing intervening
inputs are set to null. Note that only optional inputs can be missing.
- Parameters
index – The index of the input tensor.
tensor – The input tensor.
class tensorrt.IReverseSequenceLayer#
A ReverseSequence layer in an INetworkDefinition
.
This layer performs batch-wise reversal, which slices the input tensor along the axis batch_axis
. For the
i
-th slice, the operation reverses the first N
elements, specified by the corresponding i
-th value
in sequence_lens
, along sequence_axis
and keeps the remaining elements unchanged. The output tensor will
have the same shape as the input tensor.
- Variables
batch_axis – int
The batch axis. Default: 1.
sequence_axis – int
The sequence axis. Default: 0.
class tensorrt.INormalizationLayer#
A Normalization layer in an INetworkDefinition
.
The normalization layer performs the following operation:
X - input Tensor
Y - output Tensor
S - scale Tensor
B - bias Tensor
Y = (X - Mean(X, axes)) / Sqrt(Variance(X) + epsilon) * S + B
Where Mean(X, axes) is a reduction over a set of axes, and Variance(X) = Mean((X - Mean(X, axes)) ^ 2, axes).
- Variables
epsilon – float
The epsilon value used for the normalization calculation. Default: 1e-5F.
axes – int
The reduction axes for the normalization calculation.
num_groups – int
The number of groups to split the channels into for the normalization calculation. Default: 1.
compute_precision – DataType
The datatype used for the compute precision of this layer. By default TensorRT will run the normalization computation in DataType.kFLOAT32 even in mixed precision mode regardless of any set builder flags to avoid overflow errors. ILayer.precision and ILayer.set_output_type can still be set to control input and output types of this layer. Only DataType.kFLOAT32 and DataType.kHALF are valid for this member. Default: Datatype.FLOAT.
火星上的油条 · error Parsing error: Unexpected token < · Issue #447 · jsx-eslint/eslint-plugin-react · GitHub 8 月前 |
安静的感冒药 · Université de Montréal 11 月前 |