.. _l-onnx-doc-QLinearConv:
===========
QLinearConv
===========
.. contents::
:local:
.. _l-onnx-op-qlinearconv-10:
QLinearConv - 10
================
**Version**
* **name**: `QLinearConv (GitHub) `_
* **domain**: **main**
* **since_version**: **10**
* **function**: False
* **support_level**: SupportType.COMMON
* **shape inference**: True
This version of the operator has been available
**since version 10**.
**Summary**
The convolution operator consumes a quantized input tensor, its scale and zero point,
a quantized filter, its scale and zero point, and output's scale and zero point,
and computes the quantized output. Each scale and zero-point pair must have same shape.
It means they must be either scalars (per tensor) or 1-D tensors (per output channel).
Each input or output and its related zero point must have same type.
When bias is present it must be quantized using scale = input scale * weight scale and
zero point as 0.
**Attributes**
* **auto_pad**:
auto_pad must be either NOTSET, SAME_UPPER, SAME_LOWER or VALID.
Where default value is NOTSET, which means explicit padding is used.
SAME_UPPER or SAME_LOWER mean pad the input so that `output_shape[i]
= ceil(input_shape[i] / strides[i])` for each axis `i`. The padding
is split between the two sides equally or almost equally (depending
on whether it is even or odd). In case the padding is an odd number,
the extra padding is added at the end for SAME_UPPER and at the
beginning for SAME_LOWER. Default value is ``'NOTSET'``.
* **dilations**:
dilation value along each spatial axis of the filter. If not
present, the dilation defaults to 1 along each spatial axis.
* **group**:
number of groups input channels and output channels are divided
into. default is 1. Default value is ``1``.
* **kernel_shape**:
The shape of the convolution kernel. If not present, should be
inferred from input 'w'.
* **pads**:
Padding for the beginning and ending along each spatial axis, it can
take any value greater than or equal to 0.The value represent the
number of pixels added to the beginning and end part of the
corresponding axis.`pads` format should be as follow [x1_begin,
x2_begin...x1_end, x2_end,...], where xi_begin the number ofpixels
added at the beginning of axis `i` and xi_end, the number of pixels
added at the end of axis `i`.This attribute cannot be used
simultaneously with auto_pad attribute. If not present, the padding
defaultsto 0 along start and end of each spatial axis.
* **strides**:
Stride along each spatial axis. If not present, the stride defaults
to 1 along each spatial axis.
**Inputs**
Between 8 and 9 inputs.
* **x** (heterogeneous) - **T1**:
Input data tensor from previous layer; has size (N x C x H x W),
where N is the batch size, C is the number of channels, and H and W
are the height and width. Note that this is for the 2D image.
Otherwise the size is (N x C x D1 x D2 ... x Dn). Optionally, if
dimension denotation is in effect, the operation expects input data
tensor to arrive with the dimension denotation of [DATA_BATCH,
DATA_CHANNEL, DATA_FEATURE, DATA_FEATURE ...].
* **x_scale** (heterogeneous) - **tensor(float)**:
Scale tensor for input 'x'. It's a scalar, which means a per-
tensor/layer quantization.
* **x_zero_point** (heterogeneous) - **T1**:
Zero point tensor for input 'x'. It's a scalar, which means a per-
tensor/layer quantization.
* **w** (heterogeneous) - **T2**:
The weight tensor that will be used in the convolutions; has size (M
x C/group x kH x kW), where C is the number of channels, and kH and
kW are the height and width of the kernel, and M is the number of
feature maps. For more than 2 dimensions, the kernel shape will be
(M x C/group x k1 x k2 x ... x kn), where (k1 x k2 x ... kn) is the
dimension of the kernel. Optionally, if dimension denotation is in
effect, the operation expects the weight tensor to arrive with the
dimension denotation of [FILTER_OUT_CHANNEL, FILTER_IN_CHANNEL,
FILTER_SPATIAL, FILTER_SPATIAL ...]. X.shape[1] == (W.shape[1] *
group) == C (assuming zero based indices for the shape array). Or in
other words FILTER_IN_CHANNEL should be equal to DATA_CHANNEL.
* **w_scale** (heterogeneous) - **tensor(float)**:
Scale tensor for input 'w'. It could be a scalar or a 1-D tensor,
which means a per-tensor/layer or per output channel quantization.
If it's a 1-D tensor, its number of elements should be equal to the
number of output channels (M).
* **w_zero_point** (heterogeneous) - **T2**:
Zero point tensor for input 'w'. It could be a scalar or a 1-D
tensor, which means a per-tensor/layer or per output channel
quantization. If it's a 1-D tensor, its number of elements should be
equal to the number of output channels (M).
* **y_scale** (heterogeneous) - **tensor(float)**:
Scale tensor for output 'y'. It's a scalar, which means a per-
tensor/layer quantization.
* **y_zero_point** (heterogeneous) - **T3**:
Zero point tensor for output 'y'. It's a scalar, which means a per-
tensor/layer quantization.
* **B** (optional, heterogeneous) - **T4**:
Optional 1D bias to be added to the convolution, has size of M. Bias
must be quantized using scale = x_scale * w_scale and zero_point = 0
**Outputs**
* **y** (heterogeneous) - **T3**:
Output data tensor that contains the result of the convolution. The
output dimensions are functions of the kernel size, stride size, and
pad lengths.
**Type Constraints**
* **T1** in (
tensor(int8),
tensor(uint8)
):
Constrain input type to 8-bit integer tensor.
* **T2** in (
tensor(int8),
tensor(uint8)
):
Constrain filter type to 8-bit integer tensor.
* **T3** in (
tensor(int8),
tensor(uint8)
):
Constrain output type to 8-bit integer tensor.
* **T4** in (
tensor(int32)
):
Constrain bias type to 32-bit integer tensor.
**Examples**
**default**
::
node = onnx.helper.make_node('QLinearConv',
inputs=['x', 'x_scale', 'x_zero_point', 'w', 'w_scale', 'w_zero_point', 'y_scale', 'y_zero_point'],
outputs=['y'],)
x = np.array([[255, 174, 162, 25, 203, 168, 58],
[15, 59, 237, 95, 129, 0, 64],
[56, 242, 153, 221, 168, 12, 166],
[232, 178, 186, 195, 237, 162, 237],
[188, 39, 124, 77, 80, 102, 43],
[127, 230, 21, 83, 41, 40, 134],
[255, 154, 92, 141, 42, 148, 247], ], dtype=np.uint8).reshape((1, 1, 7, 7))
x_scale = np.float32(0.00369204697)
x_zero_point = np.uint8(132)
w = np.array([0], dtype=np.uint8).reshape((1, 1, 1, 1))
w_scale = np.array([0.00172794575], dtype=np.float32)
w_zero_point = np.array([255], dtype=np.uint8)
y_scale = np.float32(0.00162681262)
y_zero_point = np.uint8(123)
output = np.array([[0, 81, 93, 230, 52, 87, 197],
[240, 196, 18, 160, 126, 255, 191],
[199, 13, 102, 34, 87, 243, 89],
[23, 77, 69, 60, 18, 93, 18],
[67, 216, 131, 178, 175, 153, 212],
[128, 25, 234, 172, 214, 215, 121],
[0, 101, 163, 114, 213, 107, 8], ], dtype=np.uint8).reshape((1, 1, 7, 7))
expect(node, inputs=[x, x_scale, x_zero_point, w, w_scale, w_zero_point, y_scale, y_zero_point], outputs=[output],
name='test_qlinearconv')