tf.raw_ops.QuantizedMatMulWithBiasAndDequantize
tf.raw_ops.QuantizedMatMulWithBiasAndDequantize(
a,
b,
bias,
min_a,
max_a,
min_b,
max_b,
min_freezed_output,
max_freezed_output,
Toutput,
transpose_a=False,
transpose_b=False,
input_quant_mode='MIN_FIRST',
name=None
)
| Args |
a | A Tensor. Must be one of the following types: qint8, quint8, qint32, qint16, quint16. |
b | A Tensor. Must be one of the following types: qint8, quint8, qint32, qint16, quint16. |
bias | A Tensor. Must be one of the following types: float32, qint32. |
min_a | A Tensor of type float32. |
max_a | A Tensor of type float32. |
min_b | A Tensor of type float32. |
max_b | A Tensor of type float32. |
min_freezed_output | A Tensor of type float32. |
max_freezed_output | A Tensor of type float32. |
Toutput | A tf.DType from: tf.float32. |
transpose_a | An optional bool. Defaults to False. |
transpose_b | An optional bool. Defaults to False. |
input_quant_mode | An optional string from: "MIN_FIRST", "SCALED". Defaults to "MIN_FIRST". |
name | A name for the operation (optional). |
| Returns |
A Tensor of type Toutput. |