DML_ELEMENT_WISE_BIT_NOT_OPERATOR_DESC structure (directml.h)
Computes the bitwise NOT for each element of the input tensor, and writes the result into the output tensor.
The bitwise operation is applied to tensor data in its native encoding. Therefore, the tensor data type is ignored except for determining the width of each element.
This operator supports in-place execution, meaning that the output tensor is permitted to alias the input tensor during binding.
Syntax
struct DML_ELEMENT_WISE_BIT_NOT_OPERATOR_DESC {
const DML_TENSOR_DESC *InputTensor;
const DML_TENSOR_DESC *OutputTensor;
};
Members
InputTensor
Type: const DML_TENSOR_DESC*
The input tensor to read from.
OutputTensor
Type: const DML_TENSOR_DESC*
The output tensor to write the results to.
Example
InputTensor: (Sizes:{2,2}, DataType:UINT8)
[[0, 128], // 0b00000000, 0b10000000
[42, 255]] // 0b00101010, 0b11111111
OutputTensor: (Sizes:{2,2}, DataType:UINT8)
[[255, 127], // 0b11111111, 0b01111111
[213, 0 ]] // 0b11010101, 0b00000000
Availability
This operator was introduced in DML_FEATURE_LEVEL_3_0
.
Tensor constraints
InputTensor and OutputTensor must have the same DataType, DimensionCount, and Sizes.
Tensor support
DML_FEATURE_LEVEL_4_1 and above
Tensor | Kind | Supported dimension counts | Supported data types |
---|---|---|---|
InputTensor | Input | 1 to 8 | FLOAT64, FLOAT32, FLOAT16, INT64, INT32, INT16, INT8, UINT64, UINT32, UINT16, UINT8 |
OutputTensor | Output | 1 to 8 | FLOAT64, FLOAT32, FLOAT16, INT64, INT32, INT16, INT8, UINT64, UINT32, UINT16, UINT8 |
DML_FEATURE_LEVEL_3_0 and above
Tensor | Kind | Supported dimension counts | Supported data types |
---|---|---|---|
InputTensor | Input | 1 to 8 | UINT32, UINT16, UINT8 |
OutputTensor | Output | 1 to 8 | UINT32, UINT16, UINT8 |
Requirements
Requirement | Value |
---|---|
Minimum supported client | Windows 10 Build 20348 |
Minimum supported server | Windows 10 Build 20348 |
Header | directml.h |