In oneDNN Graph API programming model, a computation graph is passed to library and then optimized sub-graphs which are called
partitions are returned by the library.
Partition is decided by oneDNN Graph API implementation. It is the key concept to satisfy the different needs of AI hardware classes by using a unified API. Typically can compile
tensor data, and execute
The key concepts in oneDNN Graph API include
compiled partition, and
tensor. Here is the relationship between these entities. Besides, oneDNN Graph API shares the common
stream concepts of oneDNN primitive API.
Logical tensor (dnnl::graph::logical_tensor) describes the metadata of the input and output tensors, like data type, number of dimensions, size for each dimension, tensor layout and property. Each logical tensor has a unique ID which is immutable during the lifetime of a logical tensor. The metadata of a logical tensor cannot be modified without creating a new one.
Op (dnnl::graph::op) represents an operation as part of a computation graph. An operation has kind, attribute, and input and output logical tensors. Operations are added into a graph object to construct a computation graph. As both operation and logical tensor contain a unique ID, the graph object knows how to connect a producer operation to a consumer operation through a logical tensor as the edge between them.
Graph (dnnl::graph::graph) contains a set of operations. A graph object is associated to a specific engine kind (dnnl::engine::kind). Multiple operations can be added (dnnl::graph::graph::add_op) along with input and output logical tensors to a graph. After finishing adding operations, finalization API (dnnl::graph::graph::finalize) can be called to indicate that the graph is ready for partitioning. By calling partitioning API (dnnl::graph::graph::get_partitions), a group of partitions from the graph will be returned .
Partition (dnnl::graph::partition) represents a collection of operations identified by library implementation as the basic unit for compilation and execution. A partition is a connected subgraph within the source graph. The partitions returned from the library must not form a dependence cycle.
A partition needs to be compiled (dnnl::graph::partition::compile) before execution. The compilation lowers down the computation logic to hardware ISA level and generates binary code. The generated code is specialized for the input and output logical tensors and engine (dnnl::engine).
The output logical tensors can have unknown dimensions during compilation. In this case, the compilation procedure should deduce the output shapes according to the input shapes and will return an error if the output shapes cannot deduced determinately. The output logical tensors can also have
any layout type (dnnl::graph::logical_tensor::layout_type::any). It means that the compilation procedure can choose the optimal layouts for the output tensors. Optimal layouts are represented as opaque layout IDs and saved in the corresponding output logical tensors.
A partition may contains many logical tensors with part of them are internal intermediate results connecting two operations inside the partition. The required inputs and outputs of a partition are also called
ports of a partition. Two APIs
get_input_ports (dnnl::graph::partition::get_input_ports) and
get_output_ports (dnnl::graph::partition::get_output_ports) are provided to query the ports and help understand which input logical tensors and output logical tensors are needed to compile a partition. The input logical tensors and output logical tensors must match IDs with ports. These in ports and out ports can also be used to track the producer and consumer of a partitions through logical tensor IDs and for framework integration, connect the partition back to the framework graph as a custom node.
Compiled partition (dnnl::graph::compiled_partition) represents the generated code specialized for a target hardware and tensor metadata passed through compilation API. To execute a compiled partition (dnnl::graph::compiled_partition::execute), both input and output tensors, and a stream (dnnl::stream) are required to pass. Input and output tensors must bind data buffers to the input and output logical tensors respectively.
An API (dnnl::graph::compiled_partition::query_logical_tensor) is provided to query output logical tensors from a compiled partition. It allows to know the output layout and memory size (dnnl::graph::logical_tensor::get_mem_size) when they specify output logical tensor with
any layout type during compilation.
Tensor (dnnl::graph::tensor) is an abstraction for multi-dimensional input and output data which is needed in the execution of a compiled partition. A tensor contains a logical tensor, an engine (dnnl::engine), and a data handle. The application is responsible for managing the data handle’s lifecycle, for example free the memory resource when it is not used anymore.