Design Topology - 2023.2 English

Vitis Unified Software Platform Documentation: Application Acceleration Development (UG1393)

Document ID
UG1393
Release Date
2023-12-13
Version
2023.2 English

In the Vitis core development kit, Alveo Data Center acceleration cards and custom embedded platforms provide a foundation for designs. Targeted devices can include AMD Versal™ adaptive SoCs, Zynq UltraScale+ MPSoCs, Kria SOMs, or AMD UltraScale+™ FPGAs. These devices contain a programmable logic (PL) region that loads and executes a device binary (.xclbin) file that contains and connects PL kernels as compiled object (.xo) files and can also contain AI Engine graphs.

Extensible Alveo acceleration cards and custom embedded platforms contain one or more interfaces to global memory (DDR or HBM), and optional streaming interfaces connected to other resources such as AI Engines and external I/O. PL kernels can access data through global memory interfaces (m_axi) or streaming interfaces (axis). The memory interfaces of PL kernels must be connected to memory interfaces of the extensible platform. The streaming interfaces of PL kernels can be connected to any streaming interfaces of the platform, of other PL kernels, or of the AI Engine array. Both memory-based and streaming connections are defined through Vitis linking options, as described in Linking the System.

Multiple kernels (.xo) can be implemented in the PL region of the AMD device binary (.xclbin), allowing for significant application acceleration. A single kernel can also be instantiated multiple times. The number of instances, or compute units of a kernel is programmable up to 31, and determined by linking options specified when building the device binary.

For Versal devices the .xclbin file can also contains the compiled AI Engine graph application (libadf.a). The graph application consists of nodes and edges where nodes represent compute kernel functions, and edges represent data connections. Kernel functions are the fundamental building blocks of an ADF graph application. Kernels operate on data streams, consuming input blocks of data and producing output blocks of data. The libadf.a and PL kernels (.xo) are linked with the target platform (.xpfm) to define the hardware design. The AI Engine can be driven by PL kernels through axis interfaces. The AI Engine can also be controlled through the Arm processor (PS) via runtime parameters (RTP) in the graph and GMIO on Versal adaptive SoC devices. Refer to AI Engine Tools and Flows User Guide (UG1076) for more information.