Skip to content

intel/pf-bb-config

Repository files navigation

Introduction

SPDX-License-Identifier: Apache-2.0 Copyright © 2021 Intel Corporation

The Physical Function (PF) Baseband Device (BBDEV) Configuration Application (pf_bb_config) provides a means to configure a baseband device at the host level. The program accesses the configuration space and sets various parameters through memory-mapped I/O (MMIO) reads and writes.

The parameters are parsed from a given configuration file (with .cfg extensions) that is specific to a particular baseband device, although they follow same format.

Building the PF BBDEV Config Application

Build the program by typing the ./build.sh command, this would also update the version string in the source code. The result is that the pf_bb_config binary file is produced.

Usage

The application requires the name of the BBDEV device to configure. Other arguments are optional. The application executes as follows:

./pf_bb_config DEVICE_NAME [-h] [-c CFG_FILE] [-p PCI_ID] [-v VFIO_TOKEN] [-f FFT_LUT_FILE]
  • DEVICE_NAME: Specifies the device to configure; for example: 'FPGA_5GNR' or 'ACC100' or 'VRB1'
  • -h: Prints help
  • -c CFG_FILE: Specifies configuration file to use
  • -p PCI_ID: Specifies PCI ID of device to configure
  • -v VFIO_TOKEN: `VFIO_TOKEN is UUID formatted VFIO VF token required when bound with vfio-pci
  • -f FFT_LUT_FILE: Specifies the FFT LUT bin file to be used if not default location.

Usage Example

This section describes an example of configuring the ACC100 device so that the workload can be run from the Virtual Function (VF). Note the usage of DPDK and bbdev-test (https://doc.dpdk.org/guides/tools/testbbdev.html) for this example.

Additional documentation can be found on DPDK for BBDEV usage (https://doc.dpdk.org/guides/prog_guide/bbdev.html) and within Intel FlexRAN documentation on the Intel Resource & Design Center (https://www.intel.com/content/www/us/en/design/resource-design-center.html).

Using vfio-pci driver

The example in this section uses the vfio-pci module and assumes that Intel® Virtualization Technology for Directed I/O (Intel® VT-d) is enabled.

First the vfio-pcio module must be loaded with the required parameters and there are a few methods available to do this. Note that setting disable_idle_d3 is only strictly required for ACC100.

  • When the module is built-in, it is automally loaded at boot time and the parameters can be passed automatically through the kernel cmdline (grub):

    vfio_pci.enable_sriov=1 vfio_pci.disable_idle_d3=1

  • Another option for built-in module, is to set these parameters manually after boot once module is loaded:

    echo 1 | sudo tee /sys/module/vfio_pci/parameters/enable_sriov

    echo 1 | sudo tee /sys/module/vfio_pci/parameters/disable_idle_d3

  • Last when the vfio-pci is a loadable kernel module the module can be loaded explictly as well:

    modprobe vfio-pci enable_sriov=1 disable_idle_d3=1

Next step is to bind the PF with the vfio-pci module:

$dpdkPath/usertools/dpdk-devbind.py --bind=vfio-pci $pfPciDeviceAddr

Configure the device using the pf_bb_config application for VF usage with both 5G and 4G enabled:

./pf_bb_config ACC100 -v 00112233-4455-6677-8899-aabbccddeeff -c acc100/acc100_config_2vf_4g5g.cfg

Create 2 VFs from the PF using the exposed sysfs interface:

echo 2 | sudo tee /sys/bus/pci/devices/0000:$pfPciDeviceAddr/sriov_numvfs

Test that the VF is functional on the device using bbdev-test:

../../build/app/dpdk-test-bbdev -c F0 -a${VF_PF_PCI_ADDR} --vfio-vf-token=00112233-4455-6677-8899-aabbccddeeff -- -c validation -v ./ldpc_dec_default.data

NOTE:

UUID(Universally Unique Identifier) used in the above example
(00112233-4455-6677-8899-aabbccddeeff) is for testing purpose only.
User should use a random generated UUID in real application, for example
using the tools like `uuidgen`.  The UUID format should match
the standard defined in RFC4122.

For binding PF interface with vfio-pci kernel driver:
VFIO VF token support is upstreamed in Linux kernel version 5.7.
Recommended to use kernel version 5.7 or above or make sure all vfio vf token
feature are back ported to your kernel version.

In case of VFIO mode, pf_bb_config runs daemon mode. To reconfigure first
kill the existing pf_bb_config process using:
pkill pf_bb_config

Notes on the IOMMU usage

To use vfio-pci module make sure to have Intel® VT-d enabled both in the kernel (intel_iommu=on iommu=pt) and in BIOS (as Intel® VT-d). Intel® VT-d is implemented in the Input-Output Memory Management Unit (IOMMU). To check that IOMMU is actually enabled at run time use the following command:

 dmesg | grep "DMAR: IOMMU"

Additional information on vfio can be found on the related kernel documentation page (https://www.kernel.org/doc/Documentation/vfio.txt). There is also valuable information regards linux drivers in the DPDK documentation https://doc.dpdk.org/guides/linux_gsg/linux_drivers.html

Using igb_uio driver on PF (legacy)

The example in this section uses the igb_uio module and assumes that Intel® Virtualization Technology for Directed I/O (Intel® VT-d) is disabled.

Bind the PF with the igb_uio module (or alternatively with pci-pf-stub):

$dpdkPath/usertools/dpdk-devbind.py --bind=igb_uio $pfPciDeviceAddr

Create 2 VFs from the PF using the exposed sysfs interface:

echo 2 | sudo tee /sys/bus/pci/devices/0000:$pfPciDeviceAddr/max_vfs

Bind both VFs using either the igb_uio module (make sure that Intel® VT-d is disabled) or using the vfio-pci module (make sure that Intel® VT-d is enabled):

  • Binding the VFs to igb_uio

    $dpdkPath/usertools/dpdk-devbind.py --bind=igb_uio $vfPciDeviceAddr1 $vfPciDeviceAddr2

  • Binding the VFs to vfio-pci

    $dpdkPath/usertools/dpdk-devbind.py --bind=vfio-pci $vfPciDeviceAddr1 $vfPciDeviceAddr2

Configure the devices using the pf_bb_config application for VF usage with both 5G and 4G enabled:

./pf_bb_config ACC100 -c acc100/acc100_config_2vf_4g5g.cfg

Test that the VF is functional on the device using bbdev-test:

../../build/app/dpdk-test-bbdev -c F0 -a${VF_PF_PCI_ADDR} -- -c validation -v ./ldpc_dec_default.data

For DPDK versions older than 20.11, the previous command must be modified as follows:

../../${RTE_TARGET}/app/testbbdev -c F0 -w${VF_PF_PCI_ADDR} -- -c validation -v ./ldpc_dec_default.data

Details for VRB1/2 Configuration (Intel® vRAN Boost)

  • Note that VRB1 refers to Intel® vRAN Boost v1.0 for 4th Gen Intel Xeon Scalable Processor (SPR-EE), while VRB2 refers to Intel® vRAN Boost v2.0 for Intel® Xeon® D Processor (GNR-D). The bbdev PMDs are described in DPDK (https://doc.dpdk.org/guides/bbdevs/vrb1.html https://doc.dpdk.org/guides/bbdevs/vrb2.html)

  • A number of configuration file examples are available in the vrb1 and vrb2 directories. These examples include the following parameters, which can be modified for a specific use case.

  • The pf_mode_en parameter refers to the case for which the workload is run from PF, or alternatively from the VF (Single Root I/O Virtualization (SR-IOV)). Default usage is using SR-IOV hence this would be set to 0.

  • On VRB1 there are 16 Queue Groups (QGroups) available (made of up to 16 queues x 16 VFs each), which can be allocated to any available operation (4GUL/4GDL/5GUL/5GDL/FFT) based on the num_qgroups parameter. For example, 4x QGroups for 5GUL and 4x QGroups for 5GDL can be allocated when only 5G processing is required, and these QGroups cover four distinct incremental priority levels. The only limitation (which would be flagged by the configuration tool if exceeded) is for the total number of QGroups x num_aqs x aq_depth x vf_bundles to be less than 64K. (The num_aqs parameter is the number of atomic queues, the aq_depth parameter is the atomic queue depth, and the vf_bundles parameter is the number of VF bundles.).

  • The num_aqs_per_groups defines the number of atomic queues within a VF bundle for a given QGroup. The range is 1 to 16 on VRB1, with 16 being the default and recommended value. There can be a maximum of 16 * 16 * 16 = 4K total queues.

  • An example of default configuration file may be vrb1/vrb1_config_16vf.cfg which includes the configuration for 16 VFs in SRIOV mode with a total of 16 x 16 queues per VF covering all the possible processing engine functions.

    ./pf_bb_config VRB1 -v 00112233-4455-6677-8899-aabbccddeeff -c vrb1/vrb1_config_16vf.cfg

  • On VRB2 the number of supported queues is increased compared to VRB1: the total number of QGroups is increased from 16 to 32, and the number of AQs per group from 16 to 64 and the number of VF bundles from 16 to 64, while the AQ depth is set to 32 by default. This also now support a new set of queue for MLD-TS operation type. A default reference configuration file to consider is vrb2/vrb2_config_vf.cfg.

    ./pf_bb_config VRB2 -v 00112233-4455-6677-8899-aabbccddeeff -c vrb2/vrb2_config_vf.cf

  • The FFT engines includes an internal memory used to define the semi-static windowing parameters. The content of that LUT table is loaded from binary files vrbx/srs_fft_windows_coefficient.bin (up to 1024 points across 16 windows and 7 iDFT sizes). There is extra description of these windows in vrbx/srs_fft_windows_coefficient.txt.

  • It is recommended to use vfio-pci to benefit from all Intel vRAN Boost features in pf_bb_config.

Error reporting and recovery (specific to VRB1 and VRB2)

It is not expected for the accelerator to get into a bad state under any condition. Still the pf_bb_config supports the capability of detecting, reporting and recovering any potential fatal error forcing the device into a bad state. In the unlikely event case this would be happening the pf_bb_config log (/var/log/pf_bb_cfg_0000:f7:00.0.log) would include information related to such events as captured in the example log below.

Example capturing error information and automatic recovery: Sat Dec 3 00:12:56 2022:ERR:HI Device Error: 0 5G_EXTRA_COMPLETION_RECVD Sat Dec 3 00:12:56 2022:ERR:Ext Device Error: 20 EXTRA_READ_STATUS_5G Sat Dec 3 00:12:56 2022:ERR:Fatal error Sat Dec 3 00:12:56 2022:INFO:Cluster reset and reconfig Sat Dec 3 00:12:57 2022:INFO:Queue Groups UL4G 2 DL4G 2 UL5G 4 DL5G 4 FFT 4 Sat Dec 3 00:12:57 2022:INFO:Configuration in VF mode Sat Dec 3 00:12:58 2022:INFO:VRB1 configuration complete Sat Dec 3 00:12:58 2022:INFO:VRB1 PF [0000:f7:00.0] Reconfiguration complete!

Details for ACC100 Configuration (aka Mount Bryce)

  • A number of configuration file examples are available in the acc100 directory. These examples include the following parameters, which can be modified for a specific use case.

  • The pf_mode_en parameter refers to the case for which the workload is run from PF, or alternatively from the VF (Single Root I/O Virtualization (SR-IOV)).

  • There are eight Queue Groups (QGroups) available (made of up to 16 queues x 16 VFs each), which can be allocated to any available operation (4GUL/4GDL/5GUL/5GDL) based on the num_qgroups parameter. For example, 4x QGroups for 5GUL and 4x QGroups for 5GDL can be allocated when only 5G processing is required, and these QGroups cover four distinct incremental priority levels. The only limitation (which would be flagged by the configuration tool if exceeded) is for the total number of QGroups x num_aqs x aq_depth x vf_bundles to be less than 32K. (The num_aqs parameter is the number of atomic queues, the aq_depth parameter is the atomic queue depth, and the vf_bundles parameter is the number of VF bundles.).

  • The default DDR size allocated per VF depends on the num_vf_bundles set (1 to 16). Assuming 4GB of DDR is populated on the board, which is then split between VFs.

  • The num_aqs_per_groups defines the number of atomic queues within a VF bundle for a given QGroup. The range is 1 to 16, with 16 being the default value. There can be a maximum of 16 * 8 * 16 = 2048 total queues.

  • Note that if an early version of the ACC100 device is used (Engineering Sample prior to PRQ), then pf_bb_config displays: Note: Not on DDR PRQ version 1302020 != 10092020. Intel recommends using the PRQ version of the ACC100 device.

Details for Intel® FPGA Programmable Acceleration Card N3000 and N6000 Configuration

  • A number of configuration file examples are available in the fpga_5gnr and fpga_lte directories. These examples include the following parameters, which can be modified for a specific usecase.

  • The pf_mode_en parameter refers to the case for which the workload is run from PF, or alternatively from the VF (Single Root I/O Virtualization (SR-IOV)).

  • The vfqmap list parameter defines how many queues are mapped for each VF, the absolute total being 32 for either UL or DL. As an example: "16, 16, 0, 0...", which means that two VFs would be supported with 16 queues each. All queues have the same priority level. This parameter is set independently for the actual number of VFs exposed through the SR-IOV PCIe configuration capability through the sysfs interface.

  • The parameters bandwidth and load_balance are optional parameters impacting arbitration but are expected to be kept as default.

Device Reset Procedure Using PF FLR and VF FLR

The accelerators are exposed as a PCIe endpoint and therefore can be reset using Function-Level Reset (FLR) as described in the PCI Express® Base Specification Revision 3.1a. The FLR mechanism enables software to quiesce and reset endpoint hardware with Function-Level granularity, that is, for either the Physical Function (PF) or one of the Virtual Functions (VFs):

  • When the function targeted is the PF, then the full device is reset. The device must then be reconfigured using pf_bb_config as is after a cold start.
  • When the function targeted is the VF, then only the internal state related to that same VF is impacted (queues are flushed, VF registers reset back to default values). This VF can then still be used without requiring pf_bb_config to be run again (that is, running the bbdev-test command as described in the Example section).

In an application, FLR can be triggered from the sysfs interface, as long as these devices are bounded by using the following commands. The actual bus:device.function (BDF) can be retrieved using the lspci command. Trigger PF FLR with the following command:

echo 1 >> /sys/bus/pci/devices/0000\:${PF_PCI_ADDR}/reset

Trigger VF FLR on one of the VFs (the one under $(VF_PCI_ADDR) BDF) with the following command:

echo 1 >> /sys/bus/pci/devices/0000\:${VF_PCI_ADDR}/reset

CLI interface

When running in daemon mode (VFIO mode) the pf_bb_config application is running as a service and exposes a socket for CLI interaction. This can notably being used for telemetry, register dump capture or other interactions at run time:

  • RESET_MODE_CMD_ID -> To set reset mode of ACC devices (pf_flr / cluster_reset).
  • AUTO_RESET_CMD_ID -> To set auto reset mode of ACC devices (on / off).
  • CLEAR_LOG_CMD_ID -> To clear the previous content of logfile.
  • EXIT_APP_CMD_ID -> To gracefully shutdown the application.
  • REG_DUMP_CMD_ID -> To dump all the register status of the ACC device (DEVICE_ID).
  • RECONFIG_ACC_CMD_ID -> To reconfigure the ACC device (DEVICE_ID).
  • MM_READ_CMD_ID -> To read/write to a register.
  • DEVICE_DATA_CMD_ID -> To dump telemetry data.

Conversion to PDF

To convert this readme to a PDF file, first install the npm tool, followed by installing the markdown-pdf conversion program:

sudo yum install npm
npm install -g markdown-pdf

Then convert to PDF by running the markdown-pdf conversion program:

markdown-pdf README.md README.pdf

License

SPDX-License-Identifier: Apache-2.0 (see included LICENSE file located at https://github.com/intel/pf-bb-config/blob/master/LICENSE)

Copyright(c) 2021 Intel Corporation