SPDX-License-Identifier: Apache-2.0
Copyright (c) 2019-2021 Intel Corporation
This document provides information on high-level system features, issues, and limitations for Intel® Smart Edge Open and Open Network Edge Services Software (OpenNESS). Beginning with release 21.09, OpenNESS has been rebranded as Intel® Smart Edge Open.
- Release history
- Changes to Existing Features
- Fixed Issues
- Known Issues and Limitations
- Release Content
- Hardware and Software Compatibility
- Supported Operating Systems
- Packages Version
- Edge Cloud Deployment options
- Controller-based deployment of Applications in Docker Containers/VM–using-Libvirt
- Controller + Kubernetes* based deployment of Applications in Docker* Containers
- OpenNESS Controller
- Support for Edge Node Orchestration
- Support for Web UI front end
- OpenNESS APIs
- Edge Application APIs
- Edge Virtualization Infrastructure APIs
- Edge Application life cycle APIs
- Core Network Configuration APIs
- Edge Application authentication APIs
- OpenNESS Controller APIs
- Platform Features
- Microservices based Appliance and Controller agent deployment
- Support for DNS for the edge
- CentOS* 7.6 / CentOS 7.6 + RT kernel
- Basic telemetry support
- Sample Reference Applications
- OpenVINO™ based Consumer Application
- Producer Application supporting OpenVINO™
- Dataplane
- DPDK/KNI based Dataplane – NTS
- Support for deployment on IP, LTE (S1, SGi and LTE CUPS)
- Cloud Adapters
- Support for running Amazon* Greengrass* cores as an OpenNESS application
- Support for running Baidu* Cloud as an OpenNESS application
- Documentation
- User Guide Enterprise and Operator Edge
- OpenNESS Architecture
- Swagger/Proto buff External API Guide
- 4G/CUPS API whitepaper
- Cloud Connector App note
- OpenVINO™ on OpenNESS App note
- Edge Cloud Deployment options
- Async method for image download to avoid timeout.
- Dataplane
- Support for OVN/OVS based Dataplane and network overlay for Network Edge (based on Kubernetes)
- Cloud Adapters
- Support for running Amazon Green grass cores as an OpenNESS application with OVN/OVS as Dataplane and network overlay
- Support for Inter-App comms
- Support for OVS-DPDK or Linux* bridge or Default interface for inter-Apps communication for OnPrem deployment
- Accelerator support
- Support for HDDL-R accelerator for interference in a container environment for OnPrem deployment
- Edge Applications
- Early Access Support for Open Visual Cloud (OVC) based Smart City App on OpenNESS OnPrem
- Support for Dynamic use of VPU or CPU for Inferences
- Gateway
- Support for Edge node and OpenNESS Controller gate way to support route-ability
- Documentation
- OpenNESS Architecture (update)
- OpenNESS Support for OVS as dataplane with OVN
- Open Visual Cloud Smart City Application on OpenNESS - Solution Overview
- Using Intel® Movidius™ Myriad™ X High Density Deep Learning (HDDL) solution in OpenNESS
- OpenNESS How-to Guide (update)
- Hardware
- Support for Cascade lake 6252N
- Support for Intel® FPGA Programmable Acceleration Card (Intel® FPGA PAC) N3000
- Edge Application
- Fully cloud native Open Visual Cloud Smart City Application pipeline on OpenNESS Network edge.
- Edge cloud
- EAA and CNCA microservice as native Kubernetes-managed services
- Support for Kubernetes version 1.16.2
- Edge Compute EPA features support for Network Edge
- CPU Manager: Support deployment of POD with dedicated pinning
- SRIOV NIC: Support deployment of POD with dedicated SRIOV VF from NIC
- SRIOV FPGA: Support deployment of POD with dedicated SRIOV VF from FPGA
- Topology Manager: Support k8s to manage the resources allocated to workloads in a NUMA topology-aware manner
- BIOS/FW Configuration service - Intel SysCfg based BIOS/FW management service
- Hugepages: Support for allocation of 1G/2M huge pages to the Pod
- Multus: Support for Multiple network interface in the PODs deployed by Kubernetes
- Node Feature discovery: Support detection of Silicon and Software features and automation of deployment of CNF and Applications
- FPGA Remote System Update service: Support the Open Programmable Acceleration Engine (OPAE) (fpgautil) based image update service for FPGA
- Non-Privileged Container: Support deployment of non-privileged pods (CNFs and Applications as reference)
- Edge Compute EPA features support for On-Premises
- Using Intel® Movidius™ Myriad™ X High Density Deep Learning (HDDL) solution in OpenNESS
- Converged Edge Experience Kits for Network and OnPremises edge
- Offline Release Package: Customers should be able to create an installer package that can be used to install OnPremises version of OpenNESS without the need for Internet access.
- 5G NR Edge Cloud deployment support
- 5G NR edge cloud deployment support with SA mode
- AF: Support for 5G NGC Application function as a microservice
- NEF: Support for 5G NGC Network Exposure function as a microservice
- Support for 5G NR UDM, UPF, AMF, PCF and SCF (not part of the release)
- DNS support
- DNS support for UE
- DNS Support for Edge applications
- Documentation
- Completely reorganized documentation structure for ease of navigation
- 5G NR Edge Cloud deployment Whitepaper
- EPA application note for each of the features
- OVN/OVS-DPDK support for dataplane
- Network Edge: Support for kube-ovn CNI with OVS or OVS-DPDK as dataplane. Support for Calico as CNI.
- OnPremises Edge: Support for OVS-DPDK CNI with OVS-DPDK as dataplane supporting application deployed in containers or VMs
- Support for VM deployments on Kubernetes mode
- Kubevirt based VM deployment support
- EPA Support for SRIOV Virtual function allocation to the VMs deployed using Kubernetes
- EPA support - OnPremises
- Support for dedicated core allocation to applications running as VMs or Containers
- Support for dedicated SRIOV VF allocation to applications running in VM or containers
- Support for system resource allocation into the application running as a container
- Mount point for shared storage
- Pass environment variables
- Configure the port rules
- Core Network Feature (5G)
- PFD Management API support (3GPP 23.502 Sec. 52.6.3 PFD Management service)
- AF: Added support for PFD Northbound API
- NEF: Added support for PFD southbound API, and Stubs to loopback the PCF calls.
- kubectl: Enhanced CNCA kubectl plugin to configure PFD parameters
- WEB UI: Enhanced CNCA WEB UI to configure PFD params in OnPerm mode
- Auth2 based authentication between 5G Network functions: (as per 3GPP Standard)
- Implemented oAuth2 based authentication and validation
- AF and NEF communication channel is updated to authenticated based on oAuth2 JWT token in addition to HTTP2.
- HTTPS support
- Enhanced the 5G OAM, CNCA (web-ui and kube-ctl) to HTTPS interface
- PFD Management API support (3GPP 23.502 Sec. 52.6.3 PFD Management service)
- Modular Playbook
- Support for customers to choose real-time or non-realtime kernel for an edge node
- Support for customers to choose CNIs - Validated with Kube-OVN and Calico
- Edge Apps
- FlexRAN: Dockerfile and pod specification for the deployment of 4G or 5G FlexRAN
- AF: Dockerfile and pod specification
- NEF: Dockerfile and pod specification
- UPF: Dockerfile and pod specification
- OpenNESS is now available in two distributions
- Open source (Apache 2.0 license)
- Intel Distribution of OpenNESS (Intel Proprietary License)
- Includes all the code from the open source distribution plus additional features and enhancements to improve the user experience
- Access requires a signed license. A request for access can be made at smartedge-open.org by navigating to the "Products" section and selecting "Intel Distribution of OpenNESS"
- Both distributions are hosted at github.com/smart-edge-open
- On premises configuration now optionally supports Kubernetes
- Dataplane
- Support for Calico eBPF as CNI
- Performance baselining of the CNIs
- Visual Compute and Media Analytics
- Intel Visual Cloud Accelerator Card - Analytics (VCAC-A) Kubernetes deployment support (CPU, GPU, and VPU)
- Node feature discovery of VCAC-A
- Telemetry support for VCAC-A
- Provide ansible and Helm -playbook support for OVC codecs Intel® Xeon® CPU mode - video analytics service (REST API) for developers
- Edge Applications
- Smart City Application Pipeline supporting CPU or VCAC-A mode with Helm chart
- CDN Content Delivery using NGINX with SR-IOV capability for higher performance with Helm chart
- CDN transcode sample application using Intel® Xeon® CPU optimized media SDK with Helm chart
- Support for Transcoding Service using Intel® Xeon® CPU optimized media SDK with Helm chart
- Intel Edge Insights application support with Helm chart
- Edge Network Functions
- FlexRAN DU with Helm Chart (FlexRAN not part of the release)
- xRAN Fronthaul with Helm CHart (xRAN app not part of the release)
- Core Network Function - Application Function with Helm chart
- Core Network Function - Network Exposure Function With Helm chart
- Core Network Function - UPF (UPF app not part of the release)
- Core network Support functions - OAM and CNTF
- Helm Chart for Kubernetes enhancements
- NFD, CMK, SRIOV-Device plugin and Multus*
- Support for local Docker registry setup
- Support for deployment-specific Flavors
- Minimal
- RAN - 4G and 5G
- Media Analytics with VCAC-A and with CPU only mode
- CDN - Transcode
- CDN - Content Delivery
- Azure - Deployment of OpenNESS cluster on Microsoft* Azure* cloud
- Support for OpenNESS on CSP Cloud
- Azure - Deployment of OpenNESS cluster on Microsoft* Azure* cloud
- Telemetry Support
- Support for Collectd backend with hardware from Intel and custom metrics
- cpu, cpufreq, load, hugepages, intel_pmu, intel_rdt, ipmi, ovs_stats, ovs_pmd_stats
- FPGA – PACN3000 (collectd) - Temp, Power draw
- VPU Device memory, VPU device thermal, VPU Device utilization
- Open Telemetry - Support for collector and exporter for metrics (e.g., heartbeat from app)
- Support for PCM counter for Prometheus* and Grafana*
- Telemetry Aware Scheduler
- Early Access support for Resource Management Daemon (RMD)
- RMD for cache allocation to the application Pods
- Ability to deploy OpenNESS Master and Node on the same platform
- Native On-premises mode
- Following from the previous release decision of pausing Native on-premises Development the code has been move to a dedicated repository “native-on-prem”
- Kubernetes based solution will now support both Network and on-premises Edge
- Service Mesh support
- Basic support for Service Mesh using Istio Service Mesh within an OpenNESS cluster.
NOTE: When deploying Istio Service Mesh in VMs, a minimum of 8 CPU core and 16GB RAM must be allocated to each worker VM so that Istio operates smoothly
- Application of Service Mesh smartedge-open 5G and Media analytics - A dedicated network for service to service communications
- Basic support for Service Mesh using Istio Service Mesh within an OpenNESS cluster.
- EAA Update
- EAA microservices has been updated to be more cloud-native friendly
- Edge Insights Application (update)
- Industrial Edge Insights Software update to version 2.3.
- Experience Kit now supports multiple detection video's – Safety equipment detection, PCB default detection and also supports external video streams.
- Early access release of Edge Multi-Cluster Orchestration(EMCO), a Geo-distributed application orchestrator for Kubernetes. This release supports EMCO deploying and managing the life cycle of the Smart City Application pipeline on the edge cluster. More details in the EMCO Release Notes.
- Reference implementation of the offline installation package for the Converged Edge Reference Architecture (OpenNESS Experience Kit) Access Edge flavor enabling installation of Kubernetes and related enhancements for Access edge deployments.
- Azure Development kit (Devkit) supporting the installation of an OpenNESS Kubernetes cluster on the Microsoft* Azure* cloud. This is typically used by a customer who wants to develop applications and services for the edge using OpenNESS building blocks.
- Support Intel® vRAN Dedicated Accelerator ACC100, Kubernetes Cloud-native deployment supporting higher capacity 4G/LTE and 5G vRANs cells/carriers for FEC offload.
- Major system Upgrades: Kubernetes 1.19.3, CentOS 7.8, Calico 3.16, and Kube-OVN 1.5.2.
- Edge Insights for Industrial updated to 2.4
- Support for Intel® Ethernet Controller E810
- Improvements to Converged Edge Reference Architecture framework including support for deploying one or more OpenNESS Kubernetes clusters
- OpenVINO upgraded to 2021.1.110
- Major system upgrades: CentOS 7.9, Kubernetes 1.20.0, Docker 20.10.2, QEMU 5.2 and Golang 1.16.
- Kubernetes CNI upgrades: Calico 3.17, SR-IOV CNI 2.6, Flannel 0.13.0.
- Telemetry upgrades: CAdvisor 0.37.5, Grafana 7.4.2, Prometheus 2.24.0, Prometheus Node Exporter 1.0.1.
- Set Calico as a default cni for cdn-transcode, central_orchestrator, core-cplane, core-uplane, media-analytics and minimal flavor.
- Intel CMK is replaced with Kubernetes native CPU manager for core resource allocation
- Resolved an intermittent issue with Kubernetes repository signature verification
- Kubernetes deployment support for Intel® QuickAssist Technology (Intel® QAT)
- Provided an option to use CGroupFS as a CGroup driver for Kubernetes and Docker
- EII upgraded from 2.4 to 2.4.1
- VCA package upgraded from R5 to R5.1
- Changed the CNI used for Azure cloud deployment from default Calico to Azure-supported kube-ovn.
- No new features updates
- Intel Edge Software Hub Reference implementation updated and revalidated to working with Intel® Smart Edge Open - 21.09.
There are no unsupported or discontinued features relevant to this release.
There are no unsupported or discontinued features relevant to this release.
There are no unsupported or discontinued features relevant to this release.
- NTS Dataplane support for Network edge is discontinued.
- Controller UI for Network edge has been discontinued except for the CNCA configuration. Customers can optionally leverage the Kubernetes dashboard to onboard applications.
- Edge node only supports non-realtime kernel.
- Support for HDDL-R only restricted to non-real-time or non-customized CentOS 7.6 default kernel.
- Offline install for Native mode OnPremises has be deprecated
- Native on-premises is now located in a dedicated repository with no further feature updates from previous release.
There are no unsupported or discontinued features relevant to this release.
- FlexRAN/Access Edge OpenNESS Experience Kit Flavor is only available in Intel Distribution of OpenNESS
- OpenNESS repositories have been consolidated to the following
- No new features updates
There are no non-Intel issues relevant to this release.
There are no non-Intel issues relevant to this release.
- VHOST HugePages dependency
- Bug in getting appId by IP address for the container
- Wrong value of appliance verification key printed by ansible script
- NTS is hanging when trying to add same traffic policy to multiple interfaces
- Application in VM cannot be started
- Bug in libvirt deployment
- Invalid status after app un-deployment
- Application memory field is in MB
- Improved usability/automation in Ansible scripts
- Realtime Kernel support for network edge with K8s.
- Modular playbooks
- Optimized the Kubernetes based deployment by supporting multiple Flavors
- Further optimized the Kubernetes based deployment by supporting multiple Flavors
- Network edge installation time is optimized using pre-built Docker images
- cAdvisor occasional failure issue is resolved
- "Traffic rule creation: cannot parse filled and cleared fields" in Legacy OnPremises is fixed
- Issue fixed when removing Edge Node from Controller when its offline and traffic policy is configured or app deployed
- Known issue with Pod that uses hugepage get stuck in terminating state on deletion hs been fixed after upgrading to Kubernetes 1.19.3
- Upgraded to Kube-OVN v1.5.2 for further Kube-OVN CNI enhancements
- Fixed EdgeDNS service building failure resulting in deployment failure
- Fixed FlexRAN and FlexRAN Telemetry flavors to eliminate deployment failures
- Fixed Calico and SR-IOV deployment failures
- Fixed TAS deployment
- Updated SR-IOV CNI and device plugin to fix issues with image build in offline package creator
- Offline deployment issues related to zlib-devel version 1.2.7-19
- CAdvisor resource utilization has been optimized using "--docker_only=true" which decreased CPU usage from 15-25% to 5-6% (confirmed with ‘docker stats’ and ‘top’ commands). Memory usage also decreased by around 15-20%.
- Resolved an intermittent issue with Kubernetes repository signature verification
- Fixed DPDK compilation issue when primary CNI is Calico with eBPF and kernel is 5.X.
- Resolved intermittent Pip installation issues.
- Corrected Azure single node deployment issues due to incorrect configuration in the setup script.
- Resolved offline installation of Ansible prerequisites.
- VCA card is now correctly brought up after node reboot.
- Fixed multi-node deployment for Azure cloud by changing CNI to kube-ovn as Calico is not supported by Azure.
- Intel Edge Software Hub Reference implementation updated and revalidated to working with Intel® Smart Edge Open - 21.09.
There are no issues relevant to this release.
There is one issue relevant to this release: it is not possible to remove the application from Edge Node in case of error during application deployment. The issue concerns applications in a Virtual Machine.
- Gateway in multi-node - will not work when few nodes will have the same public IP (they will be behind one common NAT)
- Ansible in K8s can cause problems when rerun on a machine:
- If after running all 3 scripts
- Script 02 will be run again (it will not remove all necessary K8s related artifacts)
- We would recommend cleaning up the installation on the node
- Gateway in multi-node - will not work when few nodes will have the same public IP (they will be behind one common NAT)
- OpenNESS On-Premises: Cannot remove a failed/disconnected the edge node information/state from the controller
- The CNCA API (4G & 5G) supported in this release is an early access reference implementation and does not support authentication
- Real-time kernel support has been temporarily disabled to address the Kubernetes 1.16.2 and Realtime kernel instability.
- On-Premises edge installation takes more than 1.5 hours because of the Docker image build for OVS-DPDK
- Network edge installation takes more than 1.5 hours because of the Docker image build for OVS-DPDK
- OpenNESS controller allows management NICs to be in the pool of configuration, which might allow configuration by mistake. Thus, disconnecting the node from control plane
- When using the SRIOV EPA feature added in 20.03 with OVNCNI, the container cannot access the CNI port. This is due to the SRIOV port being set by changing the network used by the container from default to a custom network. This overwrites the OVNCNI network setting configured before this to enable the container to work with OVNCNI. Another issue with the SRIOV, is that this also overwrites the network configuration with the EAA and edgedns, agents, which prevents the SRIOV enabled container from communicating with the agents.
- Cannot remove Edge Node from Controller when its offline and traffic policy is configured or the app is deployed.
- On-Premises edge installation takes 1.5hrs because of the Docker image build for OVS-DPDK
- Network edge installation takes 1.5hrs because of docker image build for OVS-DPDK
- OpenNESS controller allows management NICs to be in the pool of configuration, which might allow configuration by mistake and thereby disconnect the node from control plane
- When using the SRIOV EPA feature added in 20.03 with OVNCNI, the container cannot access the CNI port. This is due to the SRIOV port being set by changing the network used by the container from default to a custom network, This overwrites the OVNCNI network setting configured prior to this to enable the container to work with OVNCNI. Another issue with the SRIOV, is that this also overwrites the network configuration with the EAA and edgedns, agents, which prevents the SRIOV enabled container from communicating with the agents.
- Cannot remove Edge Node from Controller when its offline and traffic policy is configured or app is deployed.
- Legacy OnPremises - Traffic rule creation: cannot parse filled and cleared fields
- There is an issue with using CDI when uploading VM images when CMK is enabled due to missing CMK taint toleration. The CDI upload pod does not get deployed and the
virtctl
plugin command times out waiting for the action to complete. A workaround for the issue is to invoke the CDI upload command, edit the taint toleration for the CDI upload to tolerate CMK, update the pod, create the PV, and let the pod run to completion. - There is a known issue with cAdvisor which in certain scenarios occasionally fails to expose the metrics for the Prometheus endpoint. See the following GitHub* link: google/cadvisor#2537
- Pod which uses hugepage get stuck in terminating state on deletion. This is a known issue on Kubernetes 1.18.x and is planned to be fixed in 1.19.x
- Calico cannot be used as secondary CNI with Multus in OpenNESS. It will work only as primary CNI. Calico must be the only network provider in each cluster. We do not currently support migrating a cluster with another network provider to use Calico networking. https://docs.projectcalico.org/getting-started/kubernetes/requirements
- Collectd Cache telemetry using RDT does not work when RMD is enabled because of resource conflict. Workaround is to disable collectd RDT plugin when using RMD - this by default is implemented globally. With this workaround customers will be able to allocate the Cache but not use Cache related telemetry. In case where RMD is not being enabled customers who desire RDT telemetry can re-enable collectd RDT.
- cAdvisor CPU utilization of Edge Node is high and could cause a delay to get an interactive SSH session. A work around is to remove CAdvisor if not needed using
helm uninstall cadvisor -n telemetry
- An issue appears when the KubeVirt Containerized Data Importer (CDI) upload pod is deployed with Kube-OVN CNI, the deployed pods readiness probe fails and pod is never in ready state. It is advised that the user uses other CNI such as Calico CNI when using CDI with OpenNESS
- Telemetry deployment with PCM enabled will cause a deployment failure in single node cluster deployments due to PCM dashboards for Grafana not being found
- Offline deployment issues related to zlib-devel version 1.2.7-19
- An issue appears when the KubeVirt Containerized Data Importer (CDI) upload pod is deployed with Kube-OVN CNI, the deployed pods readiness probe fails and pod is never in ready state. Calico CNI is used by default in OpenNESS when using CDI
- Telemetry deployment with PCM enabled will cause a deployment failure in single node cluster deployments due to conflict with CollectD deployment, it is advised to not use PCM and CollectD at the same time in OpenNESS at this time
- Kafka and Zookeeper resource consumption is on the higher side. When deployed in the context of uCPE and SD-WAN users need to consider this.
- When flannel CNI is being used and worker node is being manually joined or re-joined to the cluster, then
kubectl patch node NODE_NAME -p '{ "spec":{ "podCIDR":"10.244.0.0/24" }}
command should be issued on controller to enable flannel CNI on that node.
- Multi-node deployment through Azure cloud fails due to the default (Calico) CNI not being supported by Azure.
OpenNESS Edge node, OpenNESS Controller, Common, Spec, and OpenNESS Applications.
OpenNESS Edge node, OpenNESS Controller, Common, Spec, and OpenNESS Applications.
OpenNESS Edge node, OpenNESS Controller, Common, Spec, and OpenNESS Applications.
OpenNESS Edge node, OpenNESS Controller, Common, Spec, and OpenNESS Applications, and Experience kit.
OpenNESS Edge node, OpenNESS Controller, Common, Spec, and OpenNESS Applications, and Experience kit.
- Open Source: Edge node, Controller, Epcforedge, Common, Spec, Applications, and Experience kit.
- IDO: IDO Edge node, IDO Controller, IDO Epcforedge, IDO Spec, and IDO Experience kit.
- Open Source: Edge node, Controller, Epcforedge, Common, Spec, Applications and Experience kit.
- IDO: IDO Edge node, IDO Controller, IDO Epcforedge, IDO Spec and IDO Experience kit.
- Open Source: Edge node, Controller, Epcforedge, Common, Spec, Applications and Experience kit.
- IDO: IDO Edge node, IDO Controller, IDO Epcforedge, IDO Spec and IDO Experience kit.
NOTE: Edge applications repo is common to Open Source and IDO
- Open Source: Edge node, Controller, Epcforedge, Common, Spec, Applications and Experience kit.
- IDO: IDO Edge node, IDO Controller, IDO Epcforedge, IDO Spec and IDO Experience kit.
- https://github.com/smart-edge-open/converged-edge-experience-kits
- https://github.com/smart-edge-open/specs
- https://github.com/smart-edge-open/edgeapps
- https://github.com/smart-edge-open/edgeservices
- https://github.com/smart-edge-open/openshift-operator
Intel® Smart Edge Open edge node, Intel® Smart Edge Open controller, Common, Spec, and Intel® Smart Edge Open applications, and experience kit. Edge node has been tested using the following hardware specification:
- Supermicro* 3U form factor chassis server, product SKU code: 835TQ-R920B
- Motherboard type: X11SDV-16C-TP8F
- Intel® Xeon® Processor D-2183IT
CLX-SP | Compute Node based on CLX-SP(6252N) |
Board | S2600WFT server board |
2 x Intel® Xeon® Gold 6252N CPU @ 2.30GHz | |
2 x associated Heatsink | |
Memory | 12x Micron 16GB DDR4 2400MHz DIMMS* [2666 for PnP] |
Chassis | 2U Rackmount Server Enclosure |
Storage | Intel M.2 SSDSCKJW360H6 360G |
NIC | 1x Intel® Fortville NIC X710DA4 SFP+ ( PCIe card to CPU-0) |
QAT | Intel® Quick Assist Adapter Device 37c8 |
(Symmetrical design) LBG integrated | |
NIC on board | Intel-Ethernet-Controller-I210 (for management) |
Other card | 2x PCIe Riser cards |
SKX-SP | Compute Node based on SKX-SP(6148) |
Board | WolfPass S2600WFQ server board(symmetrical QAT)CPU |
2 x Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz | |
2 x associated Heatsink | |
Memory | 12x Micron 16GB DDR4 2400MHz DIMMS* [2666 for PnP] |
Chassis | 2U Rackmount Server Enclosure |
Storage | Intel® M.2 SSDSCKJW360H6 360G |
NIC | 1x Intel® Fortville NIC X710DA4 SFP+ ( PCIe card to CPU-0) |
QAT | Intel® Quick Assist Adapter Device 37c8 |
(Symmetrical design) LBG integrated | |
NIC on board | Intel-Ethernet-Controller-I210 (for management) |
Other card | 2x PCIe Riser cards |
HDDL-R | Mouser Mustang-V100 |
VCAC-A | VCAC-A Accelerator for Media Analytics |
PAC-N3000 | Intel® FPGA Programmable Acceleration Card (Intel® FPGA PAC) N3000 |
ACC100 | Intel® vRAN Dedicated Accelerator ACC100 |
Intel® Smart Edge Open 21.09 was tested on Ubuntu 20.04.2 LTS (Focal Fossa). OpenNESS was tested on CentOS Linux release 7.9.2009 (Core).
NOTE: OpenNESS is tested with CentOS 7.9 PREEMPT_RT kernel to ensure VNFs and Applications can co-exist. There is no requirement to run on a PREEMPT_RT kernel.
Package: telemetry, cadvisor 0.37.5, grafana 7.4.2, prometheus 2.24.0, prometheus: node exporter 1.0.1, golang 1.16, docker 20.10.2, kubernetes 1.20.0, dpdk 19.11.1, ovs 2.14.0, ovn 2.14.0, helm 3.1.2, kubeovn 1.5.2, flannel 0.13.0, calico 3.17.0, multus 3.6, sriov cni 2.6, nfd 0.6.0, cmk v1.4.1, TAS (from specific commit "a13708825e854da919c6fdf05d50753113d04831"), openssl 1.1.1i, QEMU 5.2