F5 to supercharge AI application delivery for service providers and enterprises with NVIDIA BlueField-3 DPUs

By Kunal Anand, CTO, F5


F5 announced the availability of BIG-IP Next for Kubernetes, an innovative AI application delivery and security solution that equips service providers and large enterprises with a centralised control point to accelerate
, secure, and streamline data traffic that flows into and out of large-scale AI infrastructures.

The solution harnesses the power of high-performance NVIDIA BlueField-3 DPUs to enhance the efficiency of data center traffic which is critical to large-scale AI deployments. With an integrated view of networking, traffic management, and security, customers will be able to maximise data center resource utilisation while achieving optimal AI application performance. This not only improves infrastructure efficiency but also enables faster, more responsive AI inference, ultimately delivering an enhanced AI-driven customer experience.

F5 BIG-IP Next for Kubernetes is a purpose-built solution for Kubernetes environments that has been proven in large-scale telco cloud and 5G infrastructures. With BIG-IP Next for Kubernetes, this technology is now tailored for leading AI use cases such as inference, retrieval-augmented generation (RAG), and seamless data management and storage. The integration with NVIDIA BlueField-3 DPUs minimises hardware footprint, enables granular multi-tenancy, and optimises energy consumption while delivering high-performance networking, security, and traffic management. The combination of F5 and NVIDIA technologies allows both mobile and fixed-line telco service providers to ease the transition to cloud-native (Kubernetes) infrastructure, addressing the growing demand for vendors to adapt their functions to a cloud-native network functions (CNFs) model. F5 BIG-IP Next for Kubernetes offloads data-heavy tasks to the BlueField-3 DPUs, freeing up CPU resources for revenue-generating applications. The solution is particularly beneficial at the network edge for virtualised RAN (vRAN) or DAA for MSO, and in the core network for 5G, enabling the future potential for 6G. Designed specifically for high-demand service providers and large-scale infrastructures, F5 BIG-IP Next for Kubernetes:

Streamlines delivery of AI services at cloud scale: BIG-IP Next for Kubernetes seamlessly integrates with customers and front-end networks, significantly reducing latency while delivering high-performance load balancing to handle the immense data demands of multi-billion-parameter AI models and trillions of operations.

Enhances control of AI deployments: The solution offers a centralised integration point into modern AI networks with rich observability and fine-grained information. BIG-IP Next for  Kubernetes supports multiple L7 protocols beyond HTTP, ensuring enhanced ingress and egress control at very high performance.

Protects the new AI landscape: Customers can fully automate the discovery and security of AI training and inference endpoints. BIG-IP Next for Kubernetes also isolates AI applications from targeted threats, bolstering data integrity and sovereignty while addressing the encryption capabilities critical for modern AI Environments. Availability for BIG-IP Next for Kubernetes running on NVIDIA BlueField-3 DPUs will begin in November.

“The proliferation of AI has catalysed an unprecedented demand for advanced semiconductors and technologies. Organisations are building out AI factories, highly optimised environments designed to train large AI models and deliver the requisite processing power for inference scale at an astounding rate, and with minimal latency. The synergy between F5’s robust application delivery and security services and NVIDIA’s full-stack accelerated computing creates a powerful ecosystem. This integration provides customers with enhanced observability, granular control, and optimised performance for their AI workloads across the entire stack, from the hardware acceleration layer to the application interface.” – Kunal Anand, Chief Technology and AI Officer, F5.

“Service providers and enterprises require accelerated computing to deliver high-performance AI applications securely and efficiently at cloud scale. NVIDIA is working with F5 to accelerate AI application delivery, better ensuring peak efficiency and seamless user experiences powered by Bluefield-3 DPUs.” – Ash Bhalgat, Sr. Director of AI Networking and Security Partnerships, NVIDIA

“Realising the potential of AI requires more data processing capabilities than the industry had previously prepared for. For many companies, deploying cutting-edge AI requires massive infrastructure buildouts that tend to be very complex and expensive, making efficient and secure operations more important than ever. F5 BIG-IP Next for Kubernetes addresses performance and security concerns for large-scale AI infrastructure. By delivering optimised traffic management, organisations gain greater data ingestion performance and server utilisation during AI model inferencing. This leads to a vastly improved customer experience for AI application users.” – Kuba Stolarski, Research Vice President, Computing Systems Research Practice, IDC

“The explosion of AI workloads has created a new wave of massive demand for scalable, optimised, and enhanced Kubernetes ingress and egress control. With F5 now delivering the established benefits of BIG-IP Next for Kubernetes directly on NVIDIA BlueField-3 DPUs, this unleashes an already proven

technology that is now deployable at an ideal insertion point for large-scale AI deployments. WWT clients will benefit from greater data ingestion performance and GPU utilisation during model training and better user experiences during inference while gaining a strategic control point for security services. Technology from F5 and NVIDIA—two of our most strategic partnerships—further strengthens our Global Cyber mission to deliver digital security excellence.” – Todd Hathaway, Global Practice Manager, AI, App,  and API Security Solutions at WWT.

AICloudsecurity
Comments (0)
Add Comment