Key Features

Accelerate Video, AI, and Graphics Workloads

The NVIDIA Ada Lovelace L4 Tensor Core GPU delivers universal acceleration and energy efficiency for video, AI, virtualized desktop, and graphics applications in the enterprise, in the cloud, and at the edge. With NVIDIA’s AI platform and full-stack approach, L4 is optimized for inference at scale for a broad range of AI applications, including recommendations, voice-based AI avatar assistants, generative AI, visual search, and contact center automation to deliver the best personalized experiences. As the most efficient NVIDIA accelerator for mainstream use, servers equipped with L4 power up to 120X higher AI video performance and 2.7X more generative AI performance over CPU solutions, as well as over 4X more graphics performance than the previous GPU generation. NVIDIA L4’s versatility and energy-efficient, single-slot, low-profile form factor make it ideal for global deployments, including edge locations.

Accelerate Workloads Efficiently and Sustainably

The NVIDIA L4 is an integral part of the NVIDIA data center platform. Built for AI, video, virtual workstations, graphics, simulation, data science, and data analytics, the platform accelerates over 3,000 applications and is available everywhere at scale, from data center to edge to cloud, delivering both dramatic performance gains and energy-efficiency opportunities. As AI and video become more pervasive, the demand for efficient, cost-effective computing is increasing more than ever. NVIDIA L4 Tensor Core GPUs deliver up to 120X better AI video performance, resulting in up to 99 percent better energy efficiency and lower total cost of ownership compared to traditional CPU-based infrastructure. This lets enterprises reduce rack space and significantly lower their carbon footprint, while being able to scale their data centers to many more users. The energy saved by switching from CPUs to NVIDIA L4s in a 2 megawatt (MW) data center can power over 2,000 homes for one year or match the carbon offset of 172,000 trees grown over 10 years.

Categories

Powered by NVIDIA Ada Lovelace Architecture

NVIDIA CUDA Cores

Double-speed processing for single-precision floating point (FP32) operations provides significant performance improvements for graphics and simulation workflows, such as complex 3D computer-aided design (CAD) and computer-aided engineering (CAE), on the desktop.

Third-Generation RT Cores

With up to 2X the throughput over the previous generation, third-generation RT Cores deliver massive speedups for workloads like photorealistic rendering of movie content, architectural design evaluations, and virtual prototyping of product designs. This technology also accelerates the rendering of ray-traced motion blur with greater visual accuracy.

Fourth-Generation Tensor Cores

Fourth-generation Tensor Cores provide faster AI compute performance, delivering more than 2X the performance of the previous generation. These new Tensor Cores support acceleration of the FP8 precision data type and provide independent floating-point and integer data paths to speed up execution of mixed floating point and integer calculations.

48 Gigabytes (GB) of GPU Memory

With 48GB GDDR6 memory, the RTX 6000 gives data scientists, engineers, and creative professionals the large memory necessary to work with massive datasets and workloads like rendering, data science, and simulation.

Virtualization-Ready

Support for NVIDIA RTX Virtual Workstation (vWS) software allows a personal workstation to be repurposed into multiple high-performance virtual workstation instances, enabling remote users to share resources to drive high-end design, AI, and compute workloads.

PCIe Gen 4

PCI Express Gen 4 support provides double the bandwidth of PCIe Gen 3, improving data-transfer speeds from CPU memory for data-intensive tasks like AI and data science.

FP32 30.3 teraFLOPs
TF32 Tensor Core 120 teraFLOPS*
FP16 Tensor Core 242 teraFLOPS*
BFLOAT16 Tensor Core 242 teraFLOPS*
FP8 Tensor Core 485 teraFLOPs*
INT8 Tensor Core 485 TOPs*
GPU memory 24GB
GPU memory bandwidth 300 GB/s
NVENC | NVDEC | JPEG decoders 2 | 4 | 4
Max thermal design power (TDP) 72W
Form factor 1-slot low-profile, PCIe
Interconnect PCIe Gen4 x16 64GB/s
Server options Partner and NVIDIA-Certified Systems with 1‒8 GPUs
* Shown with sparsity. Specifications 1/2 lower without sparsity.

Speak with an expert to learn more.