GEFORCE

NVIDIA L40S

NVIDIA graphics card specifications and benchmark scores

48 GB
VRAM
2520
MHz Boost
300W
TDP
384
Bus Width
Ray Tracing 🤖Tensor Cores

NVIDIA L40S Specifications

⚙️

L40S GPU Core

Shader units and compute resources

The NVIDIA L40S GPU core specifications define its raw processing power for graphics and compute workloads. Shading units (also called CUDA cores, stream processors, or execution units depending on manufacturer) handle the parallel calculations required for rendering. TMUs (Texture Mapping Units) process texture data, while ROPs (Render Output Units) handle final pixel output. Higher shader counts generally translate to better GPU benchmark performance, especially in demanding games and 3D applications.

Shading Units
18,176
Shaders
18,176
TMUs
568
ROPs
192
SM Count
142
⏱️

L40S Clock Speeds

GPU and memory frequencies

Clock speeds directly impact the L40S's performance in GPU benchmarks and real-world gaming. The base clock represents the minimum guaranteed frequency, while the boost clock indicates peak performance under optimal thermal conditions. Memory clock speed affects texture loading and frame buffer operations. The L40S by NVIDIA dynamically adjusts frequencies based on workload, temperature, and power limits to maximize performance while maintaining stability.

Base Clock
1110 MHz
Base Clock
1,110 MHz
Boost Clock
2520 MHz
Boost Clock
2,520 MHz
Memory Clock
2250 MHz 18 Gbps effective
GDDR GDDR 6X 6X

NVIDIA's L40S Memory

VRAM capacity and bandwidth

VRAM (Video RAM) is dedicated memory for storing textures, frame buffers, and shader data. The L40S's memory capacity determines how well it handles high-resolution textures and multiple displays. Memory bandwidth, measured in GB/s, affects how quickly data moves between the GPU and VRAM. Higher bandwidth improves performance in memory-intensive scenarios like 4K gaming. The memory bus width and type (GDDR6, GDDR6X, HBM) significantly influence overall GPU benchmark scores.

Memory Size
48 GB
VRAM
49,152 MB
Memory Type
GDDR6
VRAM Type
GDDR6
Memory Bus
384 bit
Bus Width
384-bit
Bandwidth
864.0 GB/s
💾

L40S by NVIDIA Cache

On-chip cache hierarchy

On-chip cache provides ultra-fast data access for the L40S, reducing the need to fetch data from slower VRAM. L1 and L2 caches store frequently accessed data close to the compute units. AMD's Infinity Cache (L3) dramatically increases effective bandwidth, improving GPU benchmark performance without requiring wider memory buses. Larger cache sizes help maintain high frame rates in memory-bound scenarios and reduce power consumption by minimizing VRAM accesses.

L1 Cache
128 KB (per SM)
L2 Cache
48 MB
📈

L40S Theoretical Performance

Compute and fill rates

Theoretical performance metrics provide a baseline for comparing the NVIDIA L40S against other graphics cards. FP32 (single-precision) performance, measured in TFLOPS, indicates compute capability for gaming and general GPU workloads. FP64 (double-precision) matters for scientific computing. Pixel and texture fill rates determine how quickly the GPU can render complex scenes. While real-world GPU benchmark results depend on many factors, these specifications help predict relative performance levels.

FP32 (Float)
91.61 TFLOPS
FP64 (Double)
1,431.4 GFLOPS (1:64)
FP16 (Half)
91.61 TFLOPS (1:1)
Pixel Rate
483.8 GPixel/s
Texture Rate
1,431.4 GTexel/s

L40S Ray Tracing & AI

Hardware acceleration features

The NVIDIA L40S includes dedicated hardware for ray tracing and AI acceleration. RT cores handle real-time ray tracing calculations for realistic lighting, reflections, and shadows in supported games. Tensor cores (NVIDIA) or XMX cores (Intel) accelerate AI workloads including DLSS, FSR, and XeSS upscaling technologies. These features enable higher visual quality without proportional performance costs, making the L40S capable of delivering both stunning graphics and smooth frame rates in modern titles.

RT Cores
142
Tensor Cores
568
🏗️

Ada Lovelace Architecture & Process

Manufacturing and design details

The NVIDIA L40S is built on NVIDIA's Ada Lovelace architecture, which defines how the GPU processes graphics and compute workloads. The manufacturing process node affects power efficiency, thermal characteristics, and maximum clock speeds. Smaller process nodes pack more transistors into the same die area, enabling higher performance per watt. Understanding the architecture helps predict how the L40S will perform in GPU benchmarks compared to previous generations.

Architecture
Ada Lovelace
GPU Name
AD102
Process Node
5 nm
Foundry
TSMC
Transistors
76,300 million
Die Size
609 mm²
Density
125.3M / mm²
🔌

NVIDIA's L40S Power & Thermal

TDP and power requirements

Power specifications for the NVIDIA L40S determine PSU requirements and thermal management needs. TDP (Thermal Design Power) indicates the heat output under typical loads, guiding cooler selection. Power connector requirements ensure adequate power delivery for stable operation during demanding GPU benchmarks. The suggested PSU wattage accounts for the entire system, not just the graphics card. Efficient power delivery enables the L40S to maintain boost clocks without throttling.

TDP
300 W
TDP
300W
Power Connectors
1x 16-pin
Suggested PSU
700 W
📐

L40S by NVIDIA Physical & Connectivity

Dimensions and outputs

Physical dimensions of the NVIDIA L40S are critical for case compatibility. Card length, height, and slot width determine whether it fits in your chassis. The PCIe interface version affects bandwidth for communication with the CPU. Display outputs define monitor connectivity options, with modern cards supporting multiple high-resolution displays simultaneously. Verify these specifications against your case and motherboard before purchasing to ensure a proper fit.

Slot Width
Dual-slot
Length
267 mm 10.5 inches
Height
111 mm 4.4 inches
Bus Interface
PCIe 4.0 x16
Display Outputs
1x HDMI 2.13x DisplayPort 1.4a
Display Outputs
1x HDMI 2.13x DisplayPort 1.4a
🎮

NVIDIA API Support

Graphics and compute APIs

API support determines which games and applications can fully utilize the NVIDIA L40S. DirectX 12 Ultimate enables advanced features like ray tracing and variable rate shading. Vulkan provides cross-platform graphics capabilities with low-level hardware access. OpenGL remains important for professional applications and older games. CUDA (NVIDIA) and OpenCL enable GPU compute for video editing, 3D rendering, and scientific applications. Higher API versions unlock newer graphical features in GPU benchmarks and games.

DirectX
12 Ultimate (12_2)
DirectX
12 Ultimate (12_2)
OpenGL
4.6
OpenGL
4.6
Vulkan
1.4
Vulkan
1.4
OpenCL
3.0
CUDA
8.9
Shader Model
6.8
📦

L40S Product Information

Release and pricing details

The NVIDIA L40S is manufactured by NVIDIA as part of their graphics card lineup. Release date and launch pricing provide context for comparing GPU benchmark results with competing products from the same era. Understanding the product lifecycle helps evaluate whether the L40S by NVIDIA represents good value at current market prices. Predecessor and successor information aids in tracking generational improvements and planning future upgrades.

Manufacturer
NVIDIA
Release Date
Oct 2022
Production
End-of-life
Predecessor
Server Ampere
Successor
Server Hopper

L40S Benchmark Scores

geekbench_openclSource

Geekbench OpenCL tests GPU compute performance using the cross-platform OpenCL API. This shows how NVIDIA L40S handles parallel computing tasks like video encoding and scientific simulations.

geekbench_opencl #3 of 582
334,437
88%
Max: 380,114
Compare with other GPUs

geekbench_vulkanSource

Geekbench Vulkan tests GPU compute using the modern low-overhead Vulkan API. This shows how NVIDIA L40S performs with next-generation graphics and compute workloads. Vulkan offers better CPU efficiency than older APIs like OpenGL.

geekbench_vulkan #6 of 386
250,769
66%
Max: 379,571

About NVIDIA L40S

The Ada Lovelace‑based NVIDIA L40S pushes the limits of workstation graphics with its massive 48 GB of GDDR6 memory. Built on a 5 nm process, the GPU delivers a base clock of 1.11 GHz and can boost up to 2.52 GHz, offering a 300 W thermal design power headroom. Benchmarks show a Geekbench OpenCL score of 334,437 points and a Vulkan score of 250,769, placing it firmly in the high‑end tier. Its PCIe 4.0 x16 interface ensures ample bandwidth for data‑intensive workloads such as AI inference and complex rendering pipelines. The card’s 48 GB capacity translates into a memory bandwidth that comfortably exceeds 1 TB/s, reducing bottlenecks in large texture streaming. With these specs, the L40S from NVIDIA stands out as a compelling choice for creators who need raw compute without compromise.

When it comes to gaming, the NVIDIA’s L40S can sustain well over 120 FPS at 4K Ultra settings in titles that are GPU‑bound. Its advanced ray‑tracing cores and tensor accelerators enable real‑time global illumination and DLSS 3 upscaling with minimal latency. The enormous VRAM pool not only fuels high‑resolution textures but also provides the headroom required for VR and multi‑monitor setups. Thermal performance is managed by a dual‑fan vapor‑chamber design that keeps hotspot temperatures under 85 °C even under sustained loads. This efficient cooling allows the card to maintain boost clocks for extended periods, which is essential for competitive e‑sports scenarios. The following list highlights the best use‑cases for extracting maximum performance from the L40S:

  1. High‑resolution 4K gaming with ray tracing enabled.
  2. Professional workloads such as 3D rendering, simulation, and AI inference.
  3. VR and multi‑monitor environments demanding large texture footprints.

The AMD Equivalent of L40S

Looking for a similar graphics card from AMD? The AMD Radeon RX 7900 XTX offers comparable performance and features in the AMD lineup.

AMD Radeon RX 7900 XTX

AMD • 24 GB VRAM

View Specs Compare

Popular NVIDIA L40S Comparisons

See how the L40S stacks up against similar graphics cards from the same generation and competing brands.

Compare L40S with Other GPUs

Select another GPU to compare specifications and benchmarks side-by-side.

Browse GPUs