Is the GeForce RTX 5080 Good for Running LLMs?

Find out if the GeForce RTX 5080 GPU is suitable for running local AI models, including large language models (LLMs) via LMStudio.

Is the GeForce RTX 5080 Good for Running LLMs? Brandon Jones / TechReviewer

Last Updated: September 23, 2025

Written by Brandon Jones

Is the GeForce RTX 5080 Good for Running LLMs?

Section Link

To run large language models (LLMs) well, a GPU needs enough VRAM, high memory bandwidth, strong compute units, and a mature software stack (for example NVIDIA CUDA or AMD ROCm).

TL;DR: GeForce RTX 5080 for LLMs

How it performs with local LLM workloads:

Verdict: With 16 GB of GDDR7 VRAM, the GeForce RTX 5080 is a solid choice for most LLM tasks.

Best for: 7B unquantized (FP16/BF16) or 8‑bit quantization; 13B–14B with 8‑bit quantization; parameter‑efficient fine‑tuning; local development.

Caveat: 13B–14B typically need 8‑bit; 30B usually requires 4‑bit and offloading on 12–16 GB of VRAM.

GeForce RTX 5080 AI Specifications
Spec Details
VRAM Capacity 16 GB
Memory Type GDDR7
Memory Bandwidth 960 GB/s
Memory Bus 256-bit
Max Recommended LLM Size 24B (quantized)
Qwen 3 8B Q4 (Min. Est. Tokens/sec) 140–173 TPS
GPT‑OSS 20B MXFP4 (Min. Est. Tokens/sec) 70–83 TPS
Qwen 3 32B Q4 (Min. Est. Tokens/sec) 16–19 TPS (partially offloaded)
Max Power Draw 360 W
PCIe Generation PCIe 5.0
PCIe Lanes x16
Architecture Blackwell

With 16 GB of VRAM and Blackwell architecture, the GeForce RTX 5080 runs models up to 24B with 4- to 6-bit quantization and occasional offload; plan around context length limits.

Upgrade your setup to handle larger and better LLMs: RTX 5090 on Amazon

Performance of the GeForce RTX 5080

Section Link

Minimum Estimated Throughput for the GeForce RTX 5080:

GeForce RTX 5080 Throughput Estimates
Metric Min. Est. Tokens/sec
Qwen 3 8B Q4 (Min. Est. Tokens/sec) 140–173 TPS
GPT‑OSS 20B MXFP4 (Min. Est. Tokens/sec) 70–83 TPS
Qwen 3 32B Q4 (Min. Est. Tokens/sec) 16–19 TPS(partially offloaded)

Performance Interpretation Guide: In chat-based applications, a throughput of 5–10 tokens per second is generally perceived as responsive, aligning closely with average human reading speeds of around 5 words (or tokens) per second in English, which supports fluid and natural-feeling interactions. That said, real-world performance can fluctuate significantly due to variables like model architecture, quantization format (such as 8-bit integer, FP16, or BF16), context window size, hardware specifications (e.g., GPU type and VRAM), driver versions, runtime environment, thermal throttling, and custom application parameters.

For example, on supported GPUs, 8-bit quantization often yields 1.5–2× greater throughput relative to unquantized FP16 or BF16 models, primarily by halving memory usage and leveraging optimized integer compute, though this depends on the workload and hardware.

Treat these minimum expected TPS figures above as indicative benchmarks (a planning floor) rather than fixed guarantees, as they represent outcomes under ideal conditions. To maximize efficiency, benchmark your setup in production-like scenarios and iterate on configurations to match your specific use case and system setup.

Cross-Vendor Reference TPS Estimates

The cross-vendor baseline figures below reflect minimum expected TPS outcomes, for use as a planning floor.

Real-World TPS Baselines
Model Variant NVIDIA (RTX 4090 · 24 GB · CUDA) AMD (RX 7900 XTX · 24 GB · ROCm) Intel (Arc A770 · 16 GB · oneAPI/IPEX)
Qwen3 7B Q4 120–150 TPS 95–115 TPS 30–45 TPS
Llama 3.1 13B Q4 70–100 TPS 56–80 TPS 20–30 TPS
Qwen3 32B Q4 30–45 TPS 24–36 TPS <10 TPS (partial offload)

LLM Fit Guide: GeForce RTX 5080 VRAM & Model Sizes

Section Link

VRAM Requirements

VRAM usage has two components: the fixed model weights and the context-dependent KV cache. Model size and precision set the weights cost, and lower precision needs less memory.

As a rule of thumb, a 30B model at 4-bit (Q4) uses about 18 to 21 GB of VRAM for weights. You should plan for roughly 19 to 22 GB of base usage on the GPU before accounting for any prompt. Then add the KV cache growth to estimate the total.

KV Cache Impact

The KV cache works like the model’s short-term memory. It grows with the length of your prompt and responses, so memory use scales with the total token count. For 30B to 34B models, expect about 0.25 to 0.30 GB of KV memory per 1,000 tokens when using FP16 with GQA, or about 0.5 to 1.0 GB per 1,000 tokens when using FP32 or without GQA.

In practice, a 30B model with 4-bit weights often totals about 20 to 24 GB at 4K to 8K tokens, including both the fixed weights and the growing cache.

Minimum VRAM for Popular LLM Models (Estimated)
Example Models Unquantized (FP16/BF16) (Approx. 2 GB per Billion Params) 8-bit Quantized (Approx. 1 GB per Billion Params) 4–5 bit Quantized (Approx. 0.56–0.69 GB per Billion Params)
~3B Gemma 3 3B
Llama 3.2 3B
Qwen 2.5 3B
6.00-8.00 GB 3.00-4.00 GB 2.00-3.50 GB
~7B-8B Mistral 7B v0.3
Gemma 3 7B
Qwen 2.5 7B
16.00-18.00 GB 8.00-9.00 GB 4.00-5.00 GB
~12B-14B Gemma 3 12B
Mistral Nemo 12B
Qwen 3 14B
28.00-32.00 GB 14.00-16.00 GB 7.00-8.50 GB
~30B-35B Qwen3-30B
Qwen 2.5-32B
Code Llama 34B
68.00-80.00 GB 34.00-40.00 GB 18.00-21.00 GB

VRAM Guidance & Classes for the GeForce RTX 5080

The GeForce RTX 5080 comes with 16 GB of VRAM. Refer to the corresponding tier below to estimate compatible workloads and compare it to common GPU classes.

VRAM Guidance & Classes
VRAM Capacity Class Guidance
Up to 8 GB Entry Local chat with 7B at 4–6‑bit quantization; basic CV; low‑res diffusion.
12–16 GB Mainstream 7B unquantized (FP16/BF16) on higher end; 13B–14B with 8‑bit quantization; long‑context RAG with tuned settings; LoRA on smaller models.
24 GB Class (e.g., RTX 4090) Enthusiast 7–13B comfortably; 30B at 4‑bit; 8‑bit often needs 32 GB+; SDXL is comfortable; supports longer prompts.
32 GB Class (e.g., RTX 5090) Enthusiast ~30B with 6–8‑bit quantization; extended contexts; ample headroom for concurrency.
48 GB (Prosumer) (e.g., RTX 6000 Ada/L40S) Prosumer 70B at 4‑bit; extra capacity for longer contexts; enables higher‑batch inference.
HBM Data Center (e.g., H100/H200/Blackwell) Data Center Large‑scale training, fine‑tuning, and high‑throughput inference.

Precision and Quantization on the GeForce RTX 5080

For the GeForce RTX 5080, the list below summarizes precision choices for inference and how each affects VRAM use, speed, and quality. Precision is the primary setting you adjust to fit a model into memory and to reach a desired throughput. Lower precision reduces the memory used by weights and often increases throughput. Total VRAM is the sum of fixed weights plus the token-dependent KV cache.

  • BF16: Similar quality to FP16 with a wider numeric range. Delivers strong speed on GPUs with native BF16 support and is a safe default when quality matters.
  • FP16: Strong quality and broad compatibility for 7B to 13B models. Widely supported across runtimes and frameworks.
  • FP8: When supported by hardware such as Hopper or Blackwell, increases throughput and provides about 2× memory efficiency compared with FP16 or BF16, usually with a small quality tradeoff.
  • 8-bit: Near-FP16 quality on most LLM tasks while using roughly 50% of the VRAM of FP16/BF16 (comparable to FP8). A practical baseline for high-throughput inference on many setups.
  • 6- or 5-bit: Further VRAM savings with a slight quality drop. Useful to fit 13B to 30B models on mid-range GPUs.
  • 4-bit: Maximizes VRAM savings. Quality depends on the quantization method (for example, GPTQ, AWQ, GGUF Q4 variants). Enables 30B and larger models on a single GPU, though prompts can be more sensitive to quantization choices.

PCIe, Offloading, & Storage for the GeForce RTX 5080

Section Link

PCIe bandwidth, storage speed, and system memory greatly influence performance during initial loading, weight offloading, and KV cache paging.

PCIe Requirements for LLMs With the GeForce RTX 5080

In AI workloads such as LLM inference, models typically run from GPU VRAM once loaded. PCIe bandwidth primarily affects the initial load and any offloading when parts of the model, KV cache, or activations spill beyond VRAM. Newer PCIe generations and wider lane widths reduce stalls during these transfers. Tokens per second are VRAM bound when nothing offloads, so faster links mainly help with initial loads and overflow traffic.

The GeForce RTX 5080 supports PCIe 5.0 x16.

PCIe LLM rating for the GeForce RTX 5080: Excellent for LLM tasks. It enables fast model loads, efficient offload bursts, and strong multi-GPU communication. Tokens per second remain unchanged once data resides in VRAM, which suits large inference or fine-tuning.

Use the primary x16 PCIe slot for best results. Motherboard, CPU, and BIOS settings can limit the effective PCIe version and lane width, so verify the active link speed with a utility like GPU-Z and confirm slot capabilities in your board manual.

PCIe Bandwidth for the GeForce RTX 5080

The GeForce RTX 5080 supports PCIe 5.0 x16, delivering approximately 63 GB/s of per-direction theoretical bandwidth.

For the GeForce RTX 5080, PCIe 4.0 x8 is generally sufficient for VRAM-resident inference tasks, while setups involving heavy offloading benefit from x16 lanes or PCIe 5.0 support. The effective link speed will depend on your motherboard's PCIe slot and the CPU's lane configuration.

Nominal PCIe Bandwidth by Link Width (Approximate Per-Direction Theoretical Limits)
x1 Bandwidth x2 Bandwidth x4 Bandwidth x8 Bandwidth x16 Bandwidth
PCIe 1.0 250 MB/s 500 MB/s 750 MB/s 2 GB/s 4 GB/s
PCIe 2.0 500 MB/s 1000 MB/s 2 GB/s 4 GB/s 8 GB/s
PCIe 3.0 1 GB/s 2 GB/s 4 GB/s 8 GB/s 16 GB/s
PCIe 4.0 2 GB/s 4 GB/s 8 GB/s 16 GB/s 32 GB/s
PCIe 5.0 4 GB/s 8 GB/s 16 GB/s 32 GB/s 63 GB/s
PCIe 6.0 8 GB/s 15 GB/s 30 GB/s 61 GB/s 121 GB/s
PCIe 7.0 15 GB/s 30 GB/s 61 GB/s 121 GB/s 242 GB/s

The bandwidth figures above are approximate per-direction theoretical limits; protocol overhead, chipset design, and shared lanes can reduce real-world throughput.

Storage

Fast NVMe (Gen4/5): Accelerates initial model loading and dataset shuffling. Once the model is in VRAM, memory bandwidth and quantization primarily drive TPS.

System RAM: Offloading and KV Cache Paging

System RAM: Higher capacity enhances performance during weight offloading or KV cache paging to system RAM.

Offload parts of the model to system RAM or NVMe when the model or its KV cache exceeds the GeForce RTX 5080 allocation of 16 GB, or when using long context lengths. Note that offloading to an NVMe would be extremely slow but may allow for larger models to run.

For example, 30B models typically exceed 16 GB on this GPU, so plan for quantization and offloading.

Offloading Overview:

  • When to offload: out-of-memory errors, large contexts such as 16k to 32k+ tokens, or 30B to 70B models on 16 to 24 GB VRAM.
  • CPU offload: moves weights or the KV cache to system RAM. This is slower than VRAM but keeps sessions running when memory is tight.
  • NVMe offload: uses fast SSD storage as overflow for weights or cache when VRAM and RAM are not sufficient. This is much slower than VRAM and can bottleneck under heavy load.
  • KV cache paging: saves VRAM by placing the attention cache in system RAM. Latency increases with longer contexts and higher batch sizes.
  • Signs of active offloading: high PCIe traffic, SSD I/O spikes, and steady but reduced GPU utilization.

Note: A 70B model needs ~35 GB for weights at 4-bit quantization and ~140 GB at FP16. Long context windows add significant memory for the KV cache, potentially tens of GB, depending on sequence length and batch size.

Power & Cooling for the GeForce RTX 5080

Section Link

When building a system for AI workloads, prioritize PSU headroom and effective cooling to handle sustained high loads without throttling or instability.

The GeForce RTX 5080 typically draws 360 W under load, though this can vary based on the specific model and usage.

The manufacturer recommends at least a 850 W system PSU for standard configurations. For AI setups with high-core CPUs, multiple drives, premium variants, or overclocking, opt for additional headroom (e.g., 100–200 W extra).

For detailed guidance on selecting a power supply, see our NVIDIA RTX 5080 PSU guide.

As a high-power card at 360 W, the GeForce RTX 5080 demands a sturdy 850 W PSU (80+ Gold efficiency or higher) and superior cooling for reliable AI operation.

AI-specific considerations:

  • Extended training can sustain near-max power for hours or days, so plan accordingly.
  • Advanced users may undervolt for 10–15% reductions in heat and consumption.
  • Custom air or liquid cooling enhances stability in dedicated AI workstations.
  • Factor in elevated electricity costs for heavy development cycles.
  • Choose 80+ Gold or Platinum PSUs for better efficiency and longevity.

Our case recommendations for cooling high‑performance GPUs:

Undervolting the GeForce RTX 5080 for Efficiency

On the GeForce RTX 5080, undervolting reduces voltage at a target clock to cut power, heat, and fan noise. Most vendors offer software controls, and third‑party tools can help. Tuned well, it keeps tokens/sec steadier in long AI or rendering runs with little to no performance loss.

How it works: Pick a sustainable clock under load, set a lower voltage at that point on the voltage‑frequency curve, and keep the curve flat to the right so the GPU does not boost above your target. Lower voltage reduces power and temperature, which also reduces throttling.

Cautions: Chip-to-chip variation can be substantial, and overly aggressive undervolting may lead to instability or silent computational errors. Tuning could potentially void warranties, so proceed cautiously with small adjustments, monitor system stability closely, and always maintain a reliable backup profile for reverting if needed.

Ecosystem & OS Support for the GeForce RTX 5080

Section Link

Software support differs by vendor and operating system, impacting setup complexity and application compatibility.

Best Supported Frameworks and Tools: PyTorch, TensorFlow, LM Studio, Ollama, and Text Generation WebUI; robust CUDA integration.

Key Considerations: NVIDIA's ecosystem is the most mature and widely optimized for AI workloads, with CUDA acceleration standard across tools. Blackwell‑series GPUs add improved support for lower‑precision modes and efficient 8‑bit quantization for enhanced inference speed on compatible models.

Compute Platform Overview for the GeForce RTX 5080

The GeForce RTX 5080 leverages CUDA for optimized performance.

Different GPU vendors employ distinct compute platforms for LLM acceleration. Here's a comparison of their compatibility for LLM workloads:

Compute Platform Compatibility for LLM Applications
Platform Main Hardware LLM Usage Compatibility Notes
NVIDIA CUDA NVIDIA GPUs PyTorch; TensorFlow; HuggingFace; vLLM; Ollama; LM Studio NVIDIA CUDA provides the most extensive ecosystem and refined tooling; the Blackwell series introduces native FP8 support for accelerated inference on compatible models.
AMD ROCm AMD Radeon/Instinct PyTorch; TensorFlow; Ollama; LM Studio AMD ROCm integrates effectively with key frameworks, prioritizing Linux while Windows compatibility continues to improve; FP8 may necessitate custom builds or updated ROCm versions.
Intel oneAPI Intel Arc/Data Center PyTorch via IPEX; TensorFlow; Ollama via IPEX; LM Studio; NPU on Core Ultra Intel oneAPI + IPEX are advancing rapidly, including NPU extensions on Core Ultra; FP8 support varies by framework and release.
Apple Metal Apple M series Ollama; PyTorch (MPS); MLX Apple Metal shines in native Mac applications, with MLX facilitating optimized workflows.

OpenCL and Vulkan offer cross-vendor alternatives to vendor-specific GPU frameworks for accelerating large language models.

LLM OS Support

The operating system you choose significantly affects driver stability, feature availability, and performance for the GeForce RTX 5080:

  • Linux: Best overall for CUDA and ROCm. Updates and drivers arrive fastest, and most vendor tooling targets Linux first. Intel oneAPI and Intel Extension for PyTorch (IPEX) are mature on Linux for both CPU and Intel GPUs.
  • Windows: Strong native CUDA support. ROCm on Windows is emerging and tied to specific versions and hardware. AMD documents WSL-based ROCm today and has stated Windows will be a first-class target with ROCm 7. Intel IPEX now provides official Windows builds for Intel GPUs and supports modern PyTorch features. User experience is friendlier for general desktop use.
  • WSL2: Well supported for CUDA workflows. NVIDIA reports near-native performance for long GPU kernels, although some workloads can see I/O or PCIe-related overhead versus native Linux. AMD provides an official ROCm-on-WSL path for select Radeon GPUs, but support is not full parity with native Linux and depends on specific driver stacks.

Final Recommendation for the GeForce RTX 5080

Section Link

The GeForce RTX 5080 is a solid pick for running LLMs.

  • The RTX 5080 handles 7B at BF16/FP16/FP8 (where supported) and 13B to 14B with 8‑bit.
  • With the RTX 5080, you can run up to 24B LLM models at practical quantization.
  • The RTX 5080 has 16 GB of GDDR7 VRAM, which makes it Good for LLM workloads.
  • With the RTX 5080, larger models need quantization or offloading; ensure software support for your platform.

What to Buy Instead

If you need more headroom for local LLMs, consider these GPUs: