RTX 3050 - Order Now
Home / Blog / Benchmarks / Stable Diffusion XL on RTX 4060 Ti: Images/sec & VRAM Usage, Category: Benchmarks, Slug: sdxl-on-rtx-4060-ti-benchmark, Excerpt: Stable Diffusion XL benchmarked on RTX 4060 Ti: 1.9 it/s, 3.8 images/min at 1024×1024, VRAM usage, and cost per 1K images., Internal links: 8 –>
Benchmarks

Stable Diffusion XL on RTX 4060 Ti: Images/sec & VRAM Usage, Category: Benchmarks, Slug: sdxl-on-rtx-4060-ti-benchmark, Excerpt: Stable Diffusion XL benchmarked on RTX 4060 Ti: 1.9 it/s, 3.8 images/min at 1024×1024, VRAM usage, and cost per 1K images., Internal links: 8 –>

Stable Diffusion XL benchmarked on RTX 4060 Ti: 1.9 it/s, 3.8 images/min at 1024x1024, VRAM usage, and cost per 1K images., Internal links: 8 -->

The RTX 4060 Ti doubles the standard 4060’s VRAM to 16 GB, which fundamentally changes what you can do with Stable Diffusion XL. Higher-resolution outputs, SDXL plus refiner in one pipeline, LoRA stacking — all become practical. And the raw speed improvement is meaningful too. We ran the numbers on GigaGPU dedicated hardware.

Benchmark Results

MetricValue
Iterations/sec1.9 it/s
Seconds per image15.79 sec (30 steps)
Images per minute3.8
Resolution1024×1024
SamplerEuler a / DPM++ 2M Karras
Performance ratingVery Good

30-step generation, 1024×1024, FP16, batch size 1. Under 16 seconds per image puts SDXL into a comfortable creative rhythm — generate, review, tweak the prompt, and generate again without losing momentum.

The VRAM Advantage

ComponentVRAM
Model weights6.5 GB
Sampling buffer~1.3 GB
Total RTX 4060 Ti VRAM16 GB
Free headroom~9.5 GB

9.5 GB free is a generous surplus for image generation. You can comfortably run the SDXL base model plus the refiner, load multiple LoRA adapters simultaneously, generate at resolutions beyond 1024×1024, or use batch sizes greater than 1 for faster throughput. This is the first GPU in the range where SDXL feels genuinely unconstrained.

Cost Efficiency

Cost MetricValue
Server cost£0.50/hr (£99/mo)
Cost per 1K images£2.19
Images per £1457

The per-image cost of £2.19/K is only marginally above the 4060 (£2.08/K), but the 4060 Ti produces 36% more images per hour and opens up workflows that the 4060 cannot support. For teams generating thousands of images daily, the £30/mo premium over the base 4060 pays for itself in throughput alone. See the best GPU for Stable Diffusion guide for a detailed comparison.

Recommendation

This is the strongest mid-range SDXL card in the lineup. Teams building production image pipelines, e-commerce product shot generators, or ComfyUI automation workflows will find the 4060 Ti’s combination of speed, VRAM, and price hard to beat below the RTX 3090 tier.

Launch it:

docker run --gpus all -p 7860:7860 ghcr.io/ai-dock/stable-diffusion-webui:latest

Guides: SDXL hosting, best GPU for SD, all benchmarks. Also: Flux.1 hosting.

SDXL Unleashed on the RTX 4060 Ti

16 GB of VRAM, 3.8 images/min, £99/mo. UK datacentre, root access included.

Order 4060 Ti Server

Need a Dedicated GPU Server?

Deploy from RTX 3050 to RTX 5090. Full root access, NVMe storage, 1Gbps — UK datacenter.

Browse GPU Servers

admin

We benchmark, deploy, and optimise GPU infrastructure for AI workloads. All data in our guides comes from real-world testing on our UK-based dedicated GPU servers.

Ready to deploy your AI workload?

Dedicated GPU servers from our UK datacenter. NVMe storage, 1Gbps networking, full root access.

Browse GPU Servers Contact Sales

Have a question? Need help?