GPU Benchmark! 4070 4080 4090 3080 3090 #2970
Replies: 53 comments 38 replies
-
2070 rtx mobile |
Beta Was this translation helpful? Give feedback.
-
Nvidia RTX 4090 FE 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:03<00:00, 6.17it/s] (edit: used default resolution first go around) |
Beta Was this translation helpful? Give feedback.
-
CPU vs GPU 1024x1024 2 steps Intel 13900KS 24,25 sec. https://files.catbox.moe/f4wb53.mp4 CPU.vs.GPU.test_2.mp4 |
Beta Was this translation helpful? Give feedback.
-
RTX 3090 |
Beta Was this translation helpful? Give feedback.
-
Just for laughs ... |
Beta Was this translation helpful? Give feedback.
-
RTX 3070 OC Drivers:~528.xx 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:08<00:00, 2.26it/s] |
Beta Was this translation helpful? Give feedback.
-
My poor old GPU 😢 |
Beta Was this translation helpful? Give feedback.
-
RTX 4080
|
Beta Was this translation helpful? Give feedback.
-
Would love to get some 3060 / 4060 data to better complete the regression... Then I'll post proposed for every 3000/4000 card. |
Beta Was this translation helpful? Give feedback.
-
RTX 4070 12Gb got prompt |
Beta Was this translation helpful? Give feedback.
-
WSL2 (6.1.x kernel) 3060Ti
Tesla P40
No xformers 3060ti
Tesla P40
|
Beta Was this translation helpful? Give feedback.
-
4090 FE, Ryzen 7900X, Samsung 980 2TB M.2 got prompt |
Beta Was this translation helpful? Give feedback.
-
RTX 3070 Laptop GPU
|
Beta Was this translation helpful? Give feedback.
-
win11 wsl2 ubuntu 100%|███████████████████████████████████████████████████████████████████████████████████| 20/20 [00:04<00:00, 4.03it/s] |
Beta Was this translation helpful? Give feedback.
-
got prompt |
Beta Was this translation helpful? Give feedback.
-
|
Beta Was this translation helpful? Give feedback.
-
RTX 4070 Ti Super 4.8 seconds, |
Beta Was this translation helpful? Give feedback.
-
RTX 3060 12GB |
Beta Was this translation helpful? Give feedback.
-
Made new benchmarks on a more recent version of ComfyUI and PyTorch 2.6(for AMD 2.7 nightly was used) This is already the second series of benchs, we tried to take into account the shortcomings of the first benchmarks and made that for each Task a new prompt is given, so that the text encoder is also tested. Similar benchmark suites were also removed, we tried to use 30 steps for generation everywhere Full interactive results link for those who is interested |
Beta Was this translation helpful? Give feedback.
-
RTX 5070ti, overclocked (+7% roughly). 100%█████████████████████████| 20/20 [00:04<00:00, 4.46it/s] Update 8 march 2025: Torch has been updated to version 2.7, testing on the new Comfi build: |
Beta Was this translation helpful? Give feedback.
-
AMD Radeon 6800XT (2400MHz) ** Platform: Linux got prompt |
Beta Was this translation helpful? Give feedback.
-
7900 XTX with torch.compile node, ROCM FlashAttention, tuning and some unpublished ComfyUI fixes (soon):
I didn't know that was even possible on this card. |
Beta Was this translation helpful? Give feedback.
-
7950X3D 48GB DDR5 6200 / RTX 5090D 32GB model weight dtype torch.float16, manual cast: None |
Beta Was this translation helpful? Give feedback.
-
Total VRAM 24090 MB, total RAM 128714 MB [Crystools .[0;32mINFO.[0m] Crystools version: 1.22.1 |
Beta Was this translation helpful? Give feedback.
-
This is mine.. |
Beta Was this translation helpful? Give feedback.
-
Total VRAM 24576 MB, total RAM 81787 MB Requested to load SDXLClipModel 512x512 got prompt 1024x1024 got prompt |
Beta Was this translation helpful? Give feedback.
-
Hardware:
Software:
|
Beta Was this translation helpful? Give feedback.
-
My new PC has Windows 11, a Nvidia GeForce RTX 3060 Ti with a AMD Ryzen 5 3600 6-Core Processor. Is it okay or meh to generates AI arts with a big picutre size? |
Beta Was this translation helpful? Give feedback.
-
Could we Do a benchmark for GPU´s ?
Need to get new hardware .
To keep it simple we could use the default load worklow on 1024*1024 with sdxl 1.0.
Seed 1
I know that ram comes in play when workflows get more complex , but for the start simple ist best.
Please post the second run. first run ist a little bit corupted by loading data.
theory pudget:

toms hardware 1111:

Beta Was this translation helpful? Give feedback.
All reactions