4EvrYng wrote:Mads Johansen wrote:The only thing I can think of is to create a bat file with ...
I ended up creating batch file that kept calling ffmpeg.exe in the loop. Load on CPU was 45-75% very evenly spread across all 24 cores. In other words, ffmpeg efficiently used every single resource at its disposal and can't be used to judge whether Resolve will benefit from GPU H264 acceleration, pattern of CPU usage between the two is obviously very different.
Mads Johansen wrote:I haven't tested if we can subtract Davinci from itself ...
You have 5070 Ti, correct? IIRC one can tell Resolve whether to use GPU or CPU for decoding, which could give an idea how much H264 GPU decoding helps in Resolve, how big is an impact on CPU. If yes what kind of file you would need to test that, please? Would 4K SI file from FX30 work?
Mads Johansen wrote:The other thing I've been thinking of is to indirectly measure how long it takes to go one frame back in a long GOP file: Record via OBS and go to the end of a known GOP file, then go one frame back. Count the frames between button press and resolve responds = GOP length/response time = decode fps. But it's tangential to the cpu usage

That's a very good idea. Using 2 machine setup with recording card would eliminate any concerns over vampire system load. Unfortunately I don't have a setup I could test that with.
We're still talking about H.264 intra right? My point is that H.264 intra is not worth accelerating because each frame can be decoded almost instantly on the cpu anyway and that other processing will take much longer almost in every case, so I do not think decoding H.264 intra is worth benchmarking.
(Edit later: I was wrong:
ffmpeg -hwaccel cuda -i S_080724_005.MP4 -f null NUL
128 fps
ffmpeg -i S_080724_005.MP4 -f null NUL
453 fps )
In general, for the flavours of H.264 and H.265 and AV1 and and and.. that are computationally expensive to decode, yes those are worth benchmarking.
Since we can't measure anything with Davinci (everything plays at 120 fps which is the maximum frame rate), I used ffmpeg.
I'm going to make a more detailed post with details and reproduction steps, but as a rough guide for decoding with a 5070TI with a H.264 4500 frame Mandelbrot file:
ffmpeg -hwaccel cuda -i mandelbrot-yuv420p-1080p.mov -f null NUL
- Code: Select all
CUDA at 1080p:
yuv420p: 645 fps
yuv420p10le: 487 fps
yuv422p: 486 fps
yuv422p10le: 370 fps
CUDA at 2160p:
yuv420p: 191 fps
yuv420p10le: 145 fps
yuv422p: 138 fps
yuv422p10le: 114 fps
- Code: Select all
Intel Ultra 7 265k at 1080p:
yuv420p: 726 fps
yuv420p10le: 721 fps
yuv422p: 523 fps
yuv422p10le: 515 fps
Intel Ultra 7 265k at 2160p:
yuv420p: 203 fps
yuv420p10le: 197 fps
yuv422p: 147 fps
yuv422p10le: 141 fps
(Which surprised me a lot. And there's no hardware decoding in the Intel)
I also don't know what a 4K SI file from FX30 is.
Now we have a better supported matrix than before at least

Davinci Resolve Studio 20 build 49, Windows 11, Ultra 7 265k, Nvidia 5070 TI, 576.80 Studio