On 25.02.2025 06:36, Scott Theisen wrote:
On 2/22/25 08:16, Timo Rothenpieler wrote:
On 22.02.2025 03:52, Scott Theisen wrote:
On 2/21/25 08:26, Timo Rothenpieler wrote:
On 20.02.2025 21:37, Scott Theisen wrote:
The default value of CuvidContext::nb_surfaces was reduced from 25 to 5 (as (CUVID_MAX_DISPLAY_DELAY + 1)) in 402d98c9d467dff6931d906ebb732b9a00334e0b.

In cuvid_is_buffer_full() delay can be 2 * CUVID_MAX_DISPLAY_DELAY with double
rate deinterlacing.  ctx->nb_surfaces is CUVID_DEFAULT_NUM_SURFACES =
(CUVID_MAX_DISPLAY_DELAY + 1) by default, in which case cuvid_is_buffer_full() will always return true and cuvid_output_frame() will never read any data since
it will not call ff_decode_get_packet().

It's been way too long since I looked at all that code, and I didn't even write most of the code involved:
https://github.com/FFmpeg/FFmpeg/commit/ bddb2343b6e594e312dadb5d21b408702929ae04 https://github.com/FFmpeg/FFmpeg/ commit/402d98c9d467dff6931d906ebb732b9a00334e0b

But doesn't this instead mean that the logic in cuvid_is_buffer_full is flawed somehow?

I think it is the number of frames ready to send to the driver + the number of frames in queue in the driver >= the number of decoded frame buffers.  However, it doesn't actually know how many frames are in queue in the driver and assumes the maximum.

Not sure if I understand you right, but the way it works is that av_fifo_can_read(ctx->frame_queue) returns how many frames have already been returned from cuvid and are ready for cuviddec.c to return them.

To that number, the maximum possible number of delayed frames is added, which could be returned by the decoder without feeding in any more input frames.

If that number reaches the desired amount of surfaces to buffer, cuvid_is_buffer_full() will report that its buffer is full, and cuviddec.c will stop fetching new input.

I think the point I was trying to make is that since it doesn't track how many frames have been sent to and received from the driver it always assumes the maximum number of delayed frames are available in the driver, which is obviously not correct when no frames have been sent to the driver.


Just increasing the default number of surfaces does not seem like the correct fix or sensible, since it will increase VRAM usage by potentially quite a bit for all users.


The changes to cuvid_handle_video_sequence() from 402d98c9d467dff6931d906ebb732b9a00334e0b will increase nb_surfaces once data has been read.

Only if the decoder reports that it will potentially buffer even more frames.

From looking at this a bit, the issue will only happen when deinterlacing, the logic in cuvid_is_buffer_full becomes stuck then, and will always claim the buffer is full. And from my understanding, it's correct in making that claim. Due to the display delay, it could in theory happen that the moment cuvid starts outputting frames, there will be more output available than what fits into ctx->frame_queue, since it delayed by 4 frames, which results in 8 surfaces, but the queue only fits 5.

So to me it looks like that the correct fix would be to double the size of the frame_queue when deinterlacing, not unconditionally.

There is nothing stopping deint_mode or drop_second_field from being changed after cuvid_decode_init() is called, so it doesn't necessarily know it will deinterlace.

Regardless, 402d98c9d467dff6931d906ebb732b9a00334e0b reduced CUVID_DEFAULT_NUM_SURFACES from 25 to *only 5* to not break playback entirely.  I don't think the intention was to break playback for double rate deinterlacing while allowing playback for only single rate deinterlacing.

Also, if AV_CODEC_FLAG_LOW_DELAY is set, then only one output surface is needed, but there are still 5.

The structs stored in the ctx->frame_queue aren't what's using all the memory. It's the frames buffered by cuvid itself, which are referred to by that buffer, so having it be larger than what cuvid will actually buffer doesn't hurt all that much.
But yeah, it could be shrunk in this case.

What this whole dance is actually trying to accomplish is to prevent the number of "ready but not-yet-returned frames" to never exceed the max value possible by cuvid set via ulNumDecodeSurfaces/ ulMaxNumDecodeSurfaces, which is determined and stored in ctx- >nb_surfaces during cuvid_handle_video_sequence().

In the default mode of operation, the buffer_full indicator will indeed stop pulling new input the moment even one frame is returned. But that's fine, since at that point already a bunch of input has been consumed, and a decent delay has been built up already. In low-delay mode frames are pretty much returned the moment it's possible.

So looking at all this, I still think the core of the issue is incorrect handling of deinterlacing in all this. CUVID treats a deinterlaced frame as one internal frame, but it's stored in the frame_queue as two frames.

So in the case of deinterlacing without drop_second_field, the size of that queue needs to be doubled, but nb_surfaces must stay the same, since for cuvid itself it's still just one frame. And in turn the is_buffer_full function has to be adjusted to multiply nb_surfaces by two if deinterlacing and not drop_second_field.

So you think it needs to be something like this would work?:
```
diff --git a/mythtv/external/FFmpeg/libavcodec/cuviddec.c b/mythtv/ external/FFmpeg/libavcodec/cuviddec.c
index 81ac54297e..535ff7afb5 100644
--- a/mythtv/external/FFmpeg/libavcodec/cuviddec.c
+++ b/mythtv/external/FFmpeg/libavcodec/cuviddec.c
@@ -317,13 +317,13 @@ static int CUDAAPI cuvid_handle_video_sequence(void *opaque, CUVIDEOFORMAT* form
          return 0;
      }

-    fifo_size_inc = ctx->nb_surfaces;
+    fifo_size_inc = av_fifo_can_read(ctx->frame_queue) + av_fifo_can_write(ctx->frame_queue);      ctx->nb_surfaces = FFMAX(ctx->nb_surfaces, format- >min_num_decode_surfaces + 3);

      if (avctx->extra_hw_frames > 0)
          ctx->nb_surfaces += avctx->extra_hw_frames;

-    fifo_size_inc = ctx->nb_surfaces - fifo_size_inc;
+    fifo_size_inc = (ctx->nb_surfaces * 2 ) - fifo_size_inc; // copy * 2 logic from cuvid_is_buffer_full()?      if (fifo_size_inc > 0 && av_fifo_grow2(ctx->frame_queue, fifo_size_inc) < 0) {          av_log(avctx, AV_LOG_ERROR, "Failed to grow frame queue on video sequence callback\n");
          ctx->internal_error = AVERROR(ENOMEM);
@@ -417,10 +417,11 @@ static int cuvid_is_buffer_full(AVCodecContext *avctx)
      CuvidContext *ctx = avctx->priv_data;

      int delay = ctx->cuparseinfo.ulMaxDisplayDelay;
+    int output_frames = 1;
     if (ctx->deint_mode != cudaVideoDeinterlaceMode_Weave && !ctx- >drop_second_field)
-        delay *= 2;
+        output_frames = 2;

-    return av_fifo_can_read(ctx->frame_queue) + delay >= ctx->nb_surfaces;
+    return av_fifo_can_read(ctx->frame_queue) + (delay * output_frames) >= (ctx->nb_surfaces * output_frames); // should this be >= av_fifo_can_read(ctx->frame_queue) + av_fifo_can_write(ctx->frame_queue) ?
  }

 static int cuvid_decode_packet(AVCodecContext *avctx, const AVPacket *avpkt) @@ -899,7 +900,7 @@ static av_cold int cuvid_decode_init(AVCodecContext *avctx)
      if(ctx->nb_surfaces < 0)
          ctx->nb_surfaces = CUVID_DEFAULT_NUM_SURFACES;

-    ctx->frame_queue = av_fifo_alloc2(ctx->nb_surfaces, sizeof(CuvidParsedFrame), 0); +    ctx->frame_queue = av_fifo_alloc2(ctx->nb_surfaces * 2, sizeof(CuvidParsedFrame), 0);
      if (!ctx->frame_queue) {
          ret = AVERROR(ENOMEM);
          goto error;
```

Something like that, I'm looking into it as well.
Been too long since I looked at that code.

It is odd that there is no function to return AVFifo::nb_elems directly.  I think ctx->nb_surfaces was incorrectly used instead of AVFifo::nb_elems.

Does frame_queue even need to grow, since ulMaxDisplayDelay is fixed at 4 (or 0)?

The frame queue contains up to ulMaxNumDecodeSurfaces surfaces, not up to ulMaxDisplayDelay. ulMaxDisplayDelay indicated how many more frames might be in-flight and be spontaneously dumped out all at once by cuvid on the next input packet, so the queue needs to be kept free enough to contain them all.

So: The frame_queue needs to be able to contain up to ulMaxNumDecodeSurfaces frames. When deinterlacing and not dropping the second field, each DecodeSurface produces two frames, so the queue needs to double in size then.


Changing that stuff at runtime is no problem, since to change anything, cuvid_handle_video_sequence() has to re-run, which updates all these sizes and will resize the fifo accordingly. And when turning it off, the only thing that happens is that buffer_full will report it's full immediately, and some frames need to be read out before accepting input again.

There's also the edge case of "half a frame" having already been returned, so the queue potentially no longer being considered full, but still all decode surfaces are in use, since the other half of that deinterlaced frame is still in the queue.
So special care must be taken to not report buffer-free too early.


I'm not understanding you here.  Aren't the elements in frame_queue already extracted from the driver and owned by FFmpeg?

No, the ctx->frame_queue frames are still cuvid frames and will stay that until cuvid_output_frame() removes them from the queue and downloads them into FFmpeg owned memory.

Thus the amount of distinct frames in ctx->frame_queue must never exceed that of ulMaxNumDecodeSurfaces. And new packets can only be fed to cuvid while the frame_queue still has space to accept up to ulMaxDisplayDelay (*2 when deinterlacing) of new frames, which might spontaneously come out of cuvid at any moment.
_______________________________________________
ffmpeg-devel mailing list
ffmpeg-devel@ffmpeg.org
https://ffmpeg.org/mailman/listinfo/ffmpeg-devel

To unsubscribe, visit link above, or email
ffmpeg-devel-requ...@ffmpeg.org with subject "unsubscribe".

Reply via email to