[libcamera-devel] [RFC PATCH v2 5/5] libcamera: pipeline: uvcvideo: Handle metadata stream

Laurent Pinchart laurent.pinchart at ideasonboard.com
Mon Aug 28 23:50:11 CEST 2023


On Mon, Aug 21, 2023 at 10:56:20PM +0100, Kieran Bingham via libcamera-devel wrote:
> Quoting Gabby George (2023-08-21 14:10:39)
> > Register the metadata stream's buffer ready callback and start
> > processing metadata buffers.  Use the timestamp from the metadata
> > buffer as the corresponding video buffer Requests' timestamp. Metadata
> > buffers are synchronized with frames coming into the video stream
> > using the sequence field of the buffers. They may come in either order
> > (video buffer first or metadata buffer first), so store relevant
> > information about the buffer required to set the metadata timestamp or
> > complete the buffer request as soon as possible.
> > 
> > The timestamp will be improved upon in the next patch. For now, use
> > the driver-provided metadata timestamp.
> > 
> > Signed-off-by: Gabby George <gabbymg94 at gmail.com>
> > ---
> >  src/libcamera/pipeline/uvcvideo/uvcvideo.cpp | 168 ++++++++++++++++++-
> >  1 file changed, 162 insertions(+), 6 deletions(-)
> > 
> > diff --git a/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp b/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp
> > index c8d6633f..215435ec 100644
> > --- a/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp
> > +++ b/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp
> > @@ -12,6 +12,8 @@
> >  #include <memory>
> >  #include <tuple>
> >  
> > +#include <linux/uvcvideo.h>
> > +
> >  #include <libcamera/base/log.h>
> >  #include <libcamera/base/utils.h>
> >  
> > @@ -34,6 +36,17 @@ namespace libcamera {
> >  
> >  LOG_DEFINE_CATEGORY(UVC)
> >  
> > +/*
> > + * The UVCH buffer contains an unsigned char array
> > + * encoding UVC timing data that needs to be recast
> > + * into usable data.

You can reflow this to 80 columns.

> > + */
> > +struct UVCTimingBuf {
> > +       __u32 pts;
> > +       __u32 stc;
> > +       __u16 sofDevice;
> > +} __attribute__((packed));
> > +
> >  class UVCCameraData : public Camera::Private
> >  {
> >  public:
> > @@ -46,6 +59,8 @@ public:
> >         void addControl(uint32_t cid, const ControlInfo &v4l2info,
> >                         ControlInfoMap::Map *ctrls);
> >         void bufferReady(FrameBuffer *buffer);
> > +       void bufferReadyMetadata(FrameBuffer *buffer);
> > +       void handleUnfinishedRequests();
> >  
> >         const std::string &id() const { return id_; }
> >  
> > @@ -56,9 +71,16 @@ public:
> >         std::map<unsigned int, MappedFrameBuffer> mappedMetadataBuffers_;
> >  
> >         std::map<PixelFormat, std::vector<SizeRange>> formats_;
> > +       std::queue<FrameBuffer *> pendingVideoBuffers_;
> > +       std::queue<std::pair<unsigned int, uint64_t>> pendingMetadata_;
> >  
> >  private:
> >         int initMetadata(MediaDevice *media);
> > +       void completeRequest(FrameBuffer *buffer, uint64_t timestamp);
> > +       void endCorruptedStream();
> > +
> > +       const unsigned int frameStart_ = 1;
> > +       const unsigned int maxVidBuffersInQueue_ = 1;
> >  
> >         bool generateId();
> >  
> > @@ -242,7 +264,7 @@ int PipelineHandlerUVC::cleanupMetadataBuffers(Camera *camera)
> >         UVCCameraData *data = cameraData(camera);
> >  
> >         if (data->metadata_)
> > -               data->metadata_->releaseBuffers();
> > +               ret = data->metadata_->releaseBuffers();
> 
> I don't think this change is specific to this patch, so it shouldn't be
> here, but I don't think it would squash to a previous patch anyway as I
> already mentioned the cleanup functions may as well be void cleanup();
> 
> >         data->metadataBuffers_.clear();
> >         data->mappedMetadataBuffers_.clear();
> >         data->metadata_ = nullptr;
> > @@ -253,7 +275,9 @@ int PipelineHandlerUVC::cleanupMetadataBuffers(Camera *camera)
> >  int PipelineHandlerUVC::cleanup(Camera *camera)
> >  {
> >         UVCCameraData *data = cameraData(camera);
> > +
> >      Y   cleanupMetadataBuffers(camera);
> > +

This should be squashed with one of the previous patches.

> >         data->video_->releaseBuffers();
> >         return 0;
> >  }
> > @@ -354,6 +378,8 @@ void PipelineHandlerUVC::stopDevice(Camera *camera)
> >         if (data->metadata_)
> >                 data->metadata_->streamOff();
> >  
> > +       data->handleUnfinishedRequests();
> > +
> >         cleanup(camera);
> >  }
> >  
> > @@ -646,8 +672,11 @@ int UVCCameraData::init(MediaDevice *media)
> >         if (ret) {
> >                 metadata_ = nullptr;
> >                 LOG(UVC, Error) << "Could not find a metadata video device.";
> > +               return 0;

This is also unrelated to this patch.

> >         }
> >  
> > +       metadata_->bufferReady.connect(this, &UVCCameraData::bufferReadyMetadata);
> > +
> >         return 0;
> >  }
> >  
> > @@ -833,18 +862,145 @@ void UVCCameraData::addControl(uint32_t cid, const ControlInfo &v4l2Info,
> >         ctrls->emplace(id, info);
> >  }
> >  
> > -void UVCCameraData::bufferReady(FrameBuffer *buffer)
> > +void UVCCameraData::completeRequest(FrameBuffer *buffer, uint64_t timestamp)
> >  {
> >         Request *request = buffer->request();
> > -
> > -       /* \todo Use the UVC metadata to calculate a more precise timestamp */
> > -       request->metadata().set(controls::SensorTimestamp,
> > -                               buffer->metadata().timestamp);
> > +       request->metadata().set(controls::SensorTimestamp, timestamp);
> >  
> >         pipe()->completeBuffer(request, buffer);
> >         pipe()->completeRequest(request);
> >  }
> >  
> > +void UVCCameraData::handleUnfinishedRequests()
> > +{
> > +       while (!pendingVideoBuffers_.empty()) {
> > +               FrameBuffer *oldBuffer = pendingVideoBuffers_.front();
> > +               Request *oldRequest = oldBuffer->request();
> > +
> > +               oldRequest->metadata().set(controls::SensorTimestamp,
> > +                                          oldBuffer->metadata().timestamp);
> > +
> > +               pipe()->completeBuffer(oldRequest, oldBuffer);
> > +               pipe()->completeRequest(oldRequest);
> 
> Should this just call
> 	completeRequest(oldBuffer, oldBuffer->metadata().timestamp); ?
> 
> > +               pendingVideoBuffers_.pop();
> > +       }
> > +}
> > +
> > +void UVCCameraData::endCorruptedStream()
> > +{
> > +       handleUnfinishedRequests();
> > +       /* Close the metadata node so we don't get inaccurate timestamps*/
> > +       metadata_ = nullptr;
> > +       LOG(UVC, Error)
> > +               << "UVC metadata stream corrupted. Reverting to driver timestamps.";
> 
> Aha - so actually this doesn't stop the whole camera streaming - it just
> stops the metadata node?
> 
> Before closing the metadata_ should it release buffers and clean up
> those ?
> 
> > +}
> > +
> > +/*
> > + * If there is a metadata buffer that hasn't been matched with a
> > + * video buffer, check to see if it matches this video buffer.
> > + *
> > + * If there is a match, use the timestamp stored in the metadata queue
> > + * for this video buffer's request. Complete this video buffer
> > + * and its request.
> > + *
> > + * If there are no metadata buffers available to check for a match,
> > + * push this video buffer's request object to the queue. It may
> > + * be that the metadata buffer has not yet arrived.
> > + * When the matching metadata buffer does come in, it will handle
> > + * completion of the buffer and request.
> > + *
> > + * If more than maxVidBuffersInQueue_ video buffers have been added
> > + * to the queue, something is wrong with the metadata stream and
> > + * we can no longer use UVC metadata packets for timestamps.
> > + * Complete all of the outstanding requests and turn off metadata
> > + * stream use.
> > + */
> > +void UVCCameraData::bufferReady(FrameBuffer *buffer)
> > +{
> > +       /* \todo Use the UVC metadata to calculate a more precise timestamp */
> > +       if (!metadata_ || buffer->metadata().sequence < frameStart_) {
> > +               completeRequest(buffer, buffer->metadata().timestamp);
> > +               return;
> > +       }
> > +
> > +       if (!pendingMetadata_.empty()) {
> > +               /* A metadata buffer was ready first. */
> > +               unsigned int mdSequence = std::get<0>(pendingMetadata_.front()) + frameStart_;
> > +               if (mdSequence == buffer->metadata().sequence) {
> > +                       completeRequest(buffer, std::get<1>(pendingMetadata_.front()));
> > +                       pendingMetadata_.pop();
> > +                       return;
> > +               } else {
> > +                       /* \todo: Is there a reason metadata buffers can arrive out of order? */
> > +                       endCorruptedStream();
> > +                       return;
> > +               }
> > +       }
> > +
> > +       pendingVideoBuffers_.push(buffer);
> > +       /*
> > +        * Deal with video buffers that haven't been completed, and with
> > +        * buffers whose metadata information arrived out of order.
> > +        */
> > +       if (pendingVideoBuffers_.size() > maxVidBuffersInQueue_) {
> > +               endCorruptedStream();
> 
> I'm worried about all cases this patch now introduces that seem to now
> mean we'll abort a running camera stream. I wonder if that's just part
> of the complexity introduced by having metadata delivered in multiple
> video nodes though.
> 
> > +       }
> > +}
> > +
> > +void UVCCameraData::bufferReadyMetadata(FrameBuffer *buffer)
> > +{
> > +       if (!metadata_ ||
> > +           buffer->metadata().status != FrameMetadata::Status::FrameSuccess) {
> > +               return;
> > +       }
> > +
> > +       /*
> > +        * The metadata stream seems to start at seq 1 and libcamera
> > +        * sets the start sequence to 0, so it's necessary to add one
> > +        * to match this buffer with the correct video frame buffer.
> > +        *
> > +        * \todo: Is there a better way to do this?  What is the root cause?
> 
> This is odd to me. Both streams should start at 0 ... as they both do
> the same operation.

Maybe this is caused by queuing buffers for the metadata stream too late
?

> > +        */
> > +       unsigned int mdSequence = buffer->metadata().sequence + frameStart_;
> > +       int pos = buffer->cookie();
> > +
> > +       Span<uint8_t> memMeta = mappedMetadataBuffers_.at(pos).planes()[0];
> > +       uvc_meta_buf *metaBuf = reinterpret_cast<uvc_meta_buf *>(memMeta.data());
> > +
> > +       //Span<uint8_t> memTime = mappedMetadataBuffers_.at(pos).planes()[0];
> > +       //UVCTimingBuf * timeBuf = reinterpret_cast<UVCTimingBuf *>(&memTime.data()[sizeof(uvc_meta_buf)]);
> 
> Patches shouldn't introduce dead-code.
> 
> > +
> > +       size_t UVCPayloadHeaderSize = sizeof(metaBuf->length) +
> > +                                     sizeof(metaBuf->flags) + sizeof(UVCTimingBuf);
> > +       if (metaBuf->length < UVCPayloadHeaderSize) {
> 
> If I understand correctly - this could be checked at the point the
> buffers are allocated, and will always be true/ok from that point onwards,
> so we could remove a fail path from the buffer completion handlers?
> 
> > +               endCorruptedStream();
> > +               return;
> > +       }
> > +
> > +       /*
> > +        * Match a pending video buffer with this buffer's sequence.  If
> > +        * there is none available, put this timestamp information on the
> > +        * queue. When the matching video buffer does come in, it will use
> > +        * a timestamp from this metadata.
> > +        */
> > +       if (!pendingVideoBuffers_.empty()) {
> > +               FrameBuffer *vidBuffer = pendingVideoBuffers_.front();
> > +               unsigned int vidSequence = vidBuffer->metadata().sequence;
> > +
> > +               if (vidSequence == mdSequence) {
> > +                       completeRequest(vidBuffer, static_cast<uint64_t>(metaBuf->ns));
> > +                       pendingVideoBuffers_.pop();
> > +               } else {
> > +                       endCorruptedStream();
> 
> The big worry here is going to be what happens when a frame gets dropped
> (for any reason, including libcamera not running fast enough, or the
> application not queuing buffers fast enough) ... so we'll have to deal
> with this in a much safer way than just giving up and closing the
> streams.

Definitely :-)

> I suspect what would really be needed is to put both buffer completions
> onto separate queues, that then triggers the function that will sort
> through both queues. When it has a match, it would complete the request.
> If it detects a drop or an offset, it would complete with an error or
> the old timestamp or such.
> 
> I wonder if the implementors of the metadata video nodes considered
> these cases...

I'm not sure what needs to be considered on the kernel side. The above
is how it should be handled in userspace.

> > +               }
> > +       } else {
> > +               pendingMetadata_.push(
> > +                       std::make_pair(buffer->metadata().sequence,
> > +                                      static_cast<uint64_t>(metaBuf->ns)));
> > +       }
> > +       metadata_->queueBuffer(buffer);
> > +}
> > +
> >  REGISTER_PIPELINE_HANDLER(PipelineHandlerUVC)
> >  
> >  } /* namespace libcamera */

-- 
Regards,

Laurent Pinchart


More information about the libcamera-devel mailing list