Why gst_base_sink_is_too_late!

Anton Olegovich ivanushkin.anton.olegovich at gmail.com
Tue Sep 2 05:29:54 PDT 2014


On 02.09.2014 15:58, Tim Müller wrote:
> On Tue, 2014-09-02 at 15:40 +0400, Anton Olegovich wrote:
>
> Hi,
>
>> Because of what can be this Logcat output:
>>
>> W/GStreamer+basesink(2096): 0:20:11.970461533 0x551bcd50
>> gstbasesink.c:2789:gst_base_sink_is_too_late:<autovideosink-actual-sink-glimage>
>> warning: A lot of buffers are being dropped.
>> W/GStreamer+basesink(2096): 0:20:11.970658764 0x551bcd50
>> gstbasesink.c:2789:gst_base_sink_is_too_late:<autovideosink-actual-sink-glimage>
>> warning: There may be a timestamping problem, or this computer to slow
>>
>> And what does it mean?
> Could you tell us more about when this happens? What's the pipeline?
>
> Often it either means the system is too slow (e.g. decoding an 1080p
> video in software on an embedded system), or there's a problem with
> latency reporting in a live streaming pipeline.
>
> Cheers
>   -Tim
>
Here are the pipelines

on transmitter:

v4l2src device=/dev/video6 num-buffers=-1 always-copy=false queue-size=1 
! TIPrepEncBuf numOutputBufs=1 contiguousInputFrame=false !  TIVidenc1 
engineName=codecServer codecName=h264enc contiguousInputFrame=true 
rateControlPreset=1 ! rtph264pay pt=96 ! udpsink (with properties)

void create_pipe_universal_elements(void)
{
   GstBus *bus;

   pipe.element.pipeline = gst_pipeline_new("camera_pipeline");
   bus = gst_pipeline_get_bus(GST_PIPELINE (pipe.element.pipeline));
   pipe.bus_watch_id = gst_bus_add_watch (bus, bus_call, pipe.loop);     
//Регистрируем message handler
   gst_object_unref(bus);
   pipe.element.video_src = gst_element_factory_make("v4l2src", "videosrc");
   g_object_set(GST_OBJECT(pipe.element.video_src), "device", 
get_isp_out_dev_name(), NULL);
   pipe.element.sink = gst_element_factory_make("udpsink", "videoudpsink");
}

int create_pipe_elements_video(void)
{
   pipe.element.video_prep_buffer = 
gst_element_factory_make("TIPrepEncBuf", "TIPrepEncBuf");
   pipe.element.video_encoder = gst_element_factory_make("TIVidenc1", 
"TIVidenc1");
   pipe.element.rtph264pay = gst_element_factory_make("rtph264pay", 
"rtph264pay");
   if(!pipe.element.video_src || !pipe.element.video_prep_buffer || 
!pipe.element.video_encoder ||\
       !pipe.element.rtph264pay || !pipe.element.sink)
     {
       warnx(ERROR_GST_ELEMENT_CREATION);
       return FALSE;
     }
   gst_bin_add_many(GST_BIN (pipe.element.pipeline), 
pipe.element.video_src, \
         pipe.element.video_prep_buffer, pipe.element.video_encoder, 
pipe.element.rtph264pay, pipe.element.sink, NULL);
   if(TRUE != gst_element_link_many(pipe.element.video_src, \
       pipe.element.video_prep_buffer, pipe.element.video_encoder, 
pipe.element.rtph264pay, pipe.element.sink, NULL))
     {
       warnx("Can not link video branch.");
       return FALSE;
     }
   g_object_set(GST_OBJECT(pipe.element.video_prep_buffer), 
"numOutputBufs", 1, "contiguousInputFrame", FALSE, NULL);
   g_object_set(GST_OBJECT(pipe.element.video_encoder), "engineName", 
"codecServer", "codecName", "h264enc",
       "contiguousInputFrame", TRUE, "rateControlPreset", 
IVIDEO_LOW_DELAY, NULL);
   g_object_set(GST_OBJECT(pipe.element.rtph264pay), "pt", 96, NULL);
   return TRUE;
}


on resiver:

"udpsrc,rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink"


data->video_pipeline = gst_pipeline_new ("videopipeline");
gst_bin_add_many(GST_BIN(data->video_pipeline),udpsrc,rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink,NULL);

   if (!gst_element_link_filtered (udpsrc,rtph264depay,udpsrc_caps)){
       GST_ERROR ("Can't link udpsrc and rtph264depay with caps");
   }

   if (!gst_element_link_many 
(rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink,NULL)){
       GST_ERROR ("Can't link many to tee");
   }
  /* if (!gst_element_link_many (queue,autovideosink,NULL)){
        GST_ERROR ("Can't link queue and videosink");
    }*/

     gst_object_unref (G_OBJECT(videorate_src_pad));//Возможно получится 
ошибка с пямятью
     gst_caps_unref(videorate_caps);///Освобождаем caps

this two pipelines impliments in applications for DM3730 and Android

How can I realise

latency reporting ?



More information about the gstreamer-devel mailing list