Why gst_base_sink_is_too_late!

Anton Olegovich ivanushkin.anton.olegovich at gmail.com
Tue Sep 2 08:40:44 PDT 2014


On 02.09.2014 16:29, Anton Olegovich wrote:
>
> On 02.09.2014 15:58, Tim Müller wrote:
>> On Tue, 2014-09-02 at 15:40 +0400, Anton Olegovich wrote:
>>
>> Hi,
>>
>>> Because of what can be this Logcat output:
>>>
>>> W/GStreamer+basesink(2096): 0:20:11.970461533 0x551bcd50
>>> gstbasesink.c:2789:gst_base_sink_is_too_late:<autovideosink-actual-sink-glimage> 
>>>
>>> warning: A lot of buffers are being dropped.
>>> W/GStreamer+basesink(2096): 0:20:11.970658764 0x551bcd50
>>> gstbasesink.c:2789:gst_base_sink_is_too_late:<autovideosink-actual-sink-glimage> 
>>>
>>> warning: There may be a timestamping problem, or this computer to slow
>>>
>>> And what does it mean?
>> Could you tell us more about when this happens? What's the pipeline?
>>
>> Often it either means the system is too slow (e.g. decoding an 1080p
>> video in software on an embedded system), or there's a problem with
>> latency reporting in a live streaming pipeline.
>>
>> Cheers
>>   -Tim
>>
> Here are the pipelines
>
> on transmitter:
>
> v4l2src device=/dev/video6 num-buffers=-1 always-copy=false 
> queue-size=1 ! TIPrepEncBuf numOutputBufs=1 contiguousInputFrame=false 
> !  TIVidenc1 engineName=codecServer codecName=h264enc 
> contiguousInputFrame=true rateControlPreset=1 ! rtph264pay pt=96 ! 
> udpsink (with properties)
>
> void create_pipe_universal_elements(void)
> {
>   GstBus *bus;
>
>   pipe.element.pipeline = gst_pipeline_new("camera_pipeline");
>   bus = gst_pipeline_get_bus(GST_PIPELINE (pipe.element.pipeline));
>   pipe.bus_watch_id = gst_bus_add_watch (bus, bus_call, 
> pipe.loop);     //Регистрируем message handler
>   gst_object_unref(bus);
>   pipe.element.video_src = gst_element_factory_make("v4l2src", 
> "videosrc");
>   g_object_set(GST_OBJECT(pipe.element.video_src), "device", 
> get_isp_out_dev_name(), NULL);
>   pipe.element.sink = gst_element_factory_make("udpsink", 
> "videoudpsink");
> }
>
> int create_pipe_elements_video(void)
> {
>   pipe.element.video_prep_buffer = 
> gst_element_factory_make("TIPrepEncBuf", "TIPrepEncBuf");
>   pipe.element.video_encoder = gst_element_factory_make("TIVidenc1", 
> "TIVidenc1");
>   pipe.element.rtph264pay = gst_element_factory_make("rtph264pay", 
> "rtph264pay");
>   if(!pipe.element.video_src || !pipe.element.video_prep_buffer || 
> !pipe.element.video_encoder ||\
>       !pipe.element.rtph264pay || !pipe.element.sink)
>     {
>       warnx(ERROR_GST_ELEMENT_CREATION);
>       return FALSE;
>     }
>   gst_bin_add_many(GST_BIN (pipe.element.pipeline), 
> pipe.element.video_src, \
>         pipe.element.video_prep_buffer, pipe.element.video_encoder, 
> pipe.element.rtph264pay, pipe.element.sink, NULL);
>   if(TRUE != gst_element_link_many(pipe.element.video_src, \
>       pipe.element.video_prep_buffer, pipe.element.video_encoder, 
> pipe.element.rtph264pay, pipe.element.sink, NULL))
>     {
>       warnx("Can not link video branch.");
>       return FALSE;
>     }
>   g_object_set(GST_OBJECT(pipe.element.video_prep_buffer), 
> "numOutputBufs", 1, "contiguousInputFrame", FALSE, NULL);
>   g_object_set(GST_OBJECT(pipe.element.video_encoder), "engineName", 
> "codecServer", "codecName", "h264enc",
>       "contiguousInputFrame", TRUE, "rateControlPreset", 
> IVIDEO_LOW_DELAY, NULL);
>   g_object_set(GST_OBJECT(pipe.element.rtph264pay), "pt", 96, NULL);
>   return TRUE;
> }
>
>
> on resiver:
>
> "udpsrc,rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink" 
>
>
>
> data->video_pipeline = gst_pipeline_new ("videopipeline");
> gst_bin_add_many(GST_BIN(data->video_pipeline),udpsrc,rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink,NULL); 
>
>
>   if (!gst_element_link_filtered (udpsrc,rtph264depay,udpsrc_caps)){
>       GST_ERROR ("Can't link udpsrc and rtph264depay with caps");
>   }
>
>   if (!gst_element_link_many 
> (rtph264depay,avdec_h264,/*videorate,*/clockoverlay,tee,/*queue,*/autovideosink,NULL)){
>       GST_ERROR ("Can't link many to tee");
>   }
>  /* if (!gst_element_link_many (queue,autovideosink,NULL)){
>        GST_ERROR ("Can't link queue and videosink");
>    }*/
>
>     gst_object_unref (G_OBJECT(videorate_src_pad));//Возможно 
> получится ошибка с пямятью
>     gst_caps_unref(videorate_caps);///Освобождаем caps
>
> this two pipelines impliments in applications for DM3730 and Android
>
> How can I realise
>
> latency reporting ?
>



More information about the gstreamer-devel mailing list