x264enc and variable rate appsrc
Gary Metalle
Gary.Metalle at rvl.co.uk
Tue Sep 22 11:27:16 UTC 2020
Hi
What happens if you use a framerate of 0/1? This is the way to define a variable framerate.
Maybe the issue is with your decoder pipeline and the not the encoding. Have you tried setting 'sync=FALSE' for the video sink you're using?
________________________________
From: gstreamer-devel <gstreamer-devel-bounces at lists.freedesktop.org> on behalf of David J <davywj at gmail.com>
Sent: 22 September 2020 12:21
To: Discussion of the development of and with GStreamer
Subject: x264enc and variable rate appsrc
I have created a test pipeline (see code below) using a custom appsrc that creates an h264 stream and sends over rtp.
I am testing it with the following command line to receive and play the video stream:
gst-launch-1.0 -v udpsrc port=5000 caps = "application/x-rtp, media=(string)video, clock-rate=(int)90000, encoding-name=(string)H264, payload=(int)96" ! rtph264depay ! decodebin ! videoconvert ! autovideosink
The command-line inside the application looks like this:
appsrc name=MyAppSrc ! video/x-raw,width=640,height=360,framerate=10/1 ! videoconvert ! identity name=ident ! x264enc zerolatency=true speed-preset=superfast ! rtph264pay ! udpsink host=127.0.0.1 port=5000
At the moment the appsrc is generating frames at precisely 10fps and I set the buffer timestamps correctly. However, unless I also specify the framerate (in the caps string following the appsrc) as 10/1 the client viewer constantly complains of:
WARNING: from element /GstPipeline:pipeline0/GstAutoVideoSink:autovideosink0/GstD3D11VideoSinkBin:autovideosink0-actual-sink-d3d11video/GstD3D11VideoSink:d3d11videosink0: A lot of buffers are being dropped.
Additional debug info:
../libs/gst/base/gstbasesink.c(3134): gst_base_sink_is_too_late (): /GstPipeline:pipeline0/GstAutoVideoSink:autovideosink0/GstD3D11VideoSinkBin:autovideosink0-actual-sink-d3d11video/GstD3D11VideoSink:d3d11videosink0:
There may be a timestamping problem, or this computer is too slow.
The problem is:
In the real application, the appsrc will produce frames at a variable rate - between about 2 and 20 fps. How can I get this to work if I'm required to set a fixed framerate?
Provided the buffer timestamps are set, why does the framerate property have any bearing on the pipeine?
Here is the application code:
// NativeBasicPipeline.cpp : This file contains the 'main' function. Program execution begins and ends there.
//
#include <gst/gst.h>
#include <gst/app/gstappsrc.h>
GTimer* timer;
GstAppSrc* appSrc;
guint sourceid = 0;
int width = 1;
int colour = 255;
bool read_data()
{
static GstClockTime time = 0;
static guint64 offset = 0;
GstFlowReturn ret;
auto ms = g_timer_elapsed(timer, NULL);
if (ms > 1.0 / 10.0) {
auto size = 640 * 3 * 360;
gpointer copy = g_malloc(size);
memset(copy, colour % 255, size);
colour += 10;
GstBuffer* buffer = gst_buffer_new_wrapped(copy, size);
//GST_BUFFER_TIMESTAMP(buffer) = pts;
GST_BUFFER_PTS(buffer) = time;
GST_BUFFER_DTS(buffer) = time;
GST_BUFFER_DURATION(buffer) = 100 * GST_MSECOND;
GST_BUFFER_OFFSET(buffer) = offset++;
GST_BUFFER_OFFSET_END(buffer) = offset;
time += 100 * GST_MSECOND;
g_signal_emit_by_name(appSrc, "push-buffer", buffer, &ret);
//ret = gst_app_src_push_buffer(appSrc, buffer);
gst_buffer_unref(buffer);
g_timer_start(timer);
return TRUE;
}
return TRUE;
}
void start_feed(GstElement* pipeline, guint size, void* unused)
{
if (sourceid == 0) {
sourceid = g_idle_add((GSourceFunc)read_data, 0);
}
}
void stop_feed(GstElement* pipeline, void* unused)
{
if (sourceid != 0) {
g_source_remove(sourceid);
sourceid = 0;
}
}
static gboolean print_field(GQuark field, const GValue* value, gpointer pfx) {
gchar* str = gst_value_serialize(value);
g_print("%s %15s: %s\n", (gchar*)pfx, g_quark_to_string(field), str);
g_free(str);
return TRUE;
}
static void print_caps(const GstCaps* caps, const gchar* pfx) {
guint i;
g_return_if_fail(caps != NULL);
if (gst_caps_is_any(caps)) {
g_print("%sANY\n", pfx);
return;
}
if (gst_caps_is_empty(caps)) {
g_print("%sEMPTY\n", pfx);
return;
}
for (i = 0; i < gst_caps_get_size(caps); i++) {
GstStructure* structure = gst_caps_get_structure(caps, i);
g_print("%s%s\n", pfx, gst_structure_get_name(structure));
gst_structure_foreach(structure, print_field, (gpointer)pfx);
}
}
void handoff(GstElement* pipeline, GstBuffer* buffer, void* unused)
{
static int count = 0;
static GstBuffer* buffers[50];
//auto ident = gst_bin_get_by_name(GST_BIN(pipeline), "ident");
auto pads = GST_ELEMENT_PADS(pipeline);
auto pad0 = GST_PAD(pads->data);
auto pad1 = GST_PAD(pads->next->data);
auto caps = gst_pad_get_current_caps(pad1);
print_caps(caps, "");
if (count < 50)
{
GstBuffer* copy = gst_buffer_copy(buffer);
buffers[count] = copy;
++count;
}
else
{
count = 100;
}
}
int main()
{
GstElement* pipeline;
GstBus* bus;
GstMessage* msg;
timer = g_timer_new();
/* Initialize GStreamer */
gst_init(0, 0);
/* Build the pipeline */
GError* err = 0;
//auto udpPipe = "videotestsrc pattern=smpte ! video/x-raw,width=640,height=360,framerate=10/1 ! videoscale ! videoconvert ! identity name=ident ! x264enc zerolatency=true speed-preset=superfast ! rtph264pay ! udpsink host=127.0.0.1 port=5000";
//auto videoPipe = "videotestsrc pattern=smpte ! video/x-raw,width=640,height=360,framerate=10/1 ! videoscale ! videoconvert ! autovideosink";
auto appSrcPipe = "appsrc name=MyAppSrc ! video/x-raw,width=640,height=360 ! videoconvert ! identity name=ident ! x264enc zerolatency=true speed-preset=superfast ! rtph264pay ! udpsink host=127.0.0.1 port=5000";
//auto appSrcPipeVideo = "appsrc name=MyAppSrc ! videoscale ! videoconvert ! autovideosink";
pipeline =
gst_parse_launch
(appSrcPipe,
&err);
appSrc = (GstAppSrc*)gst_bin_get_by_name(GST_BIN(pipeline), "MyAppSrc");
g_object_set(G_OBJECT(appSrc), "format", GST_FORMAT_TIME, NULL);
g_signal_connect(appSrc, "need-data", G_CALLBACK(start_feed), 0);
g_signal_connect(appSrc, "enough-data", G_CALLBACK(stop_feed), 0);
auto caps = gst_caps_new_simple("video/x-raw",
"format", G_TYPE_STRING, "RGB",
"bpp", G_TYPE_INT, 24,
"depth", G_TYPE_INT, 24,
"width", G_TYPE_INT, 640,
"height", G_TYPE_INT, 360,
NULL);
gst_app_src_set_caps((GstAppSrc*)appSrc, caps);
//GstPad* pad = gst_element_get_static_pad(appSrc, "src");
//auto ident = gst_bin_get_by_name(GST_BIN(pipeline), "ident");
//g_signal_connect(ident, "handoff", G_CALLBACK(handoff), 0);
/* Start playing */
gst_element_set_state(pipeline, GST_STATE_PLAYING);
auto main_loop = g_main_loop_new(NULL, FALSE);
g_main_loop_run(main_loop);
/* Wait until error or EOS */
bus = gst_element_get_bus(pipeline);
msg =
gst_bus_timed_pop_filtered(bus, GST_CLOCK_TIME_NONE,
(GstMessageType)(GST_MESSAGE_ERROR | GST_MESSAGE_EOS));
/* Free resources */
if (msg != NULL)
gst_message_unref(msg);
gst_object_unref(bus);
gst_element_set_state(pipeline, GST_STATE_NULL);
gst_object_unref(pipeline);
return 0;
}
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.freedesktop.org/archives/gstreamer-devel/attachments/20200922/ac6933fd/attachment.htm>
More information about the gstreamer-devel
mailing list