[Spice-devel] [spice-gtk v1 09/11] channel-display: don't debug latency for each frame
Frediano Ziglio
fziglio at redhat.com
Tue Apr 17 11:46:42 UTC 2018
>
> From: Victor Toso <me at victortoso.com>
>
> Becomes quite hard to find meaning on something that is printed every
> time. Only print latency value if it is a new min/max or if average
> latency is 10% bigger/lower then usual.
>
> Not aiming to perfect statistics in latency here, just to avoid too
> verbose logging. Removing latency debug isn't cool as we could infer
> issues with streaming based on it.
>
> Signed-off-by: Victor Toso <victortoso at redhat.com>
> ---
> src/channel-display-priv.h | 3 +++
> src/channel-display.c | 18 +++++++++++++++++-
> 2 files changed, 20 insertions(+), 1 deletion(-)
>
> diff --git a/src/channel-display-priv.h b/src/channel-display-priv.h
> index 95ad7d8..e7758cc 100644
> --- a/src/channel-display-priv.h
> +++ b/src/channel-display-priv.h
> @@ -136,6 +136,9 @@ struct display_stream {
> drops_sequence_stats cur_drops_seq_stats;
> GArray *drops_seqs_stats_arr;
> uint32_t num_drops_seqs;
> + uint32_t latency_min;
> + uint32_t latency_max;
> + uint32_t latency_avg;
>
In the style documentation is explicitly state that we should not
column align.
> uint32_t playback_sync_drops_seq_len;
>
> diff --git a/src/channel-display.c b/src/channel-display.c
> index 4757aa6..3901cd1 100644
> --- a/src/channel-display.c
> +++ b/src/channel-display.c
> @@ -1547,6 +1547,10 @@ static void display_stream_stats_save(display_stream
> *st,
> guint32 client_mmtime)
> {
> gint32 latency = server_mmtime - client_mmtime;
> + gint32 min_latency = st->latency_min == 0 ? latency :
> MIN(st->latency_min, latency);
why not initializing latency_min with INT32_MAX?
> + gint32 max_latency = MAX(st->latency_max, latency);
as latency can be <0 latency_max should be initialized to INT32_MIN, not 0.
> + gdouble avg_latency = (st->latency_avg * st->num_input_frames + latency)
> /
> + ((double) (st->num_input_frames + 1));
>
I would use a latency_total in the display_stream structure. I think int64_t is
safe.
> if (!st->num_input_frames) {
> st->first_frame_mm_time = server_mmtime;
> @@ -1567,7 +1571,19 @@ static void display_stream_stats_save(display_stream
> *st,
> return;
> }
>
> - CHANNEL_DEBUG(st->channel, "video latency: %d", latency);
> + /* Only debug latency value if it matters otherwise it can be too
> verbose */
> + if (min_latency != st->latency_min ||
> + max_latency != st->latency_max ||
> + avg_latency < 0.90 * st->latency_avg ||
> + avg_latency > 1.10 * st->latency_avg) {
> + CHANNEL_DEBUG(st->channel,
> + "video latency: %d | (%d , %0.2f , %d)",
> + latency, min_latency, avg_latency, max_latency);
> + st->latency_min = min_latency;
> + st->latency_max = max_latency;
> + }
> + st->latency_avg = avg_latency;
> +
> if (st->cur_drops_seq_stats.len) {
> st->cur_drops_seq_stats.duration = server_mmtime -
> st->cur_drops_seq_stats.start_mm_time;
Frediano
More information about the Spice-devel
mailing list