ast: cursor flashing softlockups

Ming Lei ming.lei at canonical.com
Tue May 17 09:12:04 UTC 2016


Hi,

On Tue, May 17, 2016 at 4:07 AM, Dann Frazier
<dann.frazier at canonical.com> wrote:
> Hi,
>  I'm observing a soft lockup issue w/ the ASPEED controller on an
> arm64 server platform. This was originally seen on Ubuntu's 4.4
> kernel, but it is reproducible w/ vanilla 4.6-rc7 as well.
>
> [   32.792656] NMI watchdog: BUG: soft lockup - CPU#38 stuck for 22s!
> [swapper/38:0]
>
> I observe this just once each time I boot into debian-installer (I'm
> using a serial console, but the ast module gets loaded during
> startup).

I have figured out that it is caused by 'mod_timer(timer, jiffies)' and
'ops->cur_blink_jiffies' is observed as zero in cursor_timer_handler()
when the issue happened.

Looks it is a real fbcon/vt issue, see following:

fbcon_init()
    <-.con_init
          <-visual_init()

reset_terminal()
    <-vc_init()

vc->vc_cur_blink_ms is just set in reset_terminal() from vc_init() path,
and ops->cur_blink_jiffies is figured out from vc->vc_cur_blink_ms
in fbcon_init().

And visual_init() is always run before vc_init(), so ops->cur_blink_jiffies
is initialized as zero and cause the soft lockup issue finally.

Thanks,
Ming

>
> perf shows that the CPU caught by the NMI is typically in code
> updating the cursor timer:
>
> -   16.92%  swapper  [kernel.kallsyms]      [k] _raw_spin_unlock_irqrestore
>    - _raw_spin_unlock_irqrestore
>       + 16.87% mod_timer
>       + 0.05% cursor_timer_handler
> -   12.15%  swapper  [kernel.kallsyms]      [k] queue_work_on
>    - queue_work_on
>       + 12.00% cursor_timer_handler
>       + 0.15% call_timer_fn
> +   10.98%  swapper  [kernel.kallsyms]      [k] run_timer_softirq
> -    2.23%  swapper  [kernel.kallsyms]      [k] mod_timer
>    - mod_timer
>       + 1.97% cursor_timer_handler
>       + 0.26% call_timer_fn
>
> During the same period, I can see that another CPU is actively
> executing the timer function:
>
> -   42.18%  kworker/u96:2  [kernel.kallsyms]  [k] ww_mutex_unlock
>    - ww_mutex_unlock
>       - 40.70% ast_dirty_update
>            ast_imageblit
>            soft_cursor
>            bit_cursor
>            fb_flashcursor
>            process_one_work
>            worker_thread
>            kthread
>            ret_from_fork
>       + 1.48% ast_imageblit
> -   40.15%  kworker/u96:2  [kernel.kallsyms]  [k] __memcpy_toio
>    - __memcpy_toio
>       + 31.54% ast_dirty_update
>       + 8.61% ast_imageblit
>
> Using the graph function tracer on fb_flashcursor(), I see that
> ast_dirty_update usually takes around 60 us, in which it makes 16
> calls to __memcpy_toio(). However, there is always one instance on
> every boot of the installer where ast_dirty_update() takes ~98 *ms* to
> complete, during which it makes 743 calls to __memcpy_toio(). While
> that  doesn't directly account for the full 22s, I do wonder if that
> maybe a smoking gun.
>
> fyi, this is being tracked at: https://bugs.launchpad.net/bugs/1574814
>
>   -dann


More information about the dri-devel mailing list