[PATCH] erofs: fix use-after-free of on-stack io[]

Henry King hongyu.jin.cn at gmail.com
Fri Apr 1 20:36:23 AEDT 2022


Gao Xiang <hsiangkao at linux.alibaba.com> 于2022年4月1日周五 14:55写道:

>
> On Fri, Apr 01, 2022 at 02:33:01PM +0800, Hongyu Jin wrote:
> > From: Hongyu Jin <hongyu.jin at unisoc.com>
> >
> > The root cause is the race as follows(k5.4):
> > Thread #1                               Thread #2(irq ctx)
> >
> > z_erofs_submit_and_unzip()
> >   struct z_erofs_vle_unzip_io io_A[]
> >   submit bio A
> >                                         z_erofs_vle_read_endio() // bio A
> >                                         z_erofs_vle_unzip_kickoff()
> >                                         spin_lock_irqsave()
> >                                         atomic_add_return()
> >   wait_event()
> >   [end of function]
> > z_erofs_submit_and_unzip() // bio B
> >                                         wake_up_locked(io_A[]) // crash
> >   struct z_erofs_vle_unzip_io io_B[]
> >   submit bio B
> >   wait_event()
>
> Thanks, good catch!
> Yet could you turn the race above into the current function names?
ok, I will chang it.
>
> >
> > Backtrace in kernel5.4:
> > [   10.129413] 8<--- cut here ---
> > [   10.129422] Unable to handle kernel paging request at virtual address eb0454a4
> > [   10.364157] CPU: 0 PID: 709 Comm: getprop Tainted: G        WC O      5.4.147-ab09225 #1
> > [   11.556325] [<c01b33b8>] (__wake_up_common) from [<c01b3300>] (__wake_up_locked+0x40/0x48)
> > [   11.565487] [<c01b3300>] (__wake_up_locked) from [<c044c8d0>] (z_erofs_vle_unzip_kickoff+0x6c/0xc0)
> > [   11.575438] [<c044c8d0>] (z_erofs_vle_unzip_kickoff) from [<c044c854>] (z_erofs_vle_read_endio+0x16c/0x17c)
> > [   11.586082] [<c044c854>] (z_erofs_vle_read_endio) from [<c06a80e8>] (clone_endio+0xb4/0x1d0)
> > [   11.595428] [<c06a80e8>] (clone_endio) from [<c04a1280>] (blk_update_request+0x150/0x4dc)
> > [   11.604516] [<c04a1280>] (blk_update_request) from [<c06dea28>] (mmc_blk_cqe_complete_rq+0x144/0x15c)
> > [   11.614640] [<c06dea28>] (mmc_blk_cqe_complete_rq) from [<c04a5d90>] (blk_done_softirq+0xb0/0xcc)
> > [   11.624419] [<c04a5d90>] (blk_done_softirq) from [<c010242c>] (__do_softirq+0x184/0x56c)
> > [   11.633419] [<c010242c>] (__do_softirq) from [<c01051e8>] (irq_exit+0xd4/0x138)
> > [   11.641640] [<c01051e8>] (irq_exit) from [<c010c314>] (__handle_domain_irq+0x94/0xd0)
> > [   11.650381] [<c010c314>] (__handle_domain_irq) from [<c04fde70>] (gic_handle_irq+0x50/0xd4)
> > [   11.659641] [<c04fde70>] (gic_handle_irq) from [<c0101b70>] (__irq_svc+0x70/0xb0)
> >
> > Signed-off-by: Hongyu Jin <hongyu.jin at unisoc.com>
> > ---
> >  fs/erofs/zdata.c | 12 ++++--------
> >  fs/erofs/zdata.h |  2 +-
> >  2 files changed, 5 insertions(+), 9 deletions(-)
> >
> > diff --git a/fs/erofs/zdata.c b/fs/erofs/zdata.c
> > index 11c7a1aaebad..4c26faa817a3 100644
> > --- a/fs/erofs/zdata.c
> > +++ b/fs/erofs/zdata.c
> > @@ -782,12 +782,9 @@ static void z_erofs_decompress_kickoff(struct z_erofs_decompressqueue *io,
> >
> >       /* wake up the caller thread for sync decompression */
> >       if (sync) {
> > -             unsigned long flags;
> > -
> > -             spin_lock_irqsave(&io->u.wait.lock, flags);
> >               if (!atomic_add_return(bios, &io->pending_bios))
> > -                     wake_up_locked(&io->u.wait);
> > -             spin_unlock_irqrestore(&io->u.wait.lock, flags);
> > +                     complete(&io->u.done);
> > +
> >               return;
> >       }
> >
> > @@ -1207,7 +1204,7 @@ jobqueue_init(struct super_block *sb,
> >       } else {
> >  fg_out:
> >               q = fgq;
> > -             init_waitqueue_head(&fgq->u.wait);
> > +             init_completion(&fgq->u.done);
> >               atomic_set(&fgq->pending_bios, 0);
> >       }
> >       q->sb = sb;
> > @@ -1370,8 +1367,7 @@ static void z_erofs_runqueue(struct super_block *sb,
> >               return;
> >
> >       /* wait until all bios are completed */
> > -     io_wait_event(io[JQ_SUBMIT].u.wait,
> > -                   !atomic_read(&io[JQ_SUBMIT].pending_bios));
> > +     wait_for_completion_io(&io[JQ_SUBMIT].u.done);
>
> Thanks, good catch!
>
> What if pending_bios is always 0 (nr_bios == 0), is it possible?
The pending_bios isn't always 0.  If bio is completed faster before
io_wait_event() called, the value of pending_bios from 1 to 0,
when enter io_wait_event(), it will not acquire lock and return immediately.
>
> Thanks,
> Gao Xiang


More information about the Linux-erofs mailing list