Skip to content

Commit b6ea95a

Browse files
Alexander Gordeevakpm00
authored andcommitted
kasan: avoid sleepable page allocation from atomic context
apply_to_pte_range() enters the lazy MMU mode and then invokes kasan_populate_vmalloc_pte() callback on each page table walk iteration. However, the callback can go into sleep when trying to allocate a single page, e.g. if an architecutre disables preemption on lazy MMU mode enter. On s390 if make arch_enter_lazy_mmu_mode() -> preempt_enable() and arch_leave_lazy_mmu_mode() -> preempt_disable(), such crash occurs: [ 0.663336] BUG: sleeping function called from invalid context at ./include/linux/sched/mm.h:321 [ 0.663348] in_atomic(): 1, irqs_disabled(): 0, non_block: 0, pid: 2, name: kthreadd [ 0.663358] preempt_count: 1, expected: 0 [ 0.663366] RCU nest depth: 0, expected: 0 [ 0.663375] no locks held by kthreadd/2. [ 0.663383] Preemption disabled at: [ 0.663386] [<0002f3284cbb4eda>] apply_to_pte_range+0xfa/0x4a0 [ 0.663405] CPU: 0 UID: 0 PID: 2 Comm: kthreadd Not tainted 6.15.0-rc5-gcc-kasan-00043-gd76bb1ebb558-dirty #162 PREEMPT [ 0.663408] Hardware name: IBM 3931 A01 701 (KVM/Linux) [ 0.663409] Call Trace: [ 0.663410] [<0002f3284c385f58>] dump_stack_lvl+0xe8/0x140 [ 0.663413] [<0002f3284c507b9e>] __might_resched+0x66e/0x700 [ 0.663415] [<0002f3284cc4f6c0>] __alloc_frozen_pages_noprof+0x370/0x4b0 [ 0.663419] [<0002f3284ccc73c0>] alloc_pages_mpol+0x1a0/0x4a0 [ 0.663421] [<0002f3284ccc8518>] alloc_frozen_pages_noprof+0x88/0xc0 [ 0.663424] [<0002f3284ccc8572>] alloc_pages_noprof+0x22/0x120 [ 0.663427] [<0002f3284cc341ac>] get_free_pages_noprof+0x2c/0xc0 [ 0.663429] [<0002f3284cceba70>] kasan_populate_vmalloc_pte+0x50/0x120 [ 0.663433] [<0002f3284cbb4ef8>] apply_to_pte_range+0x118/0x4a0 [ 0.663435] [<0002f3284cbc7c14>] apply_to_pmd_range+0x194/0x3e0 [ 0.663437] [<0002f3284cbc99be>] __apply_to_page_range+0x2fe/0x7a0 [ 0.663440] [<0002f3284cbc9e88>] apply_to_page_range+0x28/0x40 [ 0.663442] [<0002f3284ccebf12>] kasan_populate_vmalloc+0x82/0xa0 [ 0.663445] [<0002f3284cc1578c>] alloc_vmap_area+0x34c/0xc10 [ 0.663448] [<0002f3284cc1c2a6>] __get_vm_area_node+0x186/0x2a0 [ 0.663451] [<0002f3284cc1e696>] __vmalloc_node_range_noprof+0x116/0x310 [ 0.663454] [<0002f3284cc1d950>] __vmalloc_node_noprof+0xd0/0x110 [ 0.663457] [<0002f3284c454b88>] alloc_thread_stack_node+0xf8/0x330 [ 0.663460] [<0002f3284c458d56>] dup_task_struct+0x66/0x4d0 [ 0.663463] [<0002f3284c45be90>] copy_process+0x280/0x4b90 [ 0.663465] [<0002f3284c460940>] kernel_clone+0xd0/0x4b0 [ 0.663467] [<0002f3284c46115e>] kernel_thread+0xbe/0xe0 [ 0.663469] [<0002f3284c4e440e>] kthreadd+0x50e/0x7f0 [ 0.663472] [<0002f3284c38c04a>] __ret_from_fork+0x8a/0xf0 [ 0.663475] [<0002f3284ed57ff2>] ret_from_fork+0xa/0x38 Instead of allocating single pages per-PTE, bulk-allocate the shadow memory prior to applying kasan_populate_vmalloc_pte() callback on a page range. Link: https://lkml.kernel.org/r/c61d3560297c93ed044f0b1af085610353a06a58.1747316918.git.agordeev@linux.ibm.com Fixes: 3c5c3cf ("kasan: support backing vmalloc space with real shadow memory") Signed-off-by: Alexander Gordeev <[email protected]> Suggested-by: Andrey Ryabinin <[email protected]> Reviewed-by: Harry Yoo <[email protected]> Cc: Daniel Axtens <[email protected]> Cc: <[email protected]> Signed-off-by: Andrew Morton <[email protected]>
1 parent 97dfbbd commit b6ea95a

File tree

1 file changed

+78
-14
lines changed

1 file changed

+78
-14
lines changed

mm/kasan/shadow.c

Lines changed: 78 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -292,33 +292,99 @@ void __init __weak kasan_populate_early_vm_area_shadow(void *start,
292292
{
293293
}
294294

295+
struct vmalloc_populate_data {
296+
unsigned long start;
297+
struct page **pages;
298+
};
299+
295300
static int kasan_populate_vmalloc_pte(pte_t *ptep, unsigned long addr,
296-
void *unused)
301+
void *_data)
297302
{
298-
unsigned long page;
303+
struct vmalloc_populate_data *data = _data;
304+
struct page *page;
299305
pte_t pte;
306+
int index;
300307

301308
if (likely(!pte_none(ptep_get(ptep))))
302309
return 0;
303310

304-
page = __get_free_page(GFP_KERNEL);
305-
if (!page)
306-
return -ENOMEM;
307-
308-
__memset((void *)page, KASAN_VMALLOC_INVALID, PAGE_SIZE);
309-
pte = pfn_pte(PFN_DOWN(__pa(page)), PAGE_KERNEL);
311+
index = PFN_DOWN(addr - data->start);
312+
page = data->pages[index];
313+
__memset(page_to_virt(page), KASAN_VMALLOC_INVALID, PAGE_SIZE);
314+
pte = pfn_pte(page_to_pfn(page), PAGE_KERNEL);
310315

311316
spin_lock(&init_mm.page_table_lock);
312317
if (likely(pte_none(ptep_get(ptep)))) {
313318
set_pte_at(&init_mm, addr, ptep, pte);
314-
page = 0;
319+
data->pages[index] = NULL;
315320
}
316321
spin_unlock(&init_mm.page_table_lock);
317-
if (page)
318-
free_page(page);
322+
323+
return 0;
324+
}
325+
326+
static void ___free_pages_bulk(struct page **pages, int nr_pages)
327+
{
328+
int i;
329+
330+
for (i = 0; i < nr_pages; i++) {
331+
if (pages[i]) {
332+
__free_pages(pages[i], 0);
333+
pages[i] = NULL;
334+
}
335+
}
336+
}
337+
338+
static int ___alloc_pages_bulk(struct page **pages, int nr_pages)
339+
{
340+
unsigned long nr_populated, nr_total = nr_pages;
341+
struct page **page_array = pages;
342+
343+
while (nr_pages) {
344+
nr_populated = alloc_pages_bulk(GFP_KERNEL, nr_pages, pages);
345+
if (!nr_populated) {
346+
___free_pages_bulk(page_array, nr_total - nr_pages);
347+
return -ENOMEM;
348+
}
349+
pages += nr_populated;
350+
nr_pages -= nr_populated;
351+
}
352+
319353
return 0;
320354
}
321355

356+
static int __kasan_populate_vmalloc(unsigned long start, unsigned long end)
357+
{
358+
unsigned long nr_pages, nr_total = PFN_UP(end - start);
359+
struct vmalloc_populate_data data;
360+
int ret = 0;
361+
362+
data.pages = (struct page **)__get_free_page(GFP_KERNEL | __GFP_ZERO);
363+
if (!data.pages)
364+
return -ENOMEM;
365+
366+
while (nr_total) {
367+
nr_pages = min(nr_total, PAGE_SIZE / sizeof(data.pages[0]));
368+
ret = ___alloc_pages_bulk(data.pages, nr_pages);
369+
if (ret)
370+
break;
371+
372+
data.start = start;
373+
ret = apply_to_page_range(&init_mm, start, nr_pages * PAGE_SIZE,
374+
kasan_populate_vmalloc_pte, &data);
375+
___free_pages_bulk(data.pages, nr_pages);
376+
if (ret)
377+
break;
378+
379+
start += nr_pages * PAGE_SIZE;
380+
nr_total -= nr_pages;
381+
}
382+
383+
free_page((unsigned long)data.pages);
384+
385+
return ret;
386+
}
387+
322388
int kasan_populate_vmalloc(unsigned long addr, unsigned long size)
323389
{
324390
unsigned long shadow_start, shadow_end;
@@ -348,9 +414,7 @@ int kasan_populate_vmalloc(unsigned long addr, unsigned long size)
348414
shadow_start = PAGE_ALIGN_DOWN(shadow_start);
349415
shadow_end = PAGE_ALIGN(shadow_end);
350416

351-
ret = apply_to_page_range(&init_mm, shadow_start,
352-
shadow_end - shadow_start,
353-
kasan_populate_vmalloc_pte, NULL);
417+
ret = __kasan_populate_vmalloc(shadow_start, shadow_end);
354418
if (ret)
355419
return ret;
356420

0 commit comments

Comments
 (0)