@@ -1669,32 +1669,32 @@ static void purge_fragmented_blocks_allcpus(void);
1669
1669
static bool __purge_vmap_area_lazy (unsigned long start , unsigned long end )
1670
1670
{
1671
1671
unsigned long resched_threshold ;
1672
- struct list_head local_pure_list ;
1672
+ struct list_head local_purge_list ;
1673
1673
struct vmap_area * va , * n_va ;
1674
1674
1675
1675
lockdep_assert_held (& vmap_purge_lock );
1676
1676
1677
1677
spin_lock (& purge_vmap_area_lock );
1678
1678
purge_vmap_area_root = RB_ROOT ;
1679
- list_replace_init (& purge_vmap_area_list , & local_pure_list );
1679
+ list_replace_init (& purge_vmap_area_list , & local_purge_list );
1680
1680
spin_unlock (& purge_vmap_area_lock );
1681
1681
1682
- if (unlikely (list_empty (& local_pure_list )))
1682
+ if (unlikely (list_empty (& local_purge_list )))
1683
1683
return false;
1684
1684
1685
1685
start = min (start ,
1686
- list_first_entry (& local_pure_list ,
1686
+ list_first_entry (& local_purge_list ,
1687
1687
struct vmap_area , list )-> va_start );
1688
1688
1689
1689
end = max (end ,
1690
- list_last_entry (& local_pure_list ,
1690
+ list_last_entry (& local_purge_list ,
1691
1691
struct vmap_area , list )-> va_end );
1692
1692
1693
1693
flush_tlb_kernel_range (start , end );
1694
1694
resched_threshold = lazy_max_pages () << 1 ;
1695
1695
1696
1696
spin_lock (& free_vmap_area_lock );
1697
- list_for_each_entry_safe (va , n_va , & local_pure_list , list ) {
1697
+ list_for_each_entry_safe (va , n_va , & local_purge_list , list ) {
1698
1698
unsigned long nr = (va -> va_end - va -> va_start ) >> PAGE_SHIFT ;
1699
1699
unsigned long orig_start = va -> va_start ;
1700
1700
unsigned long orig_end = va -> va_end ;
0 commit comments