@@ -1669,32 +1669,32 @@ static void purge_fragmented_blocks_allcpus(void);
16691669static bool __purge_vmap_area_lazy (unsigned long start , unsigned long end )
16701670{
16711671 unsigned long resched_threshold ;
1672- struct list_head local_pure_list ;
1672+ struct list_head local_purge_list ;
16731673 struct vmap_area * va , * n_va ;
16741674
16751675 lockdep_assert_held (& vmap_purge_lock );
16761676
16771677 spin_lock (& purge_vmap_area_lock );
16781678 purge_vmap_area_root = RB_ROOT ;
1679- list_replace_init (& purge_vmap_area_list , & local_pure_list );
1679+ list_replace_init (& purge_vmap_area_list , & local_purge_list );
16801680 spin_unlock (& purge_vmap_area_lock );
16811681
1682- if (unlikely (list_empty (& local_pure_list )))
1682+ if (unlikely (list_empty (& local_purge_list )))
16831683 return false;
16841684
16851685 start = min (start ,
1686- list_first_entry (& local_pure_list ,
1686+ list_first_entry (& local_purge_list ,
16871687 struct vmap_area , list )-> va_start );
16881688
16891689 end = max (end ,
1690- list_last_entry (& local_pure_list ,
1690+ list_last_entry (& local_purge_list ,
16911691 struct vmap_area , list )-> va_end );
16921692
16931693 flush_tlb_kernel_range (start , end );
16941694 resched_threshold = lazy_max_pages () << 1 ;
16951695
16961696 spin_lock (& free_vmap_area_lock );
1697- list_for_each_entry_safe (va , n_va , & local_pure_list , list ) {
1697+ list_for_each_entry_safe (va , n_va , & local_purge_list , list ) {
16981698 unsigned long nr = (va -> va_end - va -> va_start ) >> PAGE_SHIFT ;
16991699 unsigned long orig_start = va -> va_start ;
17001700 unsigned long orig_end = va -> va_end ;
0 commit comments