@@ -2897,12 +2897,11 @@ static int __do_readpage(struct extent_io_tree *tree,
2897
2897
struct block_device * bdev ;
2898
2898
int ret ;
2899
2899
int nr = 0 ;
2900
- int parent_locked = * bio_flags & EXTENT_BIO_PARENT_LOCKED ;
2901
2900
size_t pg_offset = 0 ;
2902
2901
size_t iosize ;
2903
2902
size_t disk_io_size ;
2904
2903
size_t blocksize = inode -> i_sb -> s_blocksize ;
2905
- unsigned long this_bio_flag = * bio_flags & EXTENT_BIO_PARENT_LOCKED ;
2904
+ unsigned long this_bio_flag = 0 ;
2906
2905
2907
2906
set_page_extent_mapped (page );
2908
2907
@@ -2942,18 +2941,16 @@ static int __do_readpage(struct extent_io_tree *tree,
2942
2941
kunmap_atomic (userpage );
2943
2942
set_extent_uptodate (tree , cur , cur + iosize - 1 ,
2944
2943
& cached , GFP_NOFS );
2945
- if (!parent_locked )
2946
- unlock_extent_cached (tree , cur ,
2947
- cur + iosize - 1 ,
2948
- & cached , GFP_NOFS );
2944
+ unlock_extent_cached (tree , cur ,
2945
+ cur + iosize - 1 ,
2946
+ & cached , GFP_NOFS );
2949
2947
break ;
2950
2948
}
2951
2949
em = __get_extent_map (inode , page , pg_offset , cur ,
2952
2950
end - cur + 1 , get_extent , em_cached );
2953
2951
if (IS_ERR_OR_NULL (em )) {
2954
2952
SetPageError (page );
2955
- if (!parent_locked )
2956
- unlock_extent (tree , cur , end );
2953
+ unlock_extent (tree , cur , end );
2957
2954
break ;
2958
2955
}
2959
2956
extent_offset = cur - em -> start ;
@@ -3038,12 +3035,9 @@ static int __do_readpage(struct extent_io_tree *tree,
3038
3035
3039
3036
set_extent_uptodate (tree , cur , cur + iosize - 1 ,
3040
3037
& cached , GFP_NOFS );
3041
- if (parent_locked )
3042
- free_extent_state (cached );
3043
- else
3044
- unlock_extent_cached (tree , cur ,
3045
- cur + iosize - 1 ,
3046
- & cached , GFP_NOFS );
3038
+ unlock_extent_cached (tree , cur ,
3039
+ cur + iosize - 1 ,
3040
+ & cached , GFP_NOFS );
3047
3041
cur = cur + iosize ;
3048
3042
pg_offset += iosize ;
3049
3043
continue ;
@@ -3052,8 +3046,7 @@ static int __do_readpage(struct extent_io_tree *tree,
3052
3046
if (test_range_bit (tree , cur , cur_end ,
3053
3047
EXTENT_UPTODATE , 1 , NULL )) {
3054
3048
check_page_uptodate (tree , page );
3055
- if (!parent_locked )
3056
- unlock_extent (tree , cur , cur + iosize - 1 );
3049
+ unlock_extent (tree , cur , cur + iosize - 1 );
3057
3050
cur = cur + iosize ;
3058
3051
pg_offset += iosize ;
3059
3052
continue ;
@@ -3063,8 +3056,7 @@ static int __do_readpage(struct extent_io_tree *tree,
3063
3056
*/
3064
3057
if (block_start == EXTENT_MAP_INLINE ) {
3065
3058
SetPageError (page );
3066
- if (!parent_locked )
3067
- unlock_extent (tree , cur , cur + iosize - 1 );
3059
+ unlock_extent (tree , cur , cur + iosize - 1 );
3068
3060
cur = cur + iosize ;
3069
3061
pg_offset += iosize ;
3070
3062
continue ;
@@ -3083,8 +3075,7 @@ static int __do_readpage(struct extent_io_tree *tree,
3083
3075
* bio_flags = this_bio_flag ;
3084
3076
} else {
3085
3077
SetPageError (page );
3086
- if (!parent_locked )
3087
- unlock_extent (tree , cur , cur + iosize - 1 );
3078
+ unlock_extent (tree , cur , cur + iosize - 1 );
3088
3079
}
3089
3080
cur = cur + iosize ;
3090
3081
pg_offset += iosize ;
@@ -3213,20 +3204,6 @@ int extent_read_full_page(struct extent_io_tree *tree, struct page *page,
3213
3204
return ret ;
3214
3205
}
3215
3206
3216
- int extent_read_full_page_nolock (struct extent_io_tree * tree , struct page * page ,
3217
- get_extent_t * get_extent , int mirror_num )
3218
- {
3219
- struct bio * bio = NULL ;
3220
- unsigned long bio_flags = EXTENT_BIO_PARENT_LOCKED ;
3221
- int ret ;
3222
-
3223
- ret = __do_readpage (tree , page , get_extent , NULL , & bio , mirror_num ,
3224
- & bio_flags , READ , NULL );
3225
- if (bio )
3226
- ret = submit_one_bio (READ , bio , mirror_num , bio_flags );
3227
- return ret ;
3228
- }
3229
-
3230
3207
static noinline void update_nr_written (struct page * page ,
3231
3208
struct writeback_control * wbc ,
3232
3209
unsigned long nr_written )
0 commit comments