Skip to content

Commit dcc25ae

Browse files
htejunaxboe
authored andcommitted
writeback: move global_dirty_limit into wb_domain
This patch is a part of the series to define wb_domain which represents a domain that wb's (bdi_writeback's) belong to and are measured against each other in. This will enable IO backpressure propagation for cgroup writeback. global_dirty_limit exists to regulate the global dirty threshold which is a property of the wb_domain. This patch moves hard_dirty_limit, dirty_lock, and update_time into wb_domain. This is pure reorganization and doesn't introduce any behavioral changes. Signed-off-by: Tejun Heo <[email protected]> Cc: Jens Axboe <[email protected]> Cc: Jan Kara <[email protected]> Cc: Wu Fengguang <[email protected]> Cc: Greg Thelen <[email protected]> Signed-off-by: Jens Axboe <[email protected]>
1 parent 380c27c commit dcc25ae

File tree

4 files changed

+44
-28
lines changed

4 files changed

+44
-28
lines changed

fs/fs-writeback.c

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -887,7 +887,7 @@ static long writeback_chunk_size(struct bdi_writeback *wb,
887887
pages = LONG_MAX;
888888
else {
889889
pages = min(wb->avg_write_bandwidth / 2,
890-
global_dirty_limit / DIRTY_SCOPE);
890+
global_wb_domain.dirty_limit / DIRTY_SCOPE);
891891
pages = min(pages, work->nr_pages);
892892
pages = round_down(pages + MIN_WRITEBACK_PAGES,
893893
MIN_WRITEBACK_PAGES);

include/linux/writeback.h

Lines changed: 16 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -95,6 +95,8 @@ struct writeback_control {
9595
* dirtyable memory accordingly.
9696
*/
9797
struct wb_domain {
98+
spinlock_t lock;
99+
98100
/*
99101
* Scale the writeback cache size proportional to the relative
100102
* writeout speed.
@@ -115,6 +117,19 @@ struct wb_domain {
115117
struct fprop_global completions;
116118
struct timer_list period_timer; /* timer for aging of completions */
117119
unsigned long period_time;
120+
121+
/*
122+
* The dirtyable memory and dirty threshold could be suddenly
123+
* knocked down by a large amount (eg. on the startup of KVM in a
124+
* swapless system). This may throw the system into deep dirty
125+
* exceeded state and throttle heavy/light dirtiers alike. To
126+
* retain good responsiveness, maintain global_dirty_limit for
127+
* tracking slowly down to the knocked down dirty threshold.
128+
*
129+
* Both fields are protected by ->lock.
130+
*/
131+
unsigned long dirty_limit_tstamp;
132+
unsigned long dirty_limit;
118133
};
119134

120135
/*
@@ -153,7 +168,7 @@ void throttle_vm_writeout(gfp_t gfp_mask);
153168
bool zone_dirty_ok(struct zone *zone);
154169
int wb_domain_init(struct wb_domain *dom, gfp_t gfp);
155170

156-
extern unsigned long global_dirty_limit;
171+
extern struct wb_domain global_wb_domain;
157172

158173
/* These are exported to sysctl. */
159174
extern int dirty_background_ratio;

include/trace/events/writeback.h

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -361,7 +361,7 @@ TRACE_EVENT(global_dirty_state,
361361
__entry->nr_written = global_page_state(NR_WRITTEN);
362362
__entry->background_thresh = background_thresh;
363363
__entry->dirty_thresh = dirty_thresh;
364-
__entry->dirty_limit = global_dirty_limit;
364+
__entry->dirty_limit = global_wb_domain.dirty_limit;
365365
),
366366

367367
TP_printk("dirty=%lu writeback=%lu unstable=%lu "
@@ -463,8 +463,9 @@ TRACE_EVENT(balance_dirty_pages,
463463
unsigned long freerun = (thresh + bg_thresh) / 2;
464464
strlcpy(__entry->bdi, dev_name(bdi->dev), 32);
465465

466-
__entry->limit = global_dirty_limit;
467-
__entry->setpoint = (global_dirty_limit + freerun) / 2;
466+
__entry->limit = global_wb_domain.dirty_limit;
467+
__entry->setpoint = (global_wb_domain.dirty_limit +
468+
freerun) / 2;
468469
__entry->dirty = dirty;
469470
__entry->bdi_setpoint = __entry->setpoint *
470471
bdi_thresh / (thresh + 1);

mm/page-writeback.c

Lines changed: 23 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -122,9 +122,7 @@ EXPORT_SYMBOL(laptop_mode);
122122

123123
/* End of sysctl-exported parameters */
124124

125-
unsigned long global_dirty_limit;
126-
127-
static struct wb_domain global_wb_domain;
125+
struct wb_domain global_wb_domain;
128126

129127
/*
130128
* Length of period for aging writeout fractions of bdis. This is an
@@ -470,9 +468,15 @@ static void writeout_period(unsigned long t)
470468
int wb_domain_init(struct wb_domain *dom, gfp_t gfp)
471469
{
472470
memset(dom, 0, sizeof(*dom));
471+
472+
spin_lock_init(&dom->lock);
473+
473474
init_timer_deferrable(&dom->period_timer);
474475
dom->period_timer.function = writeout_period;
475476
dom->period_timer.data = (unsigned long)dom;
477+
478+
dom->dirty_limit_tstamp = jiffies;
479+
476480
return fprop_global_init(&dom->completions, gfp);
477481
}
478482

@@ -532,7 +536,9 @@ static unsigned long dirty_freerun_ceiling(unsigned long thresh,
532536

533537
static unsigned long hard_dirty_limit(unsigned long thresh)
534538
{
535-
return max(thresh, global_dirty_limit);
539+
struct wb_domain *dom = &global_wb_domain;
540+
541+
return max(thresh, dom->dirty_limit);
536542
}
537543

538544
/**
@@ -916,17 +922,10 @@ static void wb_update_write_bandwidth(struct bdi_writeback *wb,
916922
wb->avg_write_bandwidth = avg;
917923
}
918924

919-
/*
920-
* The global dirtyable memory and dirty threshold could be suddenly knocked
921-
* down by a large amount (eg. on the startup of KVM in a swapless system).
922-
* This may throw the system into deep dirty exceeded state and throttle
923-
* heavy/light dirtiers alike. To retain good responsiveness, maintain
924-
* global_dirty_limit for tracking slowly down to the knocked down dirty
925-
* threshold.
926-
*/
927925
static void update_dirty_limit(unsigned long thresh, unsigned long dirty)
928926
{
929-
unsigned long limit = global_dirty_limit;
927+
struct wb_domain *dom = &global_wb_domain;
928+
unsigned long limit = dom->dirty_limit;
930929

931930
/*
932931
* Follow up in one step.
@@ -939,7 +938,7 @@ static void update_dirty_limit(unsigned long thresh, unsigned long dirty)
939938
/*
940939
* Follow down slowly. Use the higher one as the target, because thresh
941940
* may drop below dirty. This is exactly the reason to introduce
942-
* global_dirty_limit which is guaranteed to lie above the dirty pages.
941+
* dom->dirty_limit which is guaranteed to lie above the dirty pages.
943942
*/
944943
thresh = max(thresh, dirty);
945944
if (limit > thresh) {
@@ -948,28 +947,27 @@ static void update_dirty_limit(unsigned long thresh, unsigned long dirty)
948947
}
949948
return;
950949
update:
951-
global_dirty_limit = limit;
950+
dom->dirty_limit = limit;
952951
}
953952

954953
static void global_update_bandwidth(unsigned long thresh,
955954
unsigned long dirty,
956955
unsigned long now)
957956
{
958-
static DEFINE_SPINLOCK(dirty_lock);
959-
static unsigned long update_time = INITIAL_JIFFIES;
957+
struct wb_domain *dom = &global_wb_domain;
960958

961959
/*
962960
* check locklessly first to optimize away locking for the most time
963961
*/
964-
if (time_before(now, update_time + BANDWIDTH_INTERVAL))
962+
if (time_before(now, dom->dirty_limit_tstamp + BANDWIDTH_INTERVAL))
965963
return;
966964

967-
spin_lock(&dirty_lock);
968-
if (time_after_eq(now, update_time + BANDWIDTH_INTERVAL)) {
965+
spin_lock(&dom->lock);
966+
if (time_after_eq(now, dom->dirty_limit_tstamp + BANDWIDTH_INTERVAL)) {
969967
update_dirty_limit(thresh, dirty);
970-
update_time = now;
968+
dom->dirty_limit_tstamp = now;
971969
}
972-
spin_unlock(&dirty_lock);
970+
spin_unlock(&dom->lock);
973971
}
974972

975973
/*
@@ -1761,10 +1759,12 @@ void laptop_sync_completion(void)
17611759

17621760
void writeback_set_ratelimit(void)
17631761
{
1762+
struct wb_domain *dom = &global_wb_domain;
17641763
unsigned long background_thresh;
17651764
unsigned long dirty_thresh;
1765+
17661766
global_dirty_limits(&background_thresh, &dirty_thresh);
1767-
global_dirty_limit = dirty_thresh;
1767+
dom->dirty_limit = dirty_thresh;
17681768
ratelimit_pages = dirty_thresh / (num_online_cpus() * 32);
17691769
if (ratelimit_pages < 16)
17701770
ratelimit_pages = 16;

0 commit comments

Comments
 (0)