block/fs: keep track of the task that dirtied the page

Background writes happen in the context of a background thread.
It is very useful to identify the actual task that generated the
request instead of background task that submited the request.
Hence keep track of the task when a page gets dirtied and dump
this task info while tracing. Not all the pages in the bio are
dirtied by the same task but most likely it will be, since the
sectors accessed on the device must be adjacent.

Change-Id: I6afba85a2063dd3350a0141ba87cf8440ce9f777
Signed-off-by: Venkat Gopalakrishnan <venkatg@codeaurora.org>
This commit is contained in:
Venkat Gopalakrishnan 2014-05-01 19:15:23 -07:00
parent 71a6aa0aa7
commit 12d982a03b
4 changed files with 81 additions and 15 deletions

View File

@ -1935,6 +1935,7 @@ EXPORT_SYMBOL(generic_make_request);
*/
void submit_bio(int rw, struct bio *bio)
{
struct task_struct *tsk = current;
bio->bi_rw |= rw;
/*
@ -1958,8 +1959,18 @@ void submit_bio(int rw, struct bio *bio)
if (unlikely(block_dump)) {
char b[BDEVNAME_SIZE];
/*
* Not all the pages in the bio are dirtied by the
* same task but most likely it will be, since the
* sectors accessed on the device must be adjacent.
*/
if (bio->bi_io_vec && bio->bi_io_vec->bv_page &&
bio->bi_io_vec->bv_page->tsk_dirty)
tsk = bio->bi_io_vec->bv_page->tsk_dirty;
printk(KERN_DEBUG "%s(%d): %s block %Lu on %s (%u sectors)\n",
current->comm, task_pid_nr(current),
tsk->comm, task_pid_nr(tsk),
(rw & WRITE) ? "WRITE" : "READ",
(unsigned long long)bio->bi_sector,
bdevname(bio->bi_bdev, b),

View File

@ -628,6 +628,8 @@ static void __set_page_dirty(struct page *page,
account_page_dirtied(page, mapping);
radix_tree_tag_set(&mapping->page_tree,
page_index(page), PAGECACHE_TAG_DIRTY);
/* Save the task that is dirtying this page */
page->tsk_dirty = current;
}
spin_unlock_irqrestore(&mapping->tree_lock, flags);
__mark_inode_dirty(mapping->host, I_DIRTY_PAGES);

View File

@ -166,6 +166,7 @@ struct page {
#ifdef CONFIG_WANT_PAGE_DEBUG_FLAGS
unsigned long debug_flags; /* Use atomic bitops on this */
#endif
struct task_struct *tsk_dirty; /* task that sets this page dirty */
#ifdef CONFIG_KMEMCHECK
/*

View File

@ -187,9 +187,9 @@ static const u32 ddir_act[2] = { BLK_TC_ACT(BLK_TC_READ),
* blk_io_trace structure and places it in a per-cpu subbuffer.
*/
static void __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes,
int rw, u32 what, int error, int pdu_len, void *pdu_data)
int rw, u32 what, int error, int pdu_len,
void *pdu_data, struct task_struct *tsk)
{
struct task_struct *tsk = current;
struct ring_buffer_event *event = NULL;
struct ring_buffer *buffer = NULL;
struct blk_io_trace *t;
@ -695,18 +695,33 @@ static void blk_add_trace_rq(struct request_queue *q, struct request *rq,
u32 what)
{
struct blk_trace *bt = q->blk_trace;
struct task_struct *tsk = current;
if (likely(!bt))
return;
/*
* Use the bio context for all events except ISSUE and
* COMPLETE events.
*
* Not all the pages in the bio are dirtied by the same task but
* most likely it will be, since the sectors accessed on the device
* must be adjacent.
*/
if (!((what == BLK_TA_ISSUE) || (what == BLK_TA_COMPLETE)) &&
bio_has_data(rq->bio) && rq->bio->bi_io_vec &&
rq->bio->bi_io_vec->bv_page &&
rq->bio->bi_io_vec->bv_page->tsk_dirty)
tsk = rq->bio->bi_io_vec->bv_page->tsk_dirty;
if (rq->cmd_type == REQ_TYPE_BLOCK_PC) {
what |= BLK_TC_ACT(BLK_TC_PC);
__blk_add_trace(bt, 0, blk_rq_bytes(rq), rq->cmd_flags,
what, rq->errors, rq->cmd_len, rq->cmd);
what, rq->errors, rq->cmd_len, rq->cmd, tsk);
} else {
what |= BLK_TC_ACT(BLK_TC_FS);
__blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq),
rq->cmd_flags, what, rq->errors, 0, NULL);
rq->cmd_flags, what, rq->errors, 0, NULL, tsk);
}
}
@ -757,15 +772,25 @@ static void blk_add_trace_bio(struct request_queue *q, struct bio *bio,
u32 what, int error)
{
struct blk_trace *bt = q->blk_trace;
struct task_struct *tsk = current;
if (likely(!bt))
return;
/*
* Not all the pages in the bio are dirtied by the same task but
* most likely it will be, since the sectors accessed on the device
* must be adjacent.
*/
if (bio_has_data(bio) && bio->bi_io_vec && bio->bi_io_vec->bv_page &&
bio->bi_io_vec->bv_page->tsk_dirty)
tsk = bio->bi_io_vec->bv_page->tsk_dirty;
if (!error && !bio_flagged(bio, BIO_UPTODATE))
error = EIO;
__blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw, what,
error, 0, NULL);
error, 0, NULL, tsk);
}
static void blk_add_trace_bio_bounce(void *ignore,
@ -813,7 +838,8 @@ static void blk_add_trace_getrq(void *ignore,
struct blk_trace *bt = q->blk_trace;
if (bt)
__blk_add_trace(bt, 0, 0, rw, BLK_TA_GETRQ, 0, 0, NULL);
__blk_add_trace(bt, 0, 0, rw, BLK_TA_GETRQ, 0, 0,
NULL, current);
}
}
@ -829,7 +855,7 @@ static void blk_add_trace_sleeprq(void *ignore,
if (bt)
__blk_add_trace(bt, 0, 0, rw, BLK_TA_SLEEPRQ,
0, 0, NULL);
0, 0, NULL, current);
}
}
@ -838,7 +864,8 @@ static void blk_add_trace_plug(void *ignore, struct request_queue *q)
struct blk_trace *bt = q->blk_trace;
if (bt)
__blk_add_trace(bt, 0, 0, 0, BLK_TA_PLUG, 0, 0, NULL);
__blk_add_trace(bt, 0, 0, 0, BLK_TA_PLUG, 0, 0, NULL,
current);
}
static void blk_add_trace_unplug(void *ignore, struct request_queue *q,
@ -855,7 +882,8 @@ static void blk_add_trace_unplug(void *ignore, struct request_queue *q,
else
what = BLK_TA_UNPLUG_TIMER;
__blk_add_trace(bt, 0, 0, 0, what, 0, sizeof(rpdu), &rpdu);
__blk_add_trace(bt, 0, 0, 0, what, 0, sizeof(rpdu), &rpdu,
current);
}
}
@ -864,13 +892,19 @@ static void blk_add_trace_split(void *ignore,
unsigned int pdu)
{
struct blk_trace *bt = q->blk_trace;
struct task_struct *tsk = current;
if (bt) {
__be64 rpdu = cpu_to_be64(pdu);
if (bio_has_data(bio) && bio->bi_io_vec &&
bio->bi_io_vec->bv_page &&
bio->bi_io_vec->bv_page->tsk_dirty)
tsk = bio->bi_io_vec->bv_page->tsk_dirty;
__blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw,
BLK_TA_SPLIT, !bio_flagged(bio, BIO_UPTODATE),
sizeof(rpdu), &rpdu);
sizeof(rpdu), &rpdu, tsk);
}
}
@ -893,6 +927,7 @@ static void blk_add_trace_bio_remap(void *ignore,
{
struct blk_trace *bt = q->blk_trace;
struct blk_io_trace_remap r;
struct task_struct *tsk = current;
if (likely(!bt))
return;
@ -901,9 +936,14 @@ static void blk_add_trace_bio_remap(void *ignore,
r.device_to = cpu_to_be32(bio->bi_bdev->bd_dev);
r.sector_from = cpu_to_be64(from);
if (bio_has_data(bio) && bio->bi_io_vec &&
bio->bi_io_vec->bv_page &&
bio->bi_io_vec->bv_page->tsk_dirty)
tsk = bio->bi_io_vec->bv_page->tsk_dirty;
__blk_add_trace(bt, bio->bi_sector, bio->bi_size, bio->bi_rw,
BLK_TA_REMAP, !bio_flagged(bio, BIO_UPTODATE),
sizeof(r), &r);
sizeof(r), &r, tsk);
}
/**
@ -926,6 +966,7 @@ static void blk_add_trace_rq_remap(void *ignore,
{
struct blk_trace *bt = q->blk_trace;
struct blk_io_trace_remap r;
struct task_struct *tsk = current;
if (likely(!bt))
return;
@ -934,9 +975,14 @@ static void blk_add_trace_rq_remap(void *ignore,
r.device_to = cpu_to_be32(disk_devt(rq->rq_disk));
r.sector_from = cpu_to_be64(from);
if (bio_has_data(rq->bio) && rq->bio->bi_io_vec &&
rq->bio->bi_io_vec->bv_page &&
rq->bio->bi_io_vec->bv_page->tsk_dirty)
tsk = rq->bio->bi_io_vec->bv_page->tsk_dirty;
__blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq),
rq_data_dir(rq), BLK_TA_REMAP, !!rq->errors,
sizeof(r), &r);
sizeof(r), &r, tsk);
}
/**
@ -955,16 +1001,22 @@ void blk_add_driver_data(struct request_queue *q,
void *data, size_t len)
{
struct blk_trace *bt = q->blk_trace;
struct task_struct *tsk = current;
if (likely(!bt))
return;
if (bio_has_data(rq->bio) && rq->bio->bi_io_vec &&
rq->bio->bi_io_vec->bv_page &&
rq->bio->bi_io_vec->bv_page->tsk_dirty)
tsk = rq->bio->bi_io_vec->bv_page->tsk_dirty;
if (rq->cmd_type == REQ_TYPE_BLOCK_PC)
__blk_add_trace(bt, 0, blk_rq_bytes(rq), 0,
BLK_TA_DRV_DATA, rq->errors, len, data);
BLK_TA_DRV_DATA, rq->errors, len, data, tsk);
else
__blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), 0,
BLK_TA_DRV_DATA, rq->errors, len, data);
BLK_TA_DRV_DATA, rq->errors, len, data, tsk);
}
EXPORT_SYMBOL_GPL(blk_add_driver_data);