projects
/
firefly-linux-kernel-4.4.55.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
Merge tag 'pm+acpi-2-3.13-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/rafae...
[firefly-linux-kernel-4.4.55.git]
/
block
/
blk-mq.c
diff --git
a/block/blk-mq.c
b/block/blk-mq.c
index 2dc8de86d0d2591bae4e9543709471fba0b1e91d..cdc629cf075b74f27f7801565ec3d90f3e299ce0 100644
(file)
--- a/
block/blk-mq.c
+++ b/
block/blk-mq.c
@@
-171,9
+171,12
@@
bool blk_mq_can_queue(struct blk_mq_hw_ctx *hctx)
}
EXPORT_SYMBOL(blk_mq_can_queue);
}
EXPORT_SYMBOL(blk_mq_can_queue);
-static void blk_mq_rq_ctx_init(struct
blk_mq_ctx *ctx, struct request *rq
,
- unsigned int rw_flags)
+static void blk_mq_rq_ctx_init(struct
request_queue *q, struct blk_mq_ctx *ctx
,
+
struct request *rq,
unsigned int rw_flags)
{
{
+ if (blk_queue_io_stat(q))
+ rw_flags |= REQ_IO_STAT;
+
rq->mq_ctx = ctx;
rq->cmd_flags = rw_flags;
ctx->rq_dispatched[rw_is_sync(rw_flags)]++;
rq->mq_ctx = ctx;
rq->cmd_flags = rw_flags;
ctx->rq_dispatched[rw_is_sync(rw_flags)]++;
@@
-197,7
+200,7
@@
static struct request *blk_mq_alloc_request_pinned(struct request_queue *q,
rq = __blk_mq_alloc_request(hctx, gfp & ~__GFP_WAIT, reserved);
if (rq) {
rq = __blk_mq_alloc_request(hctx, gfp & ~__GFP_WAIT, reserved);
if (rq) {
- blk_mq_rq_ctx_init(ctx, rq, rw);
+ blk_mq_rq_ctx_init(
q,
ctx, rq, rw);
break;
} else if (!(gfp & __GFP_WAIT))
break;
break;
} else if (!(gfp & __GFP_WAIT))
break;
@@
-319,7
+322,7
@@
void __blk_mq_end_io(struct request *rq, int error)
blk_mq_complete_request(rq, error);
}
blk_mq_complete_request(rq, error);
}
-#if defined(CONFIG_SMP)
&& defined(CONFIG_USE_GENERIC_SMP_HELPERS)
+#if defined(CONFIG_SMP)
/*
* Called with interrupts disabled.
/*
* Called with interrupts disabled.
@@
-361,7
+364,7
@@
static int ipi_remote_cpu(struct blk_mq_ctx *ctx, const int cpu,
return true;
}
return true;
}
-#else /* CONFIG_SMP
&& CONFIG_USE_GENERIC_SMP_HELPERS
*/
+#else /* CONFIG_SMP */
static int ipi_remote_cpu(struct blk_mq_ctx *ctx, const int cpu,
struct request *rq, const int error)
{
static int ipi_remote_cpu(struct blk_mq_ctx *ctx, const int cpu,
struct request *rq, const int error)
{
@@
-718,6
+721,8
@@
static void __blk_mq_insert_request(struct blk_mq_hw_ctx *hctx,
{
struct blk_mq_ctx *ctx = rq->mq_ctx;
{
struct blk_mq_ctx *ctx = rq->mq_ctx;
+ trace_block_rq_insert(hctx->queue, rq);
+
list_add_tail(&rq->queuelist, &ctx->rq_list);
blk_mq_hctx_mark_pending(hctx, ctx);
list_add_tail(&rq->queuelist, &ctx->rq_list);
blk_mq_hctx_mark_pending(hctx, ctx);
@@
-921,7
+926,7
@@
static void blk_mq_make_request(struct request_queue *q, struct bio *bio)
trace_block_getrq(q, bio, rw);
rq = __blk_mq_alloc_request(hctx, GFP_ATOMIC, false);
if (likely(rq))
trace_block_getrq(q, bio, rw);
rq = __blk_mq_alloc_request(hctx, GFP_ATOMIC, false);
if (likely(rq))
- blk_mq_rq_ctx_init(ctx, rq, rw);
+ blk_mq_rq_ctx_init(
q,
ctx, rq, rw);
else {
blk_mq_put_ctx(ctx);
trace_block_sleeprq(q, bio, rw);
else {
blk_mq_put_ctx(ctx);
trace_block_sleeprq(q, bio, rw);
@@
-950,7
+955,7
@@
static void blk_mq_make_request(struct request_queue *q, struct bio *bio)
if (plug) {
blk_mq_bio_to_request(rq, bio);
if (plug) {
blk_mq_bio_to_request(rq, bio);
- if (list_empty(&plug->list))
+ if (list_empty(&plug->
mq_
list))
trace_block_plug(q);
else if (request_count >= BLK_MAX_REQUEST_COUNT) {
blk_flush_plug_list(plug, false);
trace_block_plug(q);
else if (request_count >= BLK_MAX_REQUEST_COUNT) {
blk_flush_plug_list(plug, false);
@@
-1377,6
+1382,7
@@
struct request_queue *blk_mq_init_queue(struct blk_mq_reg *reg,
q->queue_hw_ctx = hctxs;
q->mq_ops = reg->ops;
q->queue_hw_ctx = hctxs;
q->mq_ops = reg->ops;
+ q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT;
blk_queue_make_request(q, blk_mq_make_request);
blk_queue_rq_timed_out(q, reg->ops->timeout);
blk_queue_make_request(q, blk_mq_make_request);
blk_queue_rq_timed_out(q, reg->ops->timeout);
@@
-1444,7
+1450,7
@@
void blk_mq_free_queue(struct request_queue *q)
EXPORT_SYMBOL(blk_mq_free_queue);
/* Basically redo blk_mq_init_queue with queue frozen */
EXPORT_SYMBOL(blk_mq_free_queue);
/* Basically redo blk_mq_init_queue with queue frozen */
-static void
__cpuinit
blk_mq_queue_reinit(struct request_queue *q)
+static void blk_mq_queue_reinit(struct request_queue *q)
{
blk_mq_freeze_queue(q);
{
blk_mq_freeze_queue(q);
@@
-1461,8
+1467,8
@@
static void __cpuinit blk_mq_queue_reinit(struct request_queue *q)
blk_mq_unfreeze_queue(q);
}
blk_mq_unfreeze_queue(q);
}
-static int
__cpuinit
blk_mq_queue_reinit_notify(struct notifier_block *nb,
- unsigned long action, void *hcpu)
+static int blk_mq_queue_reinit_notify(struct notifier_block *nb,
+
unsigned long action, void *hcpu)
{
struct request_queue *q;
{
struct request_queue *q;