Merge remote-tracking branch 'lsk/v3.10/topic/gator' into linux-linaro-lsk
[firefly-linux-kernel-4.4.55.git] / drivers / md / dm-thin.c
index e36f81e282eb373d402487d9f80acd0ab0d113f6..86a2a5e3b26bacdc0c773076cf7283cacd188701 100644 (file)
@@ -512,6 +512,7 @@ struct dm_thin_new_mapping {
        unsigned quiesced:1;
        unsigned prepared:1;
        unsigned pass_discard:1;
+       unsigned definitely_not_shared:1;
 
        struct thin_c *tc;
        dm_block_t virt_block;
@@ -683,7 +684,15 @@ static void process_prepared_discard_passdown(struct dm_thin_new_mapping *m)
        cell_defer_no_holder(tc, m->cell2);
 
        if (m->pass_discard)
-               remap_and_issue(tc, m->bio, m->data_block);
+               if (m->definitely_not_shared)
+                       remap_and_issue(tc, m->bio, m->data_block);
+               else {
+                       bool used = false;
+                       if (dm_pool_block_is_used(tc->pool->pmd, m->data_block, &used) || used)
+                               bio_endio(m->bio, 0);
+                       else
+                               remap_and_issue(tc, m->bio, m->data_block);
+               }
        else
                bio_endio(m->bio, 0);
 
@@ -751,13 +760,17 @@ static int ensure_next_mapping(struct pool *pool)
 
 static struct dm_thin_new_mapping *get_next_mapping(struct pool *pool)
 {
-       struct dm_thin_new_mapping *r = pool->next_mapping;
+       struct dm_thin_new_mapping *m = pool->next_mapping;
 
        BUG_ON(!pool->next_mapping);
 
+       memset(m, 0, sizeof(struct dm_thin_new_mapping));
+       INIT_LIST_HEAD(&m->list);
+       m->bio = NULL;
+
        pool->next_mapping = NULL;
 
-       return r;
+       return m;
 }
 
 static void schedule_copy(struct thin_c *tc, dm_block_t virt_block,
@@ -769,15 +782,10 @@ static void schedule_copy(struct thin_c *tc, dm_block_t virt_block,
        struct pool *pool = tc->pool;
        struct dm_thin_new_mapping *m = get_next_mapping(pool);
 
-       INIT_LIST_HEAD(&m->list);
-       m->quiesced = 0;
-       m->prepared = 0;
        m->tc = tc;
        m->virt_block = virt_block;
        m->data_block = data_dest;
        m->cell = cell;
-       m->err = 0;
-       m->bio = NULL;
 
        if (!dm_deferred_set_add_work(pool->shared_read_ds, &m->list))
                m->quiesced = 1;
@@ -840,15 +848,12 @@ static void schedule_zero(struct thin_c *tc, dm_block_t virt_block,
        struct pool *pool = tc->pool;
        struct dm_thin_new_mapping *m = get_next_mapping(pool);
 
-       INIT_LIST_HEAD(&m->list);
        m->quiesced = 1;
        m->prepared = 0;
        m->tc = tc;
        m->virt_block = virt_block;
        m->data_block = data_block;
        m->cell = cell;
-       m->err = 0;
-       m->bio = NULL;
 
        /*
         * If the whole block of data is being overwritten or we are not
@@ -1032,12 +1037,12 @@ static void process_discard(struct thin_c *tc, struct bio *bio)
                         */
                        m = get_next_mapping(pool);
                        m->tc = tc;
-                       m->pass_discard = (!lookup_result.shared) && pool->pf.discard_passdown;
+                       m->pass_discard = pool->pf.discard_passdown;
+                       m->definitely_not_shared = !lookup_result.shared;
                        m->virt_block = block;
                        m->data_block = lookup_result.block;
                        m->cell = cell;
                        m->cell2 = cell2;
-                       m->err = 0;
                        m->bio = bio;
 
                        if (!dm_deferred_set_add_work(pool->all_io_ds, &m->list)) {
@@ -1317,9 +1322,9 @@ static void process_deferred_bios(struct pool *pool)
                 */
                if (ensure_next_mapping(pool)) {
                        spin_lock_irqsave(&pool->lock, flags);
+                       bio_list_add(&pool->deferred_bios, bio);
                        bio_list_merge(&pool->deferred_bios, &bios);
                        spin_unlock_irqrestore(&pool->lock, flags);
-
                        break;
                }
 
@@ -1339,7 +1344,8 @@ static void process_deferred_bios(struct pool *pool)
        bio_list_init(&pool->deferred_flush_bios);
        spin_unlock_irqrestore(&pool->lock, flags);
 
-       if (bio_list_empty(&bios) && !need_commit_due_to_time(pool))
+       if (bio_list_empty(&bios) &&
+           !(dm_pool_changed_this_transaction(pool->pmd) && need_commit_due_to_time(pool)))
                return;
 
        if (commit_or_fallback(pool)) {
@@ -2641,7 +2647,8 @@ static void set_discard_limits(struct pool_c *pt, struct queue_limits *limits)
         */
        if (pt->adjusted_pf.discard_passdown) {
                data_limits = &bdev_get_queue(pt->data_dev->bdev)->limits;
-               limits->discard_granularity = data_limits->discard_granularity;
+               limits->discard_granularity = max(data_limits->discard_granularity,
+                                                 pool->sectors_per_block << SECTOR_SHIFT);
        } else
                limits->discard_granularity = pool->sectors_per_block << SECTOR_SHIFT;
 }
@@ -2778,6 +2785,7 @@ static int thin_ctr(struct dm_target *ti, unsigned argc, char **argv)
 
        if (get_pool_mode(tc->pool) == PM_FAIL) {
                ti->error = "Couldn't open thin device, Pool is in fail mode";
+               r = -EINVAL;
                goto bad_thin_open;
        }
 
@@ -2789,7 +2797,7 @@ static int thin_ctr(struct dm_target *ti, unsigned argc, char **argv)
 
        r = dm_set_target_max_io_len(ti, tc->pool->sectors_per_block);
        if (r)
-               goto bad_thin_open;
+               goto bad_target_max_io_len;
 
        ti->num_flush_bios = 1;
        ti->flush_supported = true;
@@ -2810,6 +2818,8 @@ static int thin_ctr(struct dm_target *ti, unsigned argc, char **argv)
 
        return 0;
 
+bad_target_max_io_len:
+       dm_pool_close_thin_device(tc->td);
 bad_thin_open:
        __pool_dec(tc->pool);
 bad_pool_lookup: