Merge branch 'linux-linaro-lsk-v4.4' into linux-linaro-lsk-v4.4-android
[firefly-linux-kernel-4.4.55.git] / drivers / mmc / card / block.c
index d8486168415ae1123f6a31d7126cfc9192fe884b..8d169d5a8b017d52a3339489cb751365dca54d05 100644 (file)
@@ -36,6 +36,8 @@
 #include <linux/compat.h>
 #include <linux/pm_runtime.h>
 
+#include <trace/events/mmc.h>
+
 #include <linux/mmc/ioctl.h>
 #include <linux/mmc/card.h>
 #include <linux/mmc/host.h>
 #include "queue.h"
 
 MODULE_ALIAS("mmc:block");
-
-#ifdef KERNEL
 #ifdef MODULE_PARAM_PREFIX
 #undef MODULE_PARAM_PREFIX
 #endif
 #define MODULE_PARAM_PREFIX "mmcblk."
-#endif
 
 #define INAND_CMD38_ARG_EXT_CSD  113
 #define INAND_CMD38_ARG_ERASE    0x00
@@ -171,11 +170,7 @@ static struct mmc_blk_data *mmc_blk_get(struct gendisk *disk)
 
 static inline int mmc_get_devidx(struct gendisk *disk)
 {
-       int devmaj = MAJOR(disk_devt(disk));
-       int devidx = MINOR(disk_devt(disk)) / perdev_minors;
-
-       if (!devmaj)
-               devidx = disk->first_minor / perdev_minors;
+       int devidx = disk->first_minor / perdev_minors;
        return devidx;
 }
 
@@ -293,6 +288,250 @@ out:
        return ret;
 }
 
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+
+static int max_read_speed, max_write_speed, cache_size = 4;
+
+module_param(max_read_speed, int, S_IRUSR | S_IRGRP);
+MODULE_PARM_DESC(max_read_speed, "maximum KB/s read speed 0=off");
+module_param(max_write_speed, int, S_IRUSR | S_IRGRP);
+MODULE_PARM_DESC(max_write_speed, "maximum KB/s write speed 0=off");
+module_param(cache_size, int, S_IRUSR | S_IRGRP);
+MODULE_PARM_DESC(cache_size, "MB high speed memory or SLC cache");
+
+/*
+ * helper macros and expectations:
+ *  size    - unsigned long number of bytes
+ *  jiffies - unsigned long HZ timestamp difference
+ *  speed   - unsigned KB/s transfer rate
+ */
+#define size_and_speed_to_jiffies(size, speed) \
+               ((size) * HZ / (speed) / 1024UL)
+#define jiffies_and_speed_to_size(jiffies, speed) \
+               (((speed) * (jiffies) * 1024UL) / HZ)
+#define jiffies_and_size_to_speed(jiffies, size) \
+               ((size) * HZ / (jiffies) / 1024UL)
+
+/* Limits to report warning */
+/* jiffies_and_size_to_speed(10*HZ, queue_max_hw_sectors(q) * 512UL) ~ 25 */
+#define MIN_SPEED(q) 250 /* 10 times faster than a floppy disk */
+#define MAX_SPEED(q) jiffies_and_size_to_speed(1, queue_max_sectors(q) * 512UL)
+
+#define speed_valid(speed) ((speed) > 0)
+
+static const char off[] = "off\n";
+
+static int max_speed_show(int speed, char *buf)
+{
+       if (speed)
+               return scnprintf(buf, PAGE_SIZE, "%uKB/s\n", speed);
+       else
+               return scnprintf(buf, PAGE_SIZE, off);
+}
+
+static int max_speed_store(const char *buf, struct request_queue *q)
+{
+       unsigned int limit, set = 0;
+
+       if (!strncasecmp(off, buf, sizeof(off) - 2))
+               return set;
+       if (kstrtouint(buf, 0, &set) || (set > INT_MAX))
+               return -EINVAL;
+       if (set == 0)
+               return set;
+       limit = MAX_SPEED(q);
+       if (set > limit)
+               pr_warn("max speed %u ineffective above %u\n", set, limit);
+       limit = MIN_SPEED(q);
+       if (set < limit)
+               pr_warn("max speed %u painful below %u\n", set, limit);
+       return set;
+}
+
+static ssize_t max_write_speed_show(struct device *dev,
+                                struct device_attribute *attr, char *buf)
+{
+       struct mmc_blk_data *md = mmc_blk_get(dev_to_disk(dev));
+       int ret = max_speed_show(atomic_read(&md->queue.max_write_speed), buf);
+
+       mmc_blk_put(md);
+       return ret;
+}
+
+static ssize_t max_write_speed_store(struct device *dev,
+                                 struct device_attribute *attr,
+                                 const char *buf, size_t count)
+{
+       struct mmc_blk_data *md = mmc_blk_get(dev_to_disk(dev));
+       int set = max_speed_store(buf, md->queue.queue);
+
+       if (set < 0) {
+               mmc_blk_put(md);
+               return set;
+       }
+
+       atomic_set(&md->queue.max_write_speed, set);
+       mmc_blk_put(md);
+       return count;
+}
+
+static const DEVICE_ATTR(max_write_speed, S_IRUGO | S_IWUSR,
+       max_write_speed_show, max_write_speed_store);
+
+static ssize_t max_read_speed_show(struct device *dev,
+                                struct device_attribute *attr, char *buf)
+{
+       struct mmc_blk_data *md = mmc_blk_get(dev_to_disk(dev));
+       int ret = max_speed_show(atomic_read(&md->queue.max_read_speed), buf);
+
+       mmc_blk_put(md);
+       return ret;
+}
+
+static ssize_t max_read_speed_store(struct device *dev,
+                                 struct device_attribute *attr,
+                                 const char *buf, size_t count)
+{
+       struct mmc_blk_data *md = mmc_blk_get(dev_to_disk(dev));
+       int set = max_speed_store(buf, md->queue.queue);
+
+       if (set < 0) {
+               mmc_blk_put(md);
+               return set;
+       }
+
+       atomic_set(&md->queue.max_read_speed, set);
+       mmc_blk_put(md);
+       return count;
+}
+
+static const DEVICE_ATTR(max_read_speed, S_IRUGO | S_IWUSR,
+       max_read_speed_show, max_read_speed_store);
+
+static ssize_t cache_size_show(struct device *dev,
+                              struct device_attribute *attr, char *buf)
+{
+       struct mmc_blk_data *md = mmc_blk_get(dev_to_disk(dev));
+       struct mmc_queue *mq = &md->queue;
+       int cache_size = atomic_read(&mq->cache_size);
+       int ret;
+
+       if (!cache_size)
+               ret = scnprintf(buf, PAGE_SIZE, off);
+       else {
+               int speed = atomic_read(&mq->max_write_speed);
+
+               if (!speed_valid(speed))
+                       ret = scnprintf(buf, PAGE_SIZE, "%uMB\n", cache_size);
+               else { /* We accept race between cache_jiffies and cache_used */
+                       unsigned long size = jiffies_and_speed_to_size(
+                               jiffies - mq->cache_jiffies, speed);
+                       long used = atomic_long_read(&mq->cache_used);
+
+                       if (size >= used)
+                               size = 0;
+                       else
+                               size = (used - size) * 100 / cache_size
+                                       / 1024UL / 1024UL;
+
+                       ret = scnprintf(buf, PAGE_SIZE, "%uMB %lu%% used\n",
+                               cache_size, size);
+               }
+       }
+
+       mmc_blk_put(md);
+       return ret;
+}
+
+static ssize_t cache_size_store(struct device *dev,
+                                 struct device_attribute *attr,
+                                 const char *buf, size_t count)
+{
+       struct mmc_blk_data *md;
+       unsigned int set = 0;
+
+       if (strncasecmp(off, buf, sizeof(off) - 2)
+        && (kstrtouint(buf, 0, &set) || (set > INT_MAX)))
+               return -EINVAL;
+
+       md = mmc_blk_get(dev_to_disk(dev));
+       atomic_set(&md->queue.cache_size, set);
+       mmc_blk_put(md);
+       return count;
+}
+
+static const DEVICE_ATTR(cache_size, S_IRUGO | S_IWUSR,
+       cache_size_show, cache_size_store);
+
+/* correct for write-back */
+static long mmc_blk_cache_used(struct mmc_queue *mq, unsigned long waitfor)
+{
+       long used = 0;
+       int speed = atomic_read(&mq->max_write_speed);
+
+       if (speed_valid(speed)) {
+               unsigned long size = jiffies_and_speed_to_size(
+                                       waitfor - mq->cache_jiffies, speed);
+               used = atomic_long_read(&mq->cache_used);
+
+               if (size >= used)
+                       used = 0;
+               else
+                       used -= size;
+       }
+
+       atomic_long_set(&mq->cache_used, used);
+       mq->cache_jiffies = waitfor;
+
+       return used;
+}
+
+static void mmc_blk_simulate_delay(
+       struct mmc_queue *mq,
+       struct request *req,
+       unsigned long waitfor)
+{
+       int max_speed;
+
+       if (!req)
+               return;
+
+       max_speed = (rq_data_dir(req) == READ)
+               ? atomic_read(&mq->max_read_speed)
+               : atomic_read(&mq->max_write_speed);
+       if (speed_valid(max_speed)) {
+               unsigned long bytes = blk_rq_bytes(req);
+
+               if (rq_data_dir(req) != READ) {
+                       int cache_size = atomic_read(&mq->cache_size);
+
+                       if (cache_size) {
+                               unsigned long size = cache_size * 1024L * 1024L;
+                               long used = mmc_blk_cache_used(mq, waitfor);
+
+                               used += bytes;
+                               atomic_long_set(&mq->cache_used, used);
+                               bytes = 0;
+                               if (used > size)
+                                       bytes = used - size;
+                       }
+               }
+               waitfor += size_and_speed_to_jiffies(bytes, max_speed);
+               if (time_is_after_jiffies(waitfor)) {
+                       long msecs = jiffies_to_msecs(waitfor - jiffies);
+
+                       if (likely(msecs > 0))
+                               msleep(msecs);
+               }
+       }
+}
+
+#else
+
+#define mmc_blk_simulate_delay(mq, req, waitfor)
+
+#endif
+
 static int mmc_blk_open(struct block_device *bdev, fmode_t mode)
 {
        struct mmc_blk_data *md = mmc_blk_get(bdev->bd_disk);
@@ -450,9 +689,11 @@ static int ioctl_do_sanitize(struct mmc_card *card)
        pr_debug("%s: %s - SANITIZE IN PROGRESS...\n",
                mmc_hostname(card->host), __func__);
 
+       trace_mmc_blk_erase_start(EXT_CSD_SANITIZE_START, 0, 0);
        err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL,
                                        EXT_CSD_SANITIZE_START, 1,
                                        MMC_SANITIZE_REQ_TIMEOUT);
+       trace_mmc_blk_erase_end(EXT_CSD_SANITIZE_START, 0, 0);
 
        if (err)
                pr_err("%s: %s - EXT_CSD_SANITIZE_START failed. err=%d\n",
@@ -596,6 +837,14 @@ static int mmc_blk_ioctl_cmd(struct block_device *bdev,
        struct mmc_card *card;
        int err = 0, ioc_err = 0;
 
+       /*
+        * The caller must have CAP_SYS_RAWIO, and must be calling this on the
+        * whole block device, not on a partition.  This prevents overspray
+        * between sibling partitions.
+        */
+       if ((!capable(CAP_SYS_RAWIO)) || (bdev != bdev->bd_contains))
+               return -EPERM;
+
        idata = mmc_blk_ioctl_copy_from_user(ic_ptr);
        if (IS_ERR(idata))
                return PTR_ERR(idata);
@@ -638,6 +887,14 @@ static int mmc_blk_ioctl_multi_cmd(struct block_device *bdev,
        int i, err = 0, ioc_err = 0;
        __u64 num_of_cmds;
 
+       /*
+        * The caller must have CAP_SYS_RAWIO, and must be calling this on the
+        * whole block device, not on a partition.  This prevents overspray
+        * between sibling partitions.
+        */
+       if ((!capable(CAP_SYS_RAWIO)) || (bdev != bdev->bd_contains))
+               return -EPERM;
+
        if (copy_from_user(&num_of_cmds, &user->num_of_cmds,
                           sizeof(num_of_cmds)))
                return -EFAULT;
@@ -693,14 +950,6 @@ cmd_err:
 static int mmc_blk_ioctl(struct block_device *bdev, fmode_t mode,
        unsigned int cmd, unsigned long arg)
 {
-       /*
-        * The caller must have CAP_SYS_RAWIO, and must be calling this on the
-        * whole block device, not on a partition.  This prevents overspray
-        * between sibling partitions.
-        */
-       if ((!capable(CAP_SYS_RAWIO)) || (bdev != bdev->bd_contains))
-               return -EPERM;
-
        switch (cmd) {
        case MMC_IOC_CMD:
                return mmc_blk_ioctl_cmd(bdev,
@@ -944,18 +1193,22 @@ static int mmc_blk_cmd_error(struct request *req, const char *name, int error,
                        req->rq_disk->disk_name, "timed out", name, status);
 
                /* If the status cmd initially failed, retry the r/w cmd */
-               if (!status_valid)
+               if (!status_valid) {
+                       pr_err("%s: status not valid, retrying timeout\n", req->rq_disk->disk_name);
                        return ERR_RETRY;
-
+               }
                /*
                 * If it was a r/w cmd crc error, or illegal command
                 * (eg, issued in wrong state) then retry - we should
                 * have corrected the state problem above.
                 */
-               if (status & (R1_COM_CRC_ERROR | R1_ILLEGAL_COMMAND))
+               if (status & (R1_COM_CRC_ERROR | R1_ILLEGAL_COMMAND)) {
+                       pr_err("%s: command error, retrying timeout\n", req->rq_disk->disk_name);
                        return ERR_RETRY;
+               }
 
                /* Otherwise abort the command */
+               pr_err("%s: not retrying timeout\n", req->rq_disk->disk_name);
                return ERR_ABORT;
 
        default:
@@ -1255,6 +1508,23 @@ static int mmc_blk_issue_flush(struct mmc_queue *mq, struct request *req)
        if (ret)
                ret = -EIO;
 
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+       else if (atomic_read(&mq->cache_size)) {
+               long used = mmc_blk_cache_used(mq, jiffies);
+
+               if (used) {
+                       int speed = atomic_read(&mq->max_write_speed);
+
+                       if (speed_valid(speed)) {
+                               unsigned long msecs = jiffies_to_msecs(
+                                       size_and_speed_to_jiffies(
+                                               used, speed));
+                               if (msecs)
+                                       msleep(msecs);
+                       }
+               }
+       }
+#endif
        blk_end_request_all(req, ret);
 
        return ret ? 0 : 1;
@@ -1747,7 +2017,7 @@ static void mmc_blk_packed_hdr_wrq_prep(struct mmc_queue_req *mqrq,
        struct mmc_blk_data *md = mq->data;
        struct mmc_packed *packed = mqrq->packed;
        bool do_rel_wr, do_data_tag;
-       u32 *packed_cmd_hdr;
+       __le32 *packed_cmd_hdr;
        u8 hdr_blocks;
        u8 i = 1;
 
@@ -1759,8 +2029,8 @@ static void mmc_blk_packed_hdr_wrq_prep(struct mmc_queue_req *mqrq,
 
        packed_cmd_hdr = packed->cmd_hdr;
        memset(packed_cmd_hdr, 0, sizeof(packed->cmd_hdr));
-       packed_cmd_hdr[0] = (packed->nr_entries << 16) |
-               (PACKED_CMD_WR << 8) | PACKED_CMD_VER;
+       packed_cmd_hdr[0] = cpu_to_le32((packed->nr_entries << 16) |
+               (PACKED_CMD_WR << 8) | PACKED_CMD_VER);
        hdr_blocks = mmc_large_sector(card) ? 8 : 1;
 
        /*
@@ -1774,14 +2044,14 @@ static void mmc_blk_packed_hdr_wrq_prep(struct mmc_queue_req *mqrq,
                        ((brq->data.blocks * brq->data.blksz) >=
                         card->ext_csd.data_tag_unit_size);
                /* Argument of CMD23 */
-               packed_cmd_hdr[(i * 2)] =
+               packed_cmd_hdr[(i * 2)] = cpu_to_le32(
                        (do_rel_wr ? MMC_CMD23_ARG_REL_WR : 0) |
                        (do_data_tag ? MMC_CMD23_ARG_TAG_REQ : 0) |
-                       blk_rq_sectors(prq);
+                       blk_rq_sectors(prq));
                /* Argument of CMD18 or CMD25 */
-               packed_cmd_hdr[((i * 2)) + 1] =
+               packed_cmd_hdr[((i * 2)) + 1] = cpu_to_le32(
                        mmc_card_blockaddr(card) ?
-                       blk_rq_pos(prq) : blk_rq_pos(prq) << 9;
+                       blk_rq_pos(prq) : blk_rq_pos(prq) << 9);
                packed->blocks += blk_rq_sectors(prq);
                i++;
        }
@@ -1934,6 +2204,9 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *rqc)
        struct mmc_async_req *areq;
        const u8 packed_nr = 2;
        u8 reqs = 0;
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+       unsigned long waitfor = jiffies;
+#endif
 
        if (!rqc && !mq->mqrq_prev->req)
                return 0;
@@ -1984,6 +2257,8 @@ static int mmc_blk_issue_rw_rq(struct mmc_queue *mq, struct request *rqc)
                         */
                        mmc_blk_reset_success(md, type);
 
+                       mmc_blk_simulate_delay(mq, rqc, waitfor);
+
                        if (mmc_packed_cmd(mq_rq->cmd_type)) {
                                ret = mmc_blk_end_packed_req(mq_rq);
                                break;
@@ -2244,6 +2519,7 @@ static struct mmc_blk_data *mmc_blk_alloc_req(struct mmc_card *card,
        md->disk->queue = md->queue.queue;
        md->disk->driverfs_dev = parent;
        set_disk_ro(md->disk, md->read_only || default_ro);
+       md->disk->flags = GENHD_FL_EXT_DEVT;
        if (area_type & (MMC_BLK_DATA_AREA_RPMB | MMC_BLK_DATA_AREA_BOOT))
                md->disk->flags |= GENHD_FL_NO_PART_SCAN;
 
@@ -2271,7 +2547,8 @@ static struct mmc_blk_data *mmc_blk_alloc_req(struct mmc_card *card,
        set_capacity(md->disk, size);
 
        if (mmc_host_cmd23(card->host)) {
-               if (mmc_card_mmc(card) ||
+               if ((mmc_card_mmc(card) &&
+                    card->csd.mmca_vsn >= CSD_SPEC_VER_3) ||
                    (mmc_card_sd(card) &&
                     card->scr.cmds & SD_SCR_CMD23_SUPPORT))
                        md->flags |= MMC_BLK_CMD23;
@@ -2401,6 +2678,14 @@ static void mmc_blk_remove_req(struct mmc_blk_data *md)
                                        card->ext_csd.boot_ro_lockable)
                                device_remove_file(disk_to_dev(md->disk),
                                        &md->power_ro_lock);
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+                       device_remove_file(disk_to_dev(md->disk),
+                                               &dev_attr_max_write_speed);
+                       device_remove_file(disk_to_dev(md->disk),
+                                               &dev_attr_max_read_speed);
+                       device_remove_file(disk_to_dev(md->disk),
+                                               &dev_attr_cache_size);
+#endif
 
                        del_gendisk(md->disk);
                }
@@ -2436,6 +2721,24 @@ static int mmc_add_disk(struct mmc_blk_data *md)
        ret = device_create_file(disk_to_dev(md->disk), &md->force_ro);
        if (ret)
                goto force_ro_fail;
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+       atomic_set(&md->queue.max_write_speed, max_write_speed);
+       ret = device_create_file(disk_to_dev(md->disk),
+                       &dev_attr_max_write_speed);
+       if (ret)
+               goto max_write_speed_fail;
+       atomic_set(&md->queue.max_read_speed, max_read_speed);
+       ret = device_create_file(disk_to_dev(md->disk),
+                       &dev_attr_max_read_speed);
+       if (ret)
+               goto max_read_speed_fail;
+       atomic_set(&md->queue.cache_size, cache_size);
+       atomic_long_set(&md->queue.cache_used, 0);
+       md->queue.cache_jiffies = jiffies;
+       ret = device_create_file(disk_to_dev(md->disk), &dev_attr_cache_size);
+       if (ret)
+               goto cache_size_fail;
+#endif
 
        if ((md->area_type & MMC_BLK_DATA_AREA_BOOT) &&
             card->ext_csd.boot_ro_lockable) {
@@ -2460,6 +2763,14 @@ static int mmc_add_disk(struct mmc_blk_data *md)
        return ret;
 
 power_ro_lock_fail:
+#ifdef CONFIG_MMC_SIMULATE_MAX_SPEED
+       device_remove_file(disk_to_dev(md->disk), &dev_attr_cache_size);
+cache_size_fail:
+       device_remove_file(disk_to_dev(md->disk), &dev_attr_max_read_speed);
+max_read_speed_fail:
+       device_remove_file(disk_to_dev(md->disk), &dev_attr_max_write_speed);
+max_write_speed_fail:
+#endif
        device_remove_file(disk_to_dev(md->disk), &md->force_ro);
 force_ro_fail:
        del_gendisk(md->disk);
@@ -2506,11 +2817,12 @@ static const struct mmc_fixup blk_fixups[] =
                  MMC_QUIRK_BLK_NO_CMD23),
 
        /*
-        * Some Micron MMC cards needs longer data read timeout than
-        * indicated in CSD.
+        * Some MMC cards need longer data read timeout than indicated in CSD.
         */
        MMC_FIXUP(CID_NAME_ANY, CID_MANFID_MICRON, 0x200, add_quirk_mmc,
                  MMC_QUIRK_LONG_READ_TIME),
+       MMC_FIXUP("008GE0", CID_MANFID_TOSHIBA, CID_OEMID_ANY, add_quirk_mmc,
+                 MMC_QUIRK_LONG_READ_TIME),
 
        /*
         * On these Samsung MoviNAND parts, performing secure erase or