libceph: kill off osd request r_data_in and r_data_out
authorAlex Elder <elder@inktank.com>
Fri, 5 Apr 2013 06:27:12 +0000 (01:27 -0500)
committerSage Weil <sage@inktank.com>
Thu, 2 May 2013 04:18:25 +0000 (21:18 -0700)
Finally!  Convert the osd op data pointers into real structures, and
make the switch over to using them instead of having all ops share
the in and/or out data structures in the osd request.

Set up a new function to traverse the set of ops and release any
data associated with them (pages).

This and the patches leading up to it resolve:
    http://tracker.ceph.com/issues/4657

Signed-off-by: Alex Elder <elder@inktank.com>
Reviewed-by: Josh Durgin <josh.durgin@inktank.com>
include/linux/ceph/osd_client.h
net/ceph/osd_client.c

index dd4ca4ba8cab1d77fddad5ef0cd0dc30dfb8cab4..4ec46c0ceaf76405dbeebd7c5b2551ff47474da9 100644 (file)
@@ -87,14 +87,14 @@ struct ceph_osd_req_op {
                        u64 offset, length;
                        u64 truncate_size;
                        u32 truncate_seq;
-                       struct ceph_osd_data *osd_data;
+                       struct ceph_osd_data osd_data;
                } extent;
                struct {
                        const char *class_name;
                        const char *method_name;
                        const void *request_data;
-                       struct ceph_osd_data *request_info;
-                       struct ceph_osd_data *response_data;
+                       struct ceph_osd_data request_info;
+                       struct ceph_osd_data response_data;
                        u32 request_data_len;
                        __u8 class_len;
                        __u8 method_len;
@@ -164,9 +164,6 @@ struct ceph_osd_request {
 
        struct ceph_file_layout r_file_layout;
        struct ceph_snap_context *r_snapc;    /* snap context for writes */
-
-       struct ceph_osd_data r_data_in;
-       struct ceph_osd_data r_data_out;
 };
 
 struct ceph_osd_event {
index cc4003fdc01f8c23c81c8ebbad85f26d5272050c..2562e4e52245771e0f455ea72abb0d3a74b470dc 100644 (file)
@@ -121,8 +121,7 @@ osd_req_op_extent_osd_data(struct ceph_osd_request *osd_req,
 {
        BUG_ON(which >= osd_req->r_num_ops);
 
-       /* return &osd_req->r_ops[which].extent.osd_data; */
-       return write_request ? &osd_req->r_data_out : &osd_req->r_data_in;
+       return &osd_req->r_ops[which].extent.osd_data;
 }
 EXPORT_SYMBOL(osd_req_op_extent_osd_data);
 
@@ -132,8 +131,7 @@ osd_req_op_cls_request_info(struct ceph_osd_request *osd_req,
 {
        BUG_ON(which >= osd_req->r_num_ops);
 
-       /* return &osd_req->r_ops[which].cls.request_info; */
-       return &osd_req->r_data_out;    /* Request data is outgoing */
+       return &osd_req->r_ops[which].cls.request_info;
 }
 EXPORT_SYMBOL(osd_req_op_cls_request_info);    /* ??? */
 
@@ -143,8 +141,7 @@ osd_req_op_cls_response_data(struct ceph_osd_request *osd_req,
 {
        BUG_ON(which >= osd_req->r_num_ops);
 
-       /* return &osd_req->r_ops[which].cls.response_data; */
-       return &osd_req->r_data_in;     /* Response data is incoming */
+       return &osd_req->r_ops[which].cls.response_data;
 }
 EXPORT_SYMBOL(osd_req_op_cls_response_data);   /* ??? */
 
@@ -158,9 +155,6 @@ void osd_req_op_extent_osd_data_pages(struct ceph_osd_request *osd_req,
        osd_data = osd_req_op_extent_osd_data(osd_req, which, write_request);
        ceph_osd_data_pages_init(osd_data, pages, length, alignment,
                                pages_from_pool, own_pages);
-
-       osd_req->r_ops[which].extent.osd_data =
-               osd_req_op_extent_osd_data(osd_req, which, write_request);
 }
 EXPORT_SYMBOL(osd_req_op_extent_osd_data_pages);
 
@@ -172,9 +166,6 @@ void osd_req_op_extent_osd_data_pagelist(struct ceph_osd_request *osd_req,
 
        osd_data = osd_req_op_extent_osd_data(osd_req, which, write_request);
        ceph_osd_data_pagelist_init(osd_data, pagelist);
-
-       osd_req->r_ops[which].extent.osd_data =
-               osd_req_op_extent_osd_data(osd_req, which, write_request);
 }
 EXPORT_SYMBOL(osd_req_op_extent_osd_data_pagelist);
 
@@ -187,9 +178,6 @@ void osd_req_op_extent_osd_data_bio(struct ceph_osd_request *osd_req,
 
        osd_data = osd_req_op_extent_osd_data(osd_req, which, write_request);
        ceph_osd_data_bio_init(osd_data, bio, bio_length);
-
-       osd_req->r_ops[which].extent.osd_data =
-               osd_req_op_extent_osd_data(osd_req, which, write_request);
 }
 EXPORT_SYMBOL(osd_req_op_extent_osd_data_bio);
 #endif /* CONFIG_BLOCK */
@@ -202,9 +190,6 @@ static void osd_req_op_cls_request_info_pagelist(
 
        osd_data = osd_req_op_cls_request_info(osd_req, which);
        ceph_osd_data_pagelist_init(osd_data, pagelist);
-
-       osd_req->r_ops[which].cls.request_info =
-               osd_req_op_cls_request_info(osd_req, which);
 }
 
 void osd_req_op_cls_response_data_pages(struct ceph_osd_request *osd_req,
@@ -216,9 +201,6 @@ void osd_req_op_cls_response_data_pages(struct ceph_osd_request *osd_req,
        osd_data = osd_req_op_cls_response_data(osd_req, which);
        ceph_osd_data_pages_init(osd_data, pages, length, alignment,
                                pages_from_pool, own_pages);
-
-       osd_req->r_ops[which].cls.response_data =
-               osd_req_op_cls_response_data(osd_req, which);
 }
 EXPORT_SYMBOL(osd_req_op_cls_response_data_pages);
 
@@ -241,18 +223,39 @@ static u64 ceph_osd_data_length(struct ceph_osd_data *osd_data)
        }
 }
 
+
 static void ceph_osd_data_release(struct ceph_osd_data *osd_data)
 {
-       if (osd_data->type != CEPH_OSD_DATA_TYPE_PAGES)
-               return;
-
-       if (osd_data->own_pages) {
+       if (osd_data->type == CEPH_OSD_DATA_TYPE_PAGES && osd_data->own_pages) {
                int num_pages;
 
                num_pages = calc_pages_for((u64)osd_data->alignment,
                                                (u64)osd_data->length);
                ceph_release_page_vector(osd_data->pages, num_pages);
        }
+       ceph_osd_data_init(osd_data);
+}
+
+static void osd_req_op_data_release(struct ceph_osd_request *osd_req,
+                       unsigned int which)
+{
+       struct ceph_osd_req_op *op;
+
+       BUG_ON(which >= osd_req->r_num_ops);
+       op = &osd_req->r_ops[which];
+
+       switch (op->op) {
+       case CEPH_OSD_OP_READ:
+       case CEPH_OSD_OP_WRITE:
+               ceph_osd_data_release(&op->extent.osd_data);
+               break;
+       case CEPH_OSD_OP_CALL:
+               ceph_osd_data_release(&op->cls.request_info);
+               ceph_osd_data_release(&op->cls.response_data);
+               break;
+       default:
+               break;
+       }
 }
 
 /*
@@ -261,6 +264,7 @@ static void ceph_osd_data_release(struct ceph_osd_data *osd_data)
 void ceph_osdc_release_request(struct kref *kref)
 {
        struct ceph_osd_request *req;
+       unsigned int which;
 
        req = container_of(kref, struct ceph_osd_request, r_kref);
        if (req->r_request)
@@ -270,8 +274,8 @@ void ceph_osdc_release_request(struct kref *kref)
                ceph_msg_put(req->r_reply);
        }
 
-       ceph_osd_data_release(&req->r_data_in);
-       ceph_osd_data_release(&req->r_data_out);
+       for (which = 0; which < req->r_num_ops; which++)
+               osd_req_op_data_release(req, which);
 
        ceph_put_snap_context(req->r_snapc);
        if (req->r_mempool)
@@ -595,27 +599,22 @@ static u64 osd_req_encode_op(struct ceph_osd_request *req,
                        cpu_to_le64(src->extent.truncate_size);
                dst->extent.truncate_seq =
                        cpu_to_le32(src->extent.truncate_seq);
-               if (src->op == CEPH_OSD_OP_WRITE) {
-                       WARN_ON(src->extent.osd_data != &req->r_data_out);
+               if (src->op == CEPH_OSD_OP_WRITE)
                        ceph_osdc_msg_data_set(req->r_request,
-                                               src->extent.osd_data);
-               } else {
-                       WARN_ON(src->extent.osd_data != &req->r_data_in);
+                                               &src->extent.osd_data);
+               else
                        ceph_osdc_msg_data_set(req->r_reply,
-                                               src->extent.osd_data);
-               }
+                                               &src->extent.osd_data);
                break;
        case CEPH_OSD_OP_CALL:
                dst->cls.class_len = src->cls.class_len;
                dst->cls.method_len = src->cls.method_len;
                dst->cls.indata_len = cpu_to_le32(src->cls.request_data_len);
-               WARN_ON(src->cls.response_data != &req->r_data_in);
-               ceph_osdc_msg_data_set(req->r_reply, src->cls.response_data);
-               WARN_ON(src->cls.request_info != &req->r_data_out);
-               ceph_osdc_msg_data_set(req->r_request, src->cls.request_info);
-               BUG_ON(src->cls.request_info->type !=
+               ceph_osdc_msg_data_set(req->r_reply, &src->cls.response_data);
+               ceph_osdc_msg_data_set(req->r_request, &src->cls.request_info);
+               BUG_ON(src->cls.request_info.type !=
                                        CEPH_OSD_DATA_TYPE_PAGELIST);
-               request_data_len = src->cls.request_info->pagelist->length;
+               request_data_len = src->cls.request_info.pagelist->length;
                break;
        case CEPH_OSD_OP_STARTSYNC:
                break;