/* * Copyright (c) 2005 Ammasso, Inc. All rights reserved. * Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses. You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include <linux/slab.h> #include "c2.h" #include "c2_vq.h" #define PBL_VIRT 1 #define PBL_PHYS 2 /* * Send all the PBL messages to convey the remainder of the PBL * Wait for the adapter's reply on the last one. * This is indicated by setting the MEM_PBL_COMPLETE in the flags. * * NOTE: vq_req is _not_ freed by this function. The VQ Host * Reply buffer _is_ freed by this function. */ static int send_pbl_messages(struct c2_dev *c2dev, __be32 stag_index, unsigned long va, u32 pbl_depth, struct c2_vq_req *vq_req, int pbl_type) { u32 pbe_count; /* amt that fits in a PBL msg */ u32 count; /* amt in this PBL MSG. */ struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */ struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */ int err, pbl_virt, pbl_index, i; switch (pbl_type) { case PBL_VIRT: pbl_virt = 1; break; case PBL_PHYS: pbl_virt = 0; break; default: return -EINVAL; break; } pbe_count = (c2dev->req_vq.msg_size - sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64); wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); if (!wr) { return -ENOMEM; } c2_wr_set_id(wr, CCWR_NSMR_PBL); /* * Only the last PBL message will generate a reply from the verbs, * so we set the context to 0 indicating there is no kernel verbs * handler blocked awaiting this reply. */ wr->hdr.context = 0; wr->rnic_handle = c2dev->adapter_handle; wr->stag_index = stag_index; /* already swapped */ wr->flags = 0; pbl_index = 0; while (pbl_depth) { count = min(pbe_count, pbl_depth); wr->addrs_length = cpu_to_be32(count); /* * If this is the last message, then reference the * vq request struct cuz we're gonna wait for a reply. * also make this PBL msg as the last one. */ if (count == pbl_depth) { /* * reference the request struct. dereferenced in the * int handler. */ vq_req_get(c2dev, vq_req); wr->flags = cpu_to_be32(MEM_PBL_COMPLETE); /* * This is the last PBL message. * Set the context to our VQ Request Object so we can * wait for the reply. */ wr->hdr.context = (unsigned long) vq_req; } /* * If pbl_virt is set then va is a virtual address * that describes a virtually contiguous memory * allocation. The wr needs the start of each virtual page * to be converted to the corresponding physical address * of the page. If pbl_virt is not set then va is an array * of physical addresses and there is no conversion to do. * Just fill in the wr with what is in the array. */ for (i = 0; i < count; i++) { if (pbl_virt) { va += PAGE_SIZE; } else { wr->paddrs[i] = cpu_to_be64(((u64 *)va)[pbl_index + i]); } } /* * Send WR to adapter */ err = vq_send_wr(c2dev, (union c2wr *) wr); if (err) { if (count <= pbe_count) { vq_req_put(c2dev, vq_req); } goto bail0; } pbl_depth -= count; pbl_index += count; } /* * Now wait for the reply... */ err = vq_wait_for_reply(c2dev, vq_req); if (err) { goto bail0; } /* * Process reply */ reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg; if (!reply) { err = -ENOMEM; goto bail0; } err = c2_errno(reply); vq_repbuf_free(c2dev, reply); bail0: kfree(wr); return err; } #define C2_PBL_MAX_DEPTH 131072 int c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list, int page_size, int pbl_depth, u32 length, u32 offset, u64 *va, enum c2_acf acf, struct c2_mr *mr) { struct c2_vq_req *vq_req; struct c2wr_nsmr_register_req *wr; struct c2wr_nsmr_register_rep *reply; u16 flags; int i, pbe_count, count; int err; if (!va || !length || !addr_list || !pbl_depth) return -EINTR; /* * Verify PBL depth is within rnic max */ if (pbl_depth > C2_PBL_MAX_DEPTH) { return -EINTR; } /* * allocate verbs request object */ vq_req = vq_req_alloc(c2dev); if (!vq_req) return -ENOMEM; wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL); if (!wr) { err = -ENOMEM; goto bail0; } /* * build the WR */ c2_wr_set_id(wr, CCWR_NSMR_REGISTER); wr->hdr.context = (unsigned long) vq_req; wr->rnic_handle = c2dev->adapter_handle; flags = (acf | MEM_VA_BASED | MEM_REMOTE); /* * compute how many pbes can fit in the message */ pbe_count = (c2dev->req_vq.msg_size - sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64); if (pbl_depth <= pbe_count) { flags |= MEM_PBL_COMPLETE; } wr->flags = cpu_to_be16(flags); wr->stag_key = 0; //stag_key; wr->va = cpu_to_be64(*va); wr->pd_id = mr->pd->pd_id; wr->pbe_size = cpu_to_be32(page_size); wr->length = cpu_to_be32(length); wr->pbl_depth = cpu_to_be32(pbl_depth); wr->fbo = cpu_to_be32(offset); count = min(pbl_depth, pbe_count); wr->addrs_length = cpu_to_be32(count); /* * fill out the PBL for this message */ for (i = 0; i < count; i++) { wr->paddrs[i] = cpu_to_be64(addr_list[i]); } /* * regerence the request struct */ vq_req_get(c2dev, vq_req); /* * send the WR to the adapter */ err = vq_send_wr(c2dev, (union c2wr *) wr); if (err) { vq_req_put(c2dev, vq_req); goto bail1; } /* * wait for reply from adapter */ err = vq_wait_for_reply(c2dev, vq_req); if (err) { goto bail1; } /* * process reply */ reply = (struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg); if (!reply) { err = -ENOMEM; goto bail1; } if ((err = c2_errno(reply))) { goto bail2; } //*p_pb_entries = be32_to_cpu(reply->pbl_depth); mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index); vq_repbuf_free(c2dev, reply); /* * if there are still more PBEs we need to send them to * the adapter and wait for a reply on the final one. * reuse vq_req for this purpose. */ pbl_depth -= count; if (pbl_depth) { vq_req->reply_msg = (unsigned long) NULL; atomic_set(&vq_req->reply_ready, 0); err = send_pbl_messages(c2dev, cpu_to_be32(mr->ibmr.lkey), (unsigned long) &addr_list[i], pbl_depth, vq_req, PBL_PHYS); if (err) { goto bail1; } } vq_req_free(c2dev, vq_req); kfree(wr); return err; bail2: vq_repbuf_free(c2dev, reply); bail1: kfree(wr); bail0: vq_req_free(c2dev, vq_req); return err; } int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index) { struct c2_vq_req *vq_req; /* verbs request object */ struct c2wr_stag_dealloc_req wr; /* work request */ struct c2wr_stag_dealloc_rep *reply; /* WR reply */ int err; /* * allocate verbs request object */ vq_req = vq_req_alloc(c2dev); if (!vq_req) { return -ENOMEM; } /* * Build the WR */ c2_wr_set_id(&wr, CCWR_STAG_DEALLOC); wr.hdr.context = (u64) (unsigned long) vq_req; wr.rnic_handle = c2dev->adapter_handle; wr.stag_index = cpu_to_be32(stag_index); /* * reference the request struct. dereferenced in the int handler. */ vq_req_get(c2dev, vq_req); /* * Send WR to adapter */ err = vq_send_wr(c2dev, (union c2wr *) & wr); if (err) { vq_req_put(c2dev, vq_req); goto bail0; } /* * Wait for reply from adapter */ err = vq_wait_for_reply(c2dev, vq_req); if (err) { goto bail0; } /* * Process reply */ reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg; if (!reply) { err = -ENOMEM; goto bail0; } err = c2_errno(reply); vq_repbuf_free(c2dev, reply); bail0: vq_req_free(c2dev, vq_req); return err; }