diff options
author | Roland Dreier <roland@topspin.com> | 2005-04-17 02:26:13 +0400 |
---|---|---|
committer | Linus Torvalds <torvalds@ppc970.osdl.org> | 2005-04-17 02:26:13 +0400 |
commit | 86562a139182bb19c984347f9625b61f3e6f7815 (patch) | |
tree | 97984c9e04c31a5d2055e4bcac16bb11c923eca1 /drivers/infiniband/hw/mthca/mthca_mr.c | |
parent | 79b61dceafce696d72661d23a02393566b1899ab (diff) | |
download | linux-86562a139182bb19c984347f9625b61f3e6f7815.tar.xz |
[PATCH] IB/mthca: map MPT/MTT context in mem-free mode
In mem-free mode, when allocating memory regions, make sure that the HCA has
context memory mapped to cover the virtual space used for the MPT and MTTs
being used.
Signed-off-by: Roland Dreier <roland@topspin.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'drivers/infiniband/hw/mthca/mthca_mr.c')
-rw-r--r-- | drivers/infiniband/hw/mthca/mthca_mr.c | 79 |
1 files changed, 68 insertions, 11 deletions
diff --git a/drivers/infiniband/hw/mthca/mthca_mr.c b/drivers/infiniband/hw/mthca/mthca_mr.c index 80a0cd97881b..5eb6e07f35bb 100644 --- a/drivers/infiniband/hw/mthca/mthca_mr.c +++ b/drivers/infiniband/hw/mthca/mthca_mr.c @@ -38,6 +38,7 @@ #include "mthca_dev.h" #include "mthca_cmd.h" +#include "mthca_memfree.h" /* * Must be packed because mtt_seg is 64 bits but only aligned to 32 bits. @@ -71,7 +72,7 @@ struct mthca_mpt_entry { * through the bitmaps) */ -static u32 mthca_alloc_mtt(struct mthca_dev *dev, int order) +static u32 __mthca_alloc_mtt(struct mthca_dev *dev, int order) { int o; int m; @@ -105,7 +106,7 @@ static u32 mthca_alloc_mtt(struct mthca_dev *dev, int order) return seg; } -static void mthca_free_mtt(struct mthca_dev *dev, u32 seg, int order) +static void __mthca_free_mtt(struct mthca_dev *dev, u32 seg, int order) { seg >>= order; @@ -122,6 +123,32 @@ static void mthca_free_mtt(struct mthca_dev *dev, u32 seg, int order) spin_unlock(&dev->mr_table.mpt_alloc.lock); } +static u32 mthca_alloc_mtt(struct mthca_dev *dev, int order) +{ + u32 seg = __mthca_alloc_mtt(dev, order); + + if (seg == -1) + return -1; + + if (dev->hca_type == ARBEL_NATIVE) + if (mthca_table_get_range(dev, dev->mr_table.mtt_table, seg, + seg + (1 << order) - 1)) { + __mthca_free_mtt(dev, seg, order); + seg = -1; + } + + return seg; +} + +static void mthca_free_mtt(struct mthca_dev *dev, u32 seg, int order) +{ + __mthca_free_mtt(dev, seg, order); + + if (dev->hca_type == ARBEL_NATIVE) + mthca_table_put_range(dev, dev->mr_table.mtt_table, seg, + seg + (1 << order) - 1); +} + static inline u32 hw_index_to_key(struct mthca_dev *dev, u32 ind) { if (dev->hca_type == ARBEL_NATIVE) @@ -141,7 +168,7 @@ static inline u32 key_to_hw_index(struct mthca_dev *dev, u32 key) int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd, u32 access, struct mthca_mr *mr) { - void *mailbox; + void *mailbox = NULL; struct mthca_mpt_entry *mpt_entry; u32 key; int err; @@ -155,11 +182,17 @@ int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd, return -ENOMEM; mr->ibmr.rkey = mr->ibmr.lkey = hw_index_to_key(dev, key); + if (dev->hca_type == ARBEL_NATIVE) { + err = mthca_table_get(dev, dev->mr_table.mpt_table, key); + if (err) + goto err_out_mpt_free; + } + mailbox = kmalloc(sizeof *mpt_entry + MTHCA_CMD_MAILBOX_EXTRA, GFP_KERNEL); if (!mailbox) { - mthca_free(&dev->mr_table.mpt_alloc, mr->ibmr.lkey); - return -ENOMEM; + err = -ENOMEM; + goto err_out_table; } mpt_entry = MAILBOX_ALIGN(mailbox); @@ -180,16 +213,27 @@ int mthca_mr_alloc_notrans(struct mthca_dev *dev, u32 pd, err = mthca_SW2HW_MPT(dev, mpt_entry, key & (dev->limits.num_mpts - 1), &status); - if (err) + if (err) { mthca_warn(dev, "SW2HW_MPT failed (%d)\n", err); - else if (status) { + goto err_out_table; + } else if (status) { mthca_warn(dev, "SW2HW_MPT returned status 0x%02x\n", status); err = -EINVAL; + goto err_out_table; } kfree(mailbox); return err; + +err_out_table: + if (dev->hca_type == ARBEL_NATIVE) + mthca_table_put(dev, dev->mr_table.mpt_table, key); + +err_out_mpt_free: + mthca_free(&dev->mr_table.mpt_alloc, mr->ibmr.lkey); + kfree(mailbox); + return err; } int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, @@ -213,6 +257,12 @@ int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, return -ENOMEM; mr->ibmr.rkey = mr->ibmr.lkey = hw_index_to_key(dev, key); + if (dev->hca_type == ARBEL_NATIVE) { + err = mthca_table_get(dev, dev->mr_table.mpt_table, key); + if (err) + goto err_out_mpt_free; + } + for (i = dev->limits.mtt_seg_size / 8, mr->order = 0; i < list_len; i <<= 1, ++mr->order) @@ -220,7 +270,7 @@ int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, mr->first_seg = mthca_alloc_mtt(dev, mr->order); if (mr->first_seg == -1) - goto err_out_mpt_free; + goto err_out_table; /* * If list_len is odd, we add one more dummy entry for @@ -307,13 +357,17 @@ int mthca_mr_alloc_phys(struct mthca_dev *dev, u32 pd, kfree(mailbox); return err; - err_out_mailbox_free: +err_out_mailbox_free: kfree(mailbox); - err_out_free_mtt: +err_out_free_mtt: mthca_free_mtt(dev, mr->first_seg, mr->order); - err_out_mpt_free: +err_out_table: + if (dev->hca_type == ARBEL_NATIVE) + mthca_table_put(dev, dev->mr_table.mpt_table, key); + +err_out_mpt_free: mthca_free(&dev->mr_table.mpt_alloc, mr->ibmr.lkey); return err; } @@ -338,6 +392,9 @@ void mthca_free_mr(struct mthca_dev *dev, struct mthca_mr *mr) if (mr->order >= 0) mthca_free_mtt(dev, mr->first_seg, mr->order); + if (dev->hca_type == ARBEL_NATIVE) + mthca_table_put(dev, dev->mr_table.mpt_table, + key_to_hw_index(dev, mr->ibmr.lkey)); mthca_free(&dev->mr_table.mpt_alloc, key_to_hw_index(dev, mr->ibmr.lkey)); } |