1 /******************************************************************* 2 * This file is part of the Emulex Linux Device Driver for * 3 * Fibre Channel Host Bus Adapters. * 4 * Copyright (C) 2004-2006 Emulex. All rights reserved. * 5 * EMULEX and SLI are trademarks of Emulex. * 6 * www.emulex.com * 7 * Portions Copyright (C) 2004-2005 Christoph Hellwig * 8 * * 9 * This program is free software; you can redistribute it and/or * 10 * modify it under the terms of version 2 of the GNU General * 11 * Public License as published by the Free Software Foundation. * 12 * This program is distributed in the hope that it will be useful. * 13 * ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND * 14 * WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, * 15 * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE * 16 * DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD * 17 * TO BE LEGALLY INVALID. See the GNU General Public License for * 18 * more details, a copy of which can be found in the file COPYING * 19 * included with this package. * 20 *******************************************************************/ 21 22 #include <linux/mempool.h> 23 #include <linux/pci.h> 24 #include <linux/interrupt.h> 25 26 #include <scsi/scsi_device.h> 27 #include <scsi/scsi_transport_fc.h> 28 29 #include <scsi/scsi.h> 30 31 #include "lpfc_hw.h" 32 #include "lpfc_sli.h" 33 #include "lpfc_disc.h" 34 #include "lpfc_scsi.h" 35 #include "lpfc.h" 36 #include "lpfc_crtn.h" 37 38 #define LPFC_MBUF_POOL_SIZE 64 /* max elements in MBUF safety pool */ 39 #define LPFC_MEM_POOL_SIZE 64 /* max elem in non-DMA safety pool */ 40 41 42 43 int 44 lpfc_mem_alloc(struct lpfc_hba * phba) 45 { 46 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; 47 int longs; 48 int i; 49 50 phba->lpfc_scsi_dma_buf_pool = pci_pool_create("lpfc_scsi_dma_buf_pool", 51 phba->pcidev, phba->cfg_sg_dma_buf_size, 8, 0); 52 if (!phba->lpfc_scsi_dma_buf_pool) 53 goto fail; 54 55 phba->lpfc_mbuf_pool = pci_pool_create("lpfc_mbuf_pool", phba->pcidev, 56 LPFC_BPL_SIZE, 8,0); 57 if (!phba->lpfc_mbuf_pool) 58 goto fail_free_dma_buf_pool; 59 60 pool->elements = kmalloc(sizeof(struct lpfc_dmabuf) * 61 LPFC_MBUF_POOL_SIZE, GFP_KERNEL); 62 if (!pool->elements) 63 goto fail_free_lpfc_mbuf_pool; 64 65 pool->max_count = 0; 66 pool->current_count = 0; 67 for ( i = 0; i < LPFC_MBUF_POOL_SIZE; i++) { 68 pool->elements[i].virt = pci_pool_alloc(phba->lpfc_mbuf_pool, 69 GFP_KERNEL, &pool->elements[i].phys); 70 if (!pool->elements[i].virt) 71 goto fail_free_mbuf_pool; 72 pool->max_count++; 73 pool->current_count++; 74 } 75 76 phba->mbox_mem_pool = mempool_create_kmalloc_pool(LPFC_MEM_POOL_SIZE, 77 sizeof(LPFC_MBOXQ_t)); 78 if (!phba->mbox_mem_pool) 79 goto fail_free_mbuf_pool; 80 81 phba->nlp_mem_pool = mempool_create_kmalloc_pool(LPFC_MEM_POOL_SIZE, 82 sizeof(struct lpfc_nodelist)); 83 if (!phba->nlp_mem_pool) 84 goto fail_free_mbox_pool; 85 86 phba->lpfc_hbq_pool = pci_pool_create("lpfc_hbq_pool",phba->pcidev, 87 LPFC_BPL_SIZE, 8, 0); 88 if (!phba->lpfc_hbq_pool) 89 goto fail_free_nlp_mem_pool; 90 91 /* vpi zero is reserved for the physical port so add 1 to max */ 92 longs = ((phba->max_vpi + 1) + BITS_PER_LONG - 1) / BITS_PER_LONG; 93 phba->vpi_bmask = kzalloc(longs * sizeof(unsigned long), GFP_KERNEL); 94 if (!phba->vpi_bmask) 95 goto fail_free_hbq_pool; 96 97 return 0; 98 99 fail_free_hbq_pool: 100 lpfc_sli_hbqbuf_free_all(phba); 101 pci_pool_destroy(phba->lpfc_hbq_pool); 102 fail_free_nlp_mem_pool: 103 mempool_destroy(phba->nlp_mem_pool); 104 phba->nlp_mem_pool = NULL; 105 fail_free_mbox_pool: 106 mempool_destroy(phba->mbox_mem_pool); 107 phba->mbox_mem_pool = NULL; 108 fail_free_mbuf_pool: 109 while (i--) 110 pci_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, 111 pool->elements[i].phys); 112 kfree(pool->elements); 113 fail_free_lpfc_mbuf_pool: 114 pci_pool_destroy(phba->lpfc_mbuf_pool); 115 phba->lpfc_mbuf_pool = NULL; 116 fail_free_dma_buf_pool: 117 pci_pool_destroy(phba->lpfc_scsi_dma_buf_pool); 118 phba->lpfc_scsi_dma_buf_pool = NULL; 119 fail: 120 return -ENOMEM; 121 } 122 123 void 124 lpfc_mem_free(struct lpfc_hba * phba) 125 { 126 struct lpfc_sli *psli = &phba->sli; 127 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; 128 LPFC_MBOXQ_t *mbox, *next_mbox; 129 struct lpfc_dmabuf *mp; 130 int i; 131 132 kfree(phba->vpi_bmask); 133 lpfc_sli_hbqbuf_free_all(phba); 134 135 list_for_each_entry_safe(mbox, next_mbox, &psli->mboxq, list) { 136 mp = (struct lpfc_dmabuf *) (mbox->context1); 137 if (mp) { 138 lpfc_mbuf_free(phba, mp->virt, mp->phys); 139 kfree(mp); 140 } 141 list_del(&mbox->list); 142 mempool_free(mbox, phba->mbox_mem_pool); 143 } 144 list_for_each_entry_safe(mbox, next_mbox, &psli->mboxq_cmpl, list) { 145 mp = (struct lpfc_dmabuf *) (mbox->context1); 146 if (mp) { 147 lpfc_mbuf_free(phba, mp->virt, mp->phys); 148 kfree(mp); 149 } 150 list_del(&mbox->list); 151 mempool_free(mbox, phba->mbox_mem_pool); 152 } 153 154 psli->sli_flag &= ~LPFC_SLI_MBOX_ACTIVE; 155 if (psli->mbox_active) { 156 mbox = psli->mbox_active; 157 mp = (struct lpfc_dmabuf *) (mbox->context1); 158 if (mp) { 159 lpfc_mbuf_free(phba, mp->virt, mp->phys); 160 kfree(mp); 161 } 162 mempool_free(mbox, phba->mbox_mem_pool); 163 psli->mbox_active = NULL; 164 } 165 166 for (i = 0; i < pool->current_count; i++) 167 pci_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, 168 pool->elements[i].phys); 169 kfree(pool->elements); 170 171 pci_pool_destroy(phba->lpfc_hbq_pool); 172 mempool_destroy(phba->nlp_mem_pool); 173 mempool_destroy(phba->mbox_mem_pool); 174 175 pci_pool_destroy(phba->lpfc_scsi_dma_buf_pool); 176 pci_pool_destroy(phba->lpfc_mbuf_pool); 177 178 phba->lpfc_hbq_pool = NULL; 179 phba->nlp_mem_pool = NULL; 180 phba->mbox_mem_pool = NULL; 181 phba->lpfc_scsi_dma_buf_pool = NULL; 182 phba->lpfc_mbuf_pool = NULL; 183 184 /* Free the iocb lookup array */ 185 kfree(psli->iocbq_lookup); 186 psli->iocbq_lookup = NULL; 187 188 } 189 190 void * 191 lpfc_mbuf_alloc(struct lpfc_hba *phba, int mem_flags, dma_addr_t *handle) 192 { 193 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; 194 unsigned long iflags; 195 void *ret; 196 197 ret = pci_pool_alloc(phba->lpfc_mbuf_pool, GFP_KERNEL, handle); 198 199 spin_lock_irqsave(&phba->hbalock, iflags); 200 if (!ret && (mem_flags & MEM_PRI) && pool->current_count) { 201 pool->current_count--; 202 ret = pool->elements[pool->current_count].virt; 203 *handle = pool->elements[pool->current_count].phys; 204 } 205 spin_unlock_irqrestore(&phba->hbalock, iflags); 206 return ret; 207 } 208 209 void 210 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) 211 { 212 struct lpfc_dma_pool *pool = &phba->lpfc_mbuf_safety_pool; 213 214 if (pool->current_count < pool->max_count) { 215 pool->elements[pool->current_count].virt = virt; 216 pool->elements[pool->current_count].phys = dma; 217 pool->current_count++; 218 } else { 219 pci_pool_free(phba->lpfc_mbuf_pool, virt, dma); 220 } 221 return; 222 } 223 224 void 225 lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) 226 { 227 unsigned long iflags; 228 229 spin_lock_irqsave(&phba->hbalock, iflags); 230 __lpfc_mbuf_free(phba, virt, dma); 231 spin_unlock_irqrestore(&phba->hbalock, iflags); 232 return; 233 } 234 235 struct hbq_dmabuf * 236 lpfc_els_hbq_alloc(struct lpfc_hba *phba) 237 { 238 struct hbq_dmabuf *hbqbp; 239 240 hbqbp = kmalloc(sizeof(struct hbq_dmabuf), GFP_KERNEL); 241 if (!hbqbp) 242 return NULL; 243 244 hbqbp->dbuf.virt = pci_pool_alloc(phba->lpfc_hbq_pool, GFP_KERNEL, 245 &hbqbp->dbuf.phys); 246 if (!hbqbp->dbuf.virt) { 247 kfree(hbqbp); 248 return NULL; 249 } 250 hbqbp->size = LPFC_BPL_SIZE; 251 return hbqbp; 252 } 253 254 void 255 lpfc_els_hbq_free(struct lpfc_hba *phba, struct hbq_dmabuf *hbqbp) 256 { 257 pci_pool_free(phba->lpfc_hbq_pool, hbqbp->dbuf.virt, hbqbp->dbuf.phys); 258 kfree(hbqbp); 259 return; 260 } 261 262 /* This is ONLY called for the LPFC_ELS_HBQ */ 263 void 264 lpfc_in_buf_free(struct lpfc_hba *phba, struct lpfc_dmabuf *mp) 265 { 266 struct hbq_dmabuf *hbq_entry; 267 unsigned long flags; 268 269 if (!mp) 270 return; 271 272 if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED) { 273 /* Check whether HBQ is still in use */ 274 spin_lock_irqsave(&phba->hbalock, flags); 275 if (!phba->hbq_in_use) { 276 spin_unlock_irqrestore(&phba->hbalock, flags); 277 return; 278 } 279 hbq_entry = container_of(mp, struct hbq_dmabuf, dbuf); 280 list_del(&hbq_entry->dbuf.list); 281 if (hbq_entry->tag == -1) { 282 (phba->hbqs[LPFC_ELS_HBQ].hbq_free_buffer) 283 (phba, hbq_entry); 284 } else { 285 lpfc_sli_free_hbq(phba, hbq_entry); 286 } 287 spin_unlock_irqrestore(&phba->hbalock, flags); 288 } else { 289 lpfc_mbuf_free(phba, mp->virt, mp->phys); 290 kfree(mp); 291 } 292 return; 293 } 294