1 /* 2 * Copyright (C) 2010-2013 Bluecherry, LLC <http://www.bluecherrydvr.com> 3 * 4 * Original author: 5 * Ben Collins <bcollins@ubuntu.com> 6 * 7 * Additional work by: 8 * John Brooks <john.brooks@bluecherry.net> 9 * 10 * This program is free software; you can redistribute it and/or modify 11 * it under the terms of the GNU General Public License as published by 12 * the Free Software Foundation; either version 2 of the License, or 13 * (at your option) any later version. 14 * 15 * This program is distributed in the hope that it will be useful, 16 * but WITHOUT ANY WARRANTY; without even the implied warranty of 17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 18 * GNU General Public License for more details. 19 */ 20 21 #include <linux/kernel.h> 22 #include <linux/module.h> 23 #include <linux/slab.h> 24 25 #include "solo6x10.h" 26 27 static int multi_p2m; 28 module_param(multi_p2m, uint, 0644); 29 MODULE_PARM_DESC(multi_p2m, 30 "Use multiple P2M DMA channels (default: no, 6010-only)"); 31 32 static int desc_mode; 33 module_param(desc_mode, uint, 0644); 34 MODULE_PARM_DESC(desc_mode, 35 "Allow use of descriptor mode DMA (default: no, 6010-only)"); 36 37 int solo_p2m_dma(struct solo_dev *solo_dev, int wr, 38 void *sys_addr, u32 ext_addr, u32 size, 39 int repeat, u32 ext_size) 40 { 41 dma_addr_t dma_addr; 42 int ret; 43 44 if (WARN_ON_ONCE((unsigned long)sys_addr & 0x03)) 45 return -EINVAL; 46 if (WARN_ON_ONCE(!size)) 47 return -EINVAL; 48 49 dma_addr = pci_map_single(solo_dev->pdev, sys_addr, size, 50 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); 51 if (pci_dma_mapping_error(solo_dev->pdev, dma_addr)) 52 return -ENOMEM; 53 54 ret = solo_p2m_dma_t(solo_dev, wr, dma_addr, ext_addr, size, 55 repeat, ext_size); 56 57 pci_unmap_single(solo_dev->pdev, dma_addr, size, 58 wr ? PCI_DMA_TODEVICE : PCI_DMA_FROMDEVICE); 59 60 return ret; 61 } 62 63 /* Mutex must be held for p2m_id before calling this!! */ 64 int solo_p2m_dma_desc(struct solo_dev *solo_dev, 65 struct solo_p2m_desc *desc, dma_addr_t desc_dma, 66 int desc_cnt) 67 { 68 struct solo_p2m_dev *p2m_dev; 69 unsigned int timeout; 70 unsigned int config = 0; 71 int ret = 0; 72 unsigned int p2m_id = 0; 73 74 /* Get next ID. According to Softlogic, 6110 has problems on !=0 P2M */ 75 if (solo_dev->type != SOLO_DEV_6110 && multi_p2m) 76 p2m_id = atomic_inc_return(&solo_dev->p2m_count) % SOLO_NR_P2M; 77 78 p2m_dev = &solo_dev->p2m_dev[p2m_id]; 79 80 if (mutex_lock_interruptible(&p2m_dev->mutex)) 81 return -EINTR; 82 83 reinit_completion(&p2m_dev->completion); 84 p2m_dev->error = 0; 85 86 if (desc_cnt > 1 && solo_dev->type != SOLO_DEV_6110 && desc_mode) { 87 /* For 6010 with more than one desc, we can do a one-shot */ 88 p2m_dev->desc_count = p2m_dev->desc_idx = 0; 89 config = solo_reg_read(solo_dev, SOLO_P2M_CONFIG(p2m_id)); 90 91 solo_reg_write(solo_dev, SOLO_P2M_DES_ADR(p2m_id), desc_dma); 92 solo_reg_write(solo_dev, SOLO_P2M_DESC_ID(p2m_id), desc_cnt); 93 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(p2m_id), config | 94 SOLO_P2M_DESC_MODE); 95 } else { 96 /* For single descriptors and 6110, we need to run each desc */ 97 p2m_dev->desc_count = desc_cnt; 98 p2m_dev->desc_idx = 1; 99 p2m_dev->descs = desc; 100 101 solo_reg_write(solo_dev, SOLO_P2M_TAR_ADR(p2m_id), 102 desc[1].dma_addr); 103 solo_reg_write(solo_dev, SOLO_P2M_EXT_ADR(p2m_id), 104 desc[1].ext_addr); 105 solo_reg_write(solo_dev, SOLO_P2M_EXT_CFG(p2m_id), 106 desc[1].cfg); 107 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(p2m_id), 108 desc[1].ctrl); 109 } 110 111 timeout = wait_for_completion_timeout(&p2m_dev->completion, 112 solo_dev->p2m_jiffies); 113 114 if (WARN_ON_ONCE(p2m_dev->error)) 115 ret = -EIO; 116 else if (timeout == 0) { 117 solo_dev->p2m_timeouts++; 118 ret = -EAGAIN; 119 } 120 121 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(p2m_id), 0); 122 123 /* Don't write here for the no_desc_mode case, because config is 0. 124 * We can't test no_desc_mode again, it might race. */ 125 if (desc_cnt > 1 && solo_dev->type != SOLO_DEV_6110 && config) 126 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(p2m_id), config); 127 128 mutex_unlock(&p2m_dev->mutex); 129 130 return ret; 131 } 132 133 void solo_p2m_fill_desc(struct solo_p2m_desc *desc, int wr, 134 dma_addr_t dma_addr, u32 ext_addr, u32 size, 135 int repeat, u32 ext_size) 136 { 137 WARN_ON_ONCE(dma_addr & 0x03); 138 WARN_ON_ONCE(!size); 139 140 desc->cfg = SOLO_P2M_COPY_SIZE(size >> 2); 141 desc->ctrl = SOLO_P2M_BURST_SIZE(SOLO_P2M_BURST_256) | 142 (wr ? SOLO_P2M_WRITE : 0) | SOLO_P2M_TRANS_ON; 143 144 if (repeat) { 145 desc->cfg |= SOLO_P2M_EXT_INC(ext_size >> 2); 146 desc->ctrl |= SOLO_P2M_PCI_INC(size >> 2) | 147 SOLO_P2M_REPEAT(repeat); 148 } 149 150 desc->dma_addr = dma_addr; 151 desc->ext_addr = ext_addr; 152 } 153 154 int solo_p2m_dma_t(struct solo_dev *solo_dev, int wr, 155 dma_addr_t dma_addr, u32 ext_addr, u32 size, 156 int repeat, u32 ext_size) 157 { 158 struct solo_p2m_desc desc[2]; 159 160 solo_p2m_fill_desc(&desc[1], wr, dma_addr, ext_addr, size, repeat, 161 ext_size); 162 163 /* No need for desc_dma since we know it is a single-shot */ 164 return solo_p2m_dma_desc(solo_dev, desc, 0, 1); 165 } 166 167 void solo_p2m_isr(struct solo_dev *solo_dev, int id) 168 { 169 struct solo_p2m_dev *p2m_dev = &solo_dev->p2m_dev[id]; 170 struct solo_p2m_desc *desc; 171 172 if (p2m_dev->desc_count <= p2m_dev->desc_idx) { 173 complete(&p2m_dev->completion); 174 return; 175 } 176 177 /* Setup next descriptor */ 178 p2m_dev->desc_idx++; 179 desc = &p2m_dev->descs[p2m_dev->desc_idx]; 180 181 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(id), 0); 182 solo_reg_write(solo_dev, SOLO_P2M_TAR_ADR(id), desc->dma_addr); 183 solo_reg_write(solo_dev, SOLO_P2M_EXT_ADR(id), desc->ext_addr); 184 solo_reg_write(solo_dev, SOLO_P2M_EXT_CFG(id), desc->cfg); 185 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(id), desc->ctrl); 186 } 187 188 void solo_p2m_error_isr(struct solo_dev *solo_dev) 189 { 190 unsigned int err = solo_reg_read(solo_dev, SOLO_PCI_ERR); 191 struct solo_p2m_dev *p2m_dev; 192 int i; 193 194 if (!(err & (SOLO_PCI_ERR_P2M | SOLO_PCI_ERR_P2M_DESC))) 195 return; 196 197 for (i = 0; i < SOLO_NR_P2M; i++) { 198 p2m_dev = &solo_dev->p2m_dev[i]; 199 p2m_dev->error = 1; 200 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(i), 0); 201 complete(&p2m_dev->completion); 202 } 203 } 204 205 void solo_p2m_exit(struct solo_dev *solo_dev) 206 { 207 int i; 208 209 for (i = 0; i < SOLO_NR_P2M; i++) 210 solo_irq_off(solo_dev, SOLO_IRQ_P2M(i)); 211 } 212 213 static int solo_p2m_test(struct solo_dev *solo_dev, int base, int size) 214 { 215 u32 *wr_buf; 216 u32 *rd_buf; 217 int i; 218 int ret = -EIO; 219 int order = get_order(size); 220 221 wr_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); 222 if (wr_buf == NULL) 223 return -1; 224 225 rd_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); 226 if (rd_buf == NULL) { 227 free_pages((unsigned long)wr_buf, order); 228 return -1; 229 } 230 231 for (i = 0; i < (size >> 3); i++) 232 *(wr_buf + i) = (i << 16) | (i + 1); 233 234 for (i = (size >> 3); i < (size >> 2); i++) 235 *(wr_buf + i) = ~((i << 16) | (i + 1)); 236 237 memset(rd_buf, 0x55, size); 238 239 if (solo_p2m_dma(solo_dev, 1, wr_buf, base, size, 0, 0)) 240 goto test_fail; 241 242 if (solo_p2m_dma(solo_dev, 0, rd_buf, base, size, 0, 0)) 243 goto test_fail; 244 245 for (i = 0; i < (size >> 2); i++) { 246 if (*(wr_buf + i) != *(rd_buf + i)) 247 goto test_fail; 248 } 249 250 ret = 0; 251 252 test_fail: 253 free_pages((unsigned long)wr_buf, order); 254 free_pages((unsigned long)rd_buf, order); 255 256 return ret; 257 } 258 259 int solo_p2m_init(struct solo_dev *solo_dev) 260 { 261 struct solo_p2m_dev *p2m_dev; 262 int i; 263 264 for (i = 0; i < SOLO_NR_P2M; i++) { 265 p2m_dev = &solo_dev->p2m_dev[i]; 266 267 mutex_init(&p2m_dev->mutex); 268 init_completion(&p2m_dev->completion); 269 270 solo_reg_write(solo_dev, SOLO_P2M_CONTROL(i), 0); 271 solo_reg_write(solo_dev, SOLO_P2M_CONFIG(i), 272 SOLO_P2M_CSC_16BIT_565 | 273 SOLO_P2M_DESC_INTR_OPT | 274 SOLO_P2M_DMA_INTERVAL(0) | 275 SOLO_P2M_PCI_MASTER_MODE); 276 solo_irq_on(solo_dev, SOLO_IRQ_P2M(i)); 277 } 278 279 /* Find correct SDRAM size */ 280 for (solo_dev->sdram_size = 0, i = 2; i >= 0; i--) { 281 solo_reg_write(solo_dev, SOLO_DMA_CTRL, 282 SOLO_DMA_CTRL_REFRESH_CYCLE(1) | 283 SOLO_DMA_CTRL_SDRAM_SIZE(i) | 284 SOLO_DMA_CTRL_SDRAM_CLK_INVERT | 285 SOLO_DMA_CTRL_READ_CLK_SELECT | 286 SOLO_DMA_CTRL_LATENCY(1)); 287 288 solo_reg_write(solo_dev, SOLO_SYS_CFG, solo_dev->sys_config | 289 SOLO_SYS_CFG_RESET); 290 solo_reg_write(solo_dev, SOLO_SYS_CFG, solo_dev->sys_config); 291 292 switch (i) { 293 case 2: 294 if (solo_p2m_test(solo_dev, 0x07ff0000, 0x00010000) || 295 solo_p2m_test(solo_dev, 0x05ff0000, 0x00010000)) 296 continue; 297 break; 298 299 case 1: 300 if (solo_p2m_test(solo_dev, 0x03ff0000, 0x00010000)) 301 continue; 302 break; 303 304 default: 305 if (solo_p2m_test(solo_dev, 0x01ff0000, 0x00010000)) 306 continue; 307 } 308 309 solo_dev->sdram_size = (32 << 20) << i; 310 break; 311 } 312 313 if (!solo_dev->sdram_size) { 314 dev_err(&solo_dev->pdev->dev, "Error detecting SDRAM size\n"); 315 return -EIO; 316 } 317 318 if (SOLO_SDRAM_END(solo_dev) > solo_dev->sdram_size) { 319 dev_err(&solo_dev->pdev->dev, 320 "SDRAM is not large enough (%u < %u)\n", 321 solo_dev->sdram_size, SOLO_SDRAM_END(solo_dev)); 322 return -EIO; 323 } 324 325 return 0; 326 } 327