1 /* 2 * This file is provided under a dual BSD/GPLv2 license. When using or 3 * redistributing this file, you may do so under either license. 4 * 5 * GPL LICENSE SUMMARY 6 * 7 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved. 8 * 9 * This program is free software; you can redistribute it and/or modify 10 * it under the terms of version 2 of the GNU General Public License as 11 * published by the Free Software Foundation. 12 * 13 * This program is distributed in the hope that it will be useful, but 14 * WITHOUT ANY WARRANTY; without even the implied warranty of 15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 16 * General Public License for more details. 17 * 18 * You should have received a copy of the GNU General Public License 19 * along with this program; if not, write to the Free Software 20 * Foundation, Inc., 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA. 21 * The full GNU General Public License is included in this distribution 22 * in the file called LICENSE.GPL. 23 * 24 * BSD LICENSE 25 * 26 * Copyright(c) 2008 - 2011 Intel Corporation. All rights reserved. 27 * All rights reserved. 28 * 29 * Redistribution and use in source and binary forms, with or without 30 * modification, are permitted provided that the following conditions 31 * are met: 32 * 33 * * Redistributions of source code must retain the above copyright 34 * notice, this list of conditions and the following disclaimer. 35 * * Redistributions in binary form must reproduce the above copyright 36 * notice, this list of conditions and the following disclaimer in 37 * the documentation and/or other materials provided with the 38 * distribution. 39 * * Neither the name of Intel Corporation nor the names of its 40 * contributors may be used to endorse or promote products derived 41 * from this software without specific prior written permission. 42 * 43 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 44 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 45 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 46 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 47 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 48 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 49 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 50 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 51 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 52 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 53 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 54 */ 55 56 #ifndef _ISCI_REQUEST_H_ 57 #define _ISCI_REQUEST_H_ 58 59 #include "isci.h" 60 #include "host.h" 61 #include "scu_task_context.h" 62 63 /** 64 * struct isci_request_status - This enum defines the possible states of an I/O 65 * request. 66 * 67 * 68 */ 69 enum isci_request_status { 70 unallocated = 0x00, 71 allocated = 0x01, 72 started = 0x02, 73 completed = 0x03, 74 aborting = 0x04, 75 aborted = 0x05, 76 terminating = 0x06, 77 dead = 0x07 78 }; 79 80 enum task_type { 81 io_task = 0, 82 tmf_task = 1 83 }; 84 85 enum sci_request_protocol { 86 SCIC_NO_PROTOCOL, 87 SCIC_SMP_PROTOCOL, 88 SCIC_SSP_PROTOCOL, 89 SCIC_STP_PROTOCOL 90 }; /* XXX remove me, use sas_task.{dev|task_proto} instead */; 91 92 /** 93 * isci_stp_request - extra request infrastructure to handle pio/atapi protocol 94 * @pio_len - number of bytes requested at PIO setup 95 * @status - pio setup ending status value to tell us if we need 96 * to wait for another fis or if the transfer is complete. Upon 97 * receipt of a d2h fis this will be the status field of that fis. 98 * @sgl - track pio transfer progress as we iterate through the sgl 99 * @device_cdb_len - atapi device advertises it's transfer constraints at setup 100 */ 101 struct isci_stp_request { 102 u32 pio_len; 103 u8 status; 104 105 struct isci_stp_pio_sgl { 106 int index; 107 u8 set; 108 u32 offset; 109 } sgl; 110 u32 device_cdb_len; 111 }; 112 113 struct isci_request { 114 enum isci_request_status status; 115 #define IREQ_COMPLETE_IN_TARGET 0 116 #define IREQ_TERMINATED 1 117 #define IREQ_TMF 2 118 #define IREQ_ACTIVE 3 119 unsigned long flags; 120 /* XXX kill ttype and ttype_ptr, allocate full sas_task */ 121 enum task_type ttype; 122 union ttype_ptr_union { 123 struct sas_task *io_task_ptr; /* When ttype==io_task */ 124 struct isci_tmf *tmf_task_ptr; /* When ttype==tmf_task */ 125 } ttype_ptr; 126 struct isci_host *isci_host; 127 /* For use in the requests_to_{complete|abort} lists: */ 128 struct list_head completed_node; 129 /* For use in the reqs_in_process list: */ 130 struct list_head dev_node; 131 spinlock_t state_lock; 132 dma_addr_t request_daddr; 133 dma_addr_t zero_scatter_daddr; 134 unsigned int num_sg_entries; 135 /* Note: "io_request_completion" is completed in two different ways 136 * depending on whether this is a TMF or regular request. 137 * - TMF requests are completed in the thread that started them; 138 * - regular requests are completed in the request completion callback 139 * function. 140 * This difference in operation allows the aborter of a TMF request 141 * to be sure that once the TMF request completes, the I/O that the 142 * TMF was aborting is guaranteed to have completed. 143 * 144 * XXX kill io_request_completion 145 */ 146 struct completion *io_request_completion; 147 struct sci_base_state_machine sm; 148 struct isci_host *owning_controller; 149 struct isci_remote_device *target_device; 150 u16 io_tag; 151 enum sci_request_protocol protocol; 152 u32 scu_status; /* hardware result */ 153 u32 sci_status; /* upper layer disposition */ 154 u32 post_context; 155 struct scu_task_context *tc; 156 /* could be larger with sg chaining */ 157 #define SCU_SGL_SIZE ((SCI_MAX_SCATTER_GATHER_ELEMENTS + 1) / 2) 158 struct scu_sgl_element_pair sg_table[SCU_SGL_SIZE] __attribute__ ((aligned(32))); 159 /* This field is a pointer to the stored rx frame data. It is used in 160 * STP internal requests and SMP response frames. If this field is 161 * non-NULL the saved frame must be released on IO request completion. 162 */ 163 u32 saved_rx_frame_index; 164 165 union { 166 struct { 167 union { 168 struct ssp_cmd_iu cmd; 169 struct ssp_task_iu tmf; 170 }; 171 union { 172 struct ssp_response_iu rsp; 173 u8 rsp_buf[SSP_RESP_IU_MAX_SIZE]; 174 }; 175 } ssp; 176 struct { 177 struct smp_resp rsp; 178 } smp; 179 struct { 180 struct isci_stp_request req; 181 struct host_to_dev_fis cmd; 182 struct dev_to_host_fis rsp; 183 } stp; 184 }; 185 }; 186 187 static inline struct isci_request *to_ireq(struct isci_stp_request *stp_req) 188 { 189 struct isci_request *ireq; 190 191 ireq = container_of(stp_req, typeof(*ireq), stp.req); 192 return ireq; 193 } 194 195 /** 196 * enum sci_base_request_states - This enumeration depicts all the states for 197 * the common request state machine. 198 * 199 * 200 */ 201 enum sci_base_request_states { 202 /* 203 * Simply the initial state for the base request state machine. 204 */ 205 SCI_REQ_INIT, 206 207 /* 208 * This state indicates that the request has been constructed. 209 * This state is entered from the INITIAL state. 210 */ 211 SCI_REQ_CONSTRUCTED, 212 213 /* 214 * This state indicates that the request has been started. This state 215 * is entered from the CONSTRUCTED state. 216 */ 217 SCI_REQ_STARTED, 218 219 SCI_REQ_STP_UDMA_WAIT_TC_COMP, 220 SCI_REQ_STP_UDMA_WAIT_D2H, 221 222 SCI_REQ_STP_NON_DATA_WAIT_H2D, 223 SCI_REQ_STP_NON_DATA_WAIT_D2H, 224 225 SCI_REQ_STP_SOFT_RESET_WAIT_H2D_ASSERTED, 226 SCI_REQ_STP_SOFT_RESET_WAIT_H2D_DIAG, 227 SCI_REQ_STP_SOFT_RESET_WAIT_D2H, 228 229 /* 230 * While in this state the IO request object is waiting for the TC 231 * completion notification for the H2D Register FIS 232 */ 233 SCI_REQ_STP_PIO_WAIT_H2D, 234 235 /* 236 * While in this state the IO request object is waiting for either a 237 * PIO Setup FIS or a D2H register FIS. The type of frame received is 238 * based on the result of the prior frame and line conditions. 239 */ 240 SCI_REQ_STP_PIO_WAIT_FRAME, 241 242 /* 243 * While in this state the IO request object is waiting for a DATA 244 * frame from the device. 245 */ 246 SCI_REQ_STP_PIO_DATA_IN, 247 248 /* 249 * While in this state the IO request object is waiting to transmit 250 * the next data frame to the device. 251 */ 252 SCI_REQ_STP_PIO_DATA_OUT, 253 254 /* 255 * The AWAIT_TC_COMPLETION sub-state indicates that the started raw 256 * task management request is waiting for the transmission of the 257 * initial frame (i.e. command, task, etc.). 258 */ 259 SCI_REQ_TASK_WAIT_TC_COMP, 260 261 /* 262 * This sub-state indicates that the started task management request 263 * is waiting for the reception of an unsolicited frame 264 * (i.e. response IU). 265 */ 266 SCI_REQ_TASK_WAIT_TC_RESP, 267 268 /* 269 * This sub-state indicates that the started task management request 270 * is waiting for the reception of an unsolicited frame 271 * (i.e. response IU). 272 */ 273 SCI_REQ_SMP_WAIT_RESP, 274 275 /* 276 * The AWAIT_TC_COMPLETION sub-state indicates that the started SMP 277 * request is waiting for the transmission of the initial frame 278 * (i.e. command, task, etc.). 279 */ 280 SCI_REQ_SMP_WAIT_TC_COMP, 281 282 /* 283 * This state indicates that the request has completed. 284 * This state is entered from the STARTED state. This state is entered 285 * from the ABORTING state. 286 */ 287 SCI_REQ_COMPLETED, 288 289 /* 290 * This state indicates that the request is in the process of being 291 * terminated/aborted. 292 * This state is entered from the CONSTRUCTED state. 293 * This state is entered from the STARTED state. 294 */ 295 SCI_REQ_ABORTING, 296 297 /* 298 * Simply the final state for the base request state machine. 299 */ 300 SCI_REQ_FINAL, 301 }; 302 303 enum sci_status sci_request_start(struct isci_request *ireq); 304 enum sci_status sci_io_request_terminate(struct isci_request *ireq); 305 enum sci_status 306 sci_io_request_event_handler(struct isci_request *ireq, 307 u32 event_code); 308 enum sci_status 309 sci_io_request_frame_handler(struct isci_request *ireq, 310 u32 frame_index); 311 enum sci_status 312 sci_task_request_terminate(struct isci_request *ireq); 313 extern enum sci_status 314 sci_request_complete(struct isci_request *ireq); 315 extern enum sci_status 316 sci_io_request_tc_completion(struct isci_request *ireq, u32 code); 317 318 /* XXX open code in caller */ 319 static inline dma_addr_t 320 sci_io_request_get_dma_addr(struct isci_request *ireq, void *virt_addr) 321 { 322 323 char *requested_addr = (char *)virt_addr; 324 char *base_addr = (char *)ireq; 325 326 BUG_ON(requested_addr < base_addr); 327 BUG_ON((requested_addr - base_addr) >= sizeof(*ireq)); 328 329 return ireq->request_daddr + (requested_addr - base_addr); 330 } 331 332 /** 333 * isci_request_change_state() - This function sets the status of the request 334 * object. 335 * @request: This parameter points to the isci_request object 336 * @status: This Parameter is the new status of the object 337 * 338 */ 339 static inline enum isci_request_status 340 isci_request_change_state(struct isci_request *isci_request, 341 enum isci_request_status status) 342 { 343 enum isci_request_status old_state; 344 unsigned long flags; 345 346 dev_dbg(&isci_request->isci_host->pdev->dev, 347 "%s: isci_request = %p, state = 0x%x\n", 348 __func__, 349 isci_request, 350 status); 351 352 BUG_ON(isci_request == NULL); 353 354 spin_lock_irqsave(&isci_request->state_lock, flags); 355 old_state = isci_request->status; 356 isci_request->status = status; 357 spin_unlock_irqrestore(&isci_request->state_lock, flags); 358 359 return old_state; 360 } 361 362 /** 363 * isci_request_change_started_to_newstate() - This function sets the status of 364 * the request object. 365 * @request: This parameter points to the isci_request object 366 * @status: This Parameter is the new status of the object 367 * 368 * state previous to any change. 369 */ 370 static inline enum isci_request_status 371 isci_request_change_started_to_newstate(struct isci_request *isci_request, 372 struct completion *completion_ptr, 373 enum isci_request_status newstate) 374 { 375 enum isci_request_status old_state; 376 unsigned long flags; 377 378 spin_lock_irqsave(&isci_request->state_lock, flags); 379 380 old_state = isci_request->status; 381 382 if (old_state == started || old_state == aborting) { 383 BUG_ON(isci_request->io_request_completion != NULL); 384 385 isci_request->io_request_completion = completion_ptr; 386 isci_request->status = newstate; 387 } 388 389 spin_unlock_irqrestore(&isci_request->state_lock, flags); 390 391 dev_dbg(&isci_request->isci_host->pdev->dev, 392 "%s: isci_request = %p, old_state = 0x%x\n", 393 __func__, 394 isci_request, 395 old_state); 396 397 return old_state; 398 } 399 400 /** 401 * isci_request_change_started_to_aborted() - This function sets the status of 402 * the request object. 403 * @request: This parameter points to the isci_request object 404 * @completion_ptr: This parameter is saved as the kernel completion structure 405 * signalled when the old request completes. 406 * 407 * state previous to any change. 408 */ 409 static inline enum isci_request_status 410 isci_request_change_started_to_aborted(struct isci_request *isci_request, 411 struct completion *completion_ptr) 412 { 413 return isci_request_change_started_to_newstate(isci_request, 414 completion_ptr, 415 aborted); 416 } 417 418 #define isci_request_access_task(req) ((req)->ttype_ptr.io_task_ptr) 419 420 #define isci_request_access_tmf(req) ((req)->ttype_ptr.tmf_task_ptr) 421 422 struct isci_request *isci_tmf_request_from_tag(struct isci_host *ihost, 423 struct isci_tmf *isci_tmf, 424 u16 tag); 425 int isci_request_execute(struct isci_host *ihost, struct isci_remote_device *idev, 426 struct sas_task *task, u16 tag); 427 void isci_terminate_pending_requests(struct isci_host *ihost, 428 struct isci_remote_device *idev); 429 enum sci_status 430 sci_task_request_construct(struct isci_host *ihost, 431 struct isci_remote_device *idev, 432 u16 io_tag, 433 struct isci_request *ireq); 434 enum sci_status 435 sci_task_request_construct_ssp(struct isci_request *ireq); 436 enum sci_status 437 sci_task_request_construct_sata(struct isci_request *ireq); 438 void sci_smp_request_copy_response(struct isci_request *ireq); 439 440 static inline int isci_task_is_ncq_recovery(struct sas_task *task) 441 { 442 return (sas_protocol_ata(task->task_proto) && 443 task->ata_task.fis.command == ATA_CMD_READ_LOG_EXT && 444 task->ata_task.fis.lbal == ATA_LOG_SATA_NCQ); 445 446 } 447 448 #endif /* !defined(_ISCI_REQUEST_H_) */ 449