1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #ifndef S390_IO_SCH_H 3 #define S390_IO_SCH_H 4 5 #include <linux/types.h> 6 #include <asm/schid.h> 7 #include <asm/ccwdev.h> 8 #include <asm/irq.h> 9 #include "css.h" 10 #include "orb.h" 11 12 struct io_subchannel_dma_area { 13 struct ccw1 sense_ccw; /* static ccw for sense command */ 14 }; 15 16 struct io_subchannel_private { 17 union orb orb; /* operation request block */ 18 struct ccw_device *cdev;/* pointer to the child ccw device */ 19 struct { 20 unsigned int suspend:1; /* allow suspend */ 21 unsigned int prefetch:1;/* deny prefetch */ 22 unsigned int inter:1; /* suppress intermediate interrupts */ 23 } __packed options; 24 struct io_subchannel_dma_area *dma_area; 25 dma_addr_t dma_area_dma; 26 } __aligned(8); 27 28 #define to_io_private(n) ((struct io_subchannel_private *) \ 29 dev_get_drvdata(&(n)->dev)) 30 #define set_io_private(n, p) (dev_set_drvdata(&(n)->dev, p)) 31 32 static inline struct ccw_device *sch_get_cdev(struct subchannel *sch) 33 { 34 struct io_subchannel_private *priv = to_io_private(sch); 35 return priv ? priv->cdev : NULL; 36 } 37 38 static inline void sch_set_cdev(struct subchannel *sch, 39 struct ccw_device *cdev) 40 { 41 struct io_subchannel_private *priv = to_io_private(sch); 42 if (priv) 43 priv->cdev = cdev; 44 } 45 46 #define MAX_CIWS 8 47 48 /* 49 * Possible status values for a CCW request's I/O. 50 */ 51 enum io_status { 52 IO_DONE, 53 IO_RUNNING, 54 IO_STATUS_ERROR, 55 IO_PATH_ERROR, 56 IO_REJECTED, 57 IO_KILLED 58 }; 59 60 /** 61 * ccw_request - Internal CCW request. 62 * @cp: channel program to start 63 * @timeout: maximum allowable time in jiffies between start I/O and interrupt 64 * @maxretries: number of retries per I/O operation and path 65 * @lpm: mask of paths to use 66 * @check: optional callback that determines if results are final 67 * @filter: optional callback to adjust request status based on IRB data 68 * @callback: final callback 69 * @data: user-defined pointer passed to all callbacks 70 * @singlepath: if set, use only one path from @lpm per start I/O 71 * @cancel: non-zero if request was cancelled 72 * @done: non-zero if request was finished 73 * @mask: current path mask 74 * @retries: current number of retries 75 * @drc: delayed return code 76 */ 77 struct ccw_request { 78 struct ccw1 *cp; 79 unsigned long timeout; 80 u16 maxretries; 81 u8 lpm; 82 int (*check)(struct ccw_device *, void *); 83 enum io_status (*filter)(struct ccw_device *, void *, struct irb *, 84 enum io_status); 85 void (*callback)(struct ccw_device *, void *, int); 86 void *data; 87 unsigned int singlepath:1; 88 /* These fields are used internally. */ 89 unsigned int cancel:1; 90 unsigned int done:1; 91 u16 mask; 92 u16 retries; 93 int drc; 94 } __attribute__((packed)); 95 96 /* 97 * sense-id response buffer layout 98 */ 99 struct senseid { 100 /* common part */ 101 u8 reserved; /* always 0x'FF' */ 102 u16 cu_type; /* control unit type */ 103 u8 cu_model; /* control unit model */ 104 u16 dev_type; /* device type */ 105 u8 dev_model; /* device model */ 106 u8 unused; /* padding byte */ 107 /* extended part */ 108 struct ciw ciw[MAX_CIWS]; /* variable # of CIWs */ 109 } __attribute__ ((packed, aligned(4))); 110 111 enum cdev_todo { 112 CDEV_TODO_NOTHING, 113 CDEV_TODO_ENABLE_CMF, 114 CDEV_TODO_REBIND, 115 CDEV_TODO_REGISTER, 116 CDEV_TODO_UNREG, 117 CDEV_TODO_UNREG_EVAL, 118 }; 119 120 #define FAKE_CMD_IRB 1 121 #define FAKE_TM_IRB 2 122 123 struct ccw_device_dma_area { 124 struct senseid senseid; /* SenseID info */ 125 struct ccw1 iccws[2]; /* ccws for SNID/SID/SPGID commands */ 126 struct irb irb; /* device status */ 127 struct pgid pgid[8]; /* path group IDs per chpid*/ 128 }; 129 130 struct ccw_device_private { 131 struct ccw_device *cdev; 132 struct subchannel *sch; 133 int state; /* device state */ 134 atomic_t onoff; 135 struct ccw_dev_id dev_id; /* device id */ 136 struct ccw_request req; /* internal I/O request */ 137 int iretry; 138 u8 pgid_valid_mask; /* mask of valid PGIDs */ 139 u8 pgid_todo_mask; /* mask of PGIDs to be adjusted */ 140 u8 pgid_reset_mask; /* mask of PGIDs which were reset */ 141 u8 path_noirq_mask; /* mask of paths for which no irq was 142 received */ 143 u8 path_notoper_mask; /* mask of paths which were found 144 not operable */ 145 u8 path_gone_mask; /* mask of paths, that became unavailable */ 146 u8 path_new_mask; /* mask of paths, that became available */ 147 u8 path_broken_mask; /* mask of paths, which were found to be 148 unusable */ 149 struct { 150 unsigned int fast:1; /* post with "channel end" */ 151 unsigned int repall:1; /* report every interrupt status */ 152 unsigned int pgroup:1; /* do path grouping */ 153 unsigned int force:1; /* allow forced online */ 154 unsigned int mpath:1; /* do multipathing */ 155 } __attribute__ ((packed)) options; 156 struct { 157 unsigned int esid:1; /* Ext. SenseID supported by HW */ 158 unsigned int dosense:1; /* delayed SENSE required */ 159 unsigned int doverify:1; /* delayed path verification */ 160 unsigned int donotify:1; /* call notify function */ 161 unsigned int recog_done:1; /* dev. recog. complete */ 162 unsigned int fake_irb:2; /* deliver faked irb */ 163 unsigned int pgroup:1; /* pathgroup is set up */ 164 unsigned int mpath:1; /* multipathing is set up */ 165 unsigned int pgid_unknown:1;/* unknown pgid state */ 166 unsigned int initialized:1; /* set if initial reference held */ 167 } __attribute__((packed)) flags; 168 unsigned long intparm; /* user interruption parameter */ 169 struct qdio_irq *qdio_data; 170 int async_kill_io_rc; 171 struct work_struct todo_work; 172 enum cdev_todo todo; 173 wait_queue_head_t wait_q; 174 struct timer_list timer; 175 void *cmb; /* measurement information */ 176 struct list_head cmb_list; /* list of measured devices */ 177 u64 cmb_start_time; /* clock value of cmb reset */ 178 void *cmb_wait; /* deferred cmb enable/disable */ 179 struct gen_pool *dma_pool; 180 struct ccw_device_dma_area *dma_area; 181 enum interruption_class int_class; 182 }; 183 184 #endif 185