xref: /openbmc/linux/drivers/s390/cio/vfio_ccw_fsm.c (revision 2243acd5)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Finite state machine for vfio-ccw device handling
4  *
5  * Copyright IBM Corp. 2017
6  * Copyright Red Hat, Inc. 2019
7  *
8  * Author(s): Dong Jia Shi <bjsdjshi@linux.vnet.ibm.com>
9  *            Cornelia Huck <cohuck@redhat.com>
10  */
11 
12 #include <linux/vfio.h>
13 
14 #include <asm/isc.h>
15 
16 #include "ioasm.h"
17 #include "vfio_ccw_private.h"
18 
19 static int fsm_io_helper(struct vfio_ccw_private *private)
20 {
21 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
22 	union orb *orb;
23 	int ccode;
24 	__u8 lpm;
25 	unsigned long flags;
26 	int ret;
27 
28 	spin_lock_irqsave(sch->lock, flags);
29 
30 	orb = cp_get_orb(&private->cp, sch);
31 	if (!orb) {
32 		ret = -EIO;
33 		goto out;
34 	}
35 
36 	VFIO_CCW_TRACE_EVENT(5, "stIO");
37 	VFIO_CCW_TRACE_EVENT(5, dev_name(&sch->dev));
38 
39 	/* Issue "Start Subchannel" */
40 	ccode = ssch(sch->schid, orb);
41 
42 	VFIO_CCW_HEX_EVENT(5, &ccode, sizeof(ccode));
43 
44 	switch (ccode) {
45 	case 0:
46 		/*
47 		 * Initialize device status information
48 		 */
49 		sch->schib.scsw.cmd.actl |= SCSW_ACTL_START_PEND;
50 		ret = 0;
51 		private->state = VFIO_CCW_STATE_CP_PENDING;
52 		break;
53 	case 1:		/* Status pending */
54 	case 2:		/* Busy */
55 		ret = -EBUSY;
56 		break;
57 	case 3:		/* Device/path not operational */
58 	{
59 		lpm = orb->cmd.lpm;
60 		if (lpm != 0)
61 			sch->lpm &= ~lpm;
62 		else
63 			sch->lpm = 0;
64 
65 		if (cio_update_schib(sch))
66 			ret = -ENODEV;
67 		else
68 			ret = sch->lpm ? -EACCES : -ENODEV;
69 		break;
70 	}
71 	default:
72 		ret = ccode;
73 	}
74 out:
75 	spin_unlock_irqrestore(sch->lock, flags);
76 	return ret;
77 }
78 
79 static int fsm_do_halt(struct vfio_ccw_private *private)
80 {
81 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
82 	unsigned long flags;
83 	int ccode;
84 	int ret;
85 
86 	spin_lock_irqsave(sch->lock, flags);
87 
88 	VFIO_CCW_TRACE_EVENT(2, "haltIO");
89 	VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
90 
91 	/* Issue "Halt Subchannel" */
92 	ccode = hsch(sch->schid);
93 
94 	VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
95 
96 	switch (ccode) {
97 	case 0:
98 		/*
99 		 * Initialize device status information
100 		 */
101 		sch->schib.scsw.cmd.actl |= SCSW_ACTL_HALT_PEND;
102 		ret = 0;
103 		break;
104 	case 1:		/* Status pending */
105 	case 2:		/* Busy */
106 		ret = -EBUSY;
107 		break;
108 	case 3:		/* Device not operational */
109 		ret = -ENODEV;
110 		break;
111 	default:
112 		ret = ccode;
113 	}
114 	spin_unlock_irqrestore(sch->lock, flags);
115 	return ret;
116 }
117 
118 static int fsm_do_clear(struct vfio_ccw_private *private)
119 {
120 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
121 	unsigned long flags;
122 	int ccode;
123 	int ret;
124 
125 	spin_lock_irqsave(sch->lock, flags);
126 
127 	VFIO_CCW_TRACE_EVENT(2, "clearIO");
128 	VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
129 
130 	/* Issue "Clear Subchannel" */
131 	ccode = csch(sch->schid);
132 
133 	VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
134 
135 	switch (ccode) {
136 	case 0:
137 		/*
138 		 * Initialize device status information
139 		 */
140 		sch->schib.scsw.cmd.actl = SCSW_ACTL_CLEAR_PEND;
141 		/* TODO: check what else we might need to clear */
142 		ret = 0;
143 		break;
144 	case 3:		/* Device not operational */
145 		ret = -ENODEV;
146 		break;
147 	default:
148 		ret = ccode;
149 	}
150 	spin_unlock_irqrestore(sch->lock, flags);
151 	return ret;
152 }
153 
154 static void fsm_notoper(struct vfio_ccw_private *private,
155 			enum vfio_ccw_event event)
156 {
157 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
158 
159 	VFIO_CCW_MSG_EVENT(2, "sch %x.%x.%04x: notoper event %x state %x\n",
160 			   sch->schid.cssid,
161 			   sch->schid.ssid,
162 			   sch->schid.sch_no,
163 			   event,
164 			   private->state);
165 
166 	/*
167 	 * TODO:
168 	 * Probably we should send the machine check to the guest.
169 	 */
170 	css_sched_sch_todo(sch, SCH_TODO_UNREG);
171 	private->state = VFIO_CCW_STATE_NOT_OPER;
172 
173 	/* This is usually handled during CLOSE event */
174 	cp_free(&private->cp);
175 }
176 
177 /*
178  * No operation action.
179  */
180 static void fsm_nop(struct vfio_ccw_private *private,
181 		    enum vfio_ccw_event event)
182 {
183 }
184 
185 static void fsm_io_error(struct vfio_ccw_private *private,
186 			 enum vfio_ccw_event event)
187 {
188 	pr_err("vfio-ccw: FSM: I/O request from state:%d\n", private->state);
189 	private->io_region->ret_code = -EIO;
190 }
191 
192 static void fsm_io_busy(struct vfio_ccw_private *private,
193 			enum vfio_ccw_event event)
194 {
195 	private->io_region->ret_code = -EBUSY;
196 }
197 
198 static void fsm_io_retry(struct vfio_ccw_private *private,
199 			 enum vfio_ccw_event event)
200 {
201 	private->io_region->ret_code = -EAGAIN;
202 }
203 
204 static void fsm_async_error(struct vfio_ccw_private *private,
205 			    enum vfio_ccw_event event)
206 {
207 	struct ccw_cmd_region *cmd_region = private->cmd_region;
208 
209 	pr_err("vfio-ccw: FSM: %s request from state:%d\n",
210 	       cmd_region->command == VFIO_CCW_ASYNC_CMD_HSCH ? "halt" :
211 	       cmd_region->command == VFIO_CCW_ASYNC_CMD_CSCH ? "clear" :
212 	       "<unknown>", private->state);
213 	cmd_region->ret_code = -EIO;
214 }
215 
216 static void fsm_async_retry(struct vfio_ccw_private *private,
217 			    enum vfio_ccw_event event)
218 {
219 	private->cmd_region->ret_code = -EAGAIN;
220 }
221 
222 static void fsm_disabled_irq(struct vfio_ccw_private *private,
223 			     enum vfio_ccw_event event)
224 {
225 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
226 
227 	/*
228 	 * An interrupt in a disabled state means a previous disable was not
229 	 * successful - should not happen, but we try to disable again.
230 	 */
231 	cio_disable_subchannel(sch);
232 }
233 inline struct subchannel_id get_schid(struct vfio_ccw_private *p)
234 {
235 	struct subchannel *sch = to_subchannel(p->vdev.dev->parent);
236 
237 	return sch->schid;
238 }
239 
240 /*
241  * Deal with the ccw command request from the userspace.
242  */
243 static void fsm_io_request(struct vfio_ccw_private *private,
244 			   enum vfio_ccw_event event)
245 {
246 	union orb *orb;
247 	union scsw *scsw = &private->scsw;
248 	struct ccw_io_region *io_region = private->io_region;
249 	char *errstr = "request";
250 	struct subchannel_id schid = get_schid(private);
251 
252 	private->state = VFIO_CCW_STATE_CP_PROCESSING;
253 	memcpy(scsw, io_region->scsw_area, sizeof(*scsw));
254 
255 	if (scsw->cmd.fctl & SCSW_FCTL_START_FUNC) {
256 		orb = (union orb *)io_region->orb_area;
257 
258 		/* Don't try to build a cp if transport mode is specified. */
259 		if (orb->tm.b) {
260 			io_region->ret_code = -EOPNOTSUPP;
261 			VFIO_CCW_MSG_EVENT(2,
262 					   "sch %x.%x.%04x: transport mode\n",
263 					   schid.cssid,
264 					   schid.ssid, schid.sch_no);
265 			errstr = "transport mode";
266 			goto err_out;
267 		}
268 		io_region->ret_code = cp_init(&private->cp, orb);
269 		if (io_region->ret_code) {
270 			VFIO_CCW_MSG_EVENT(2,
271 					   "sch %x.%x.%04x: cp_init=%d\n",
272 					   schid.cssid,
273 					   schid.ssid, schid.sch_no,
274 					   io_region->ret_code);
275 			errstr = "cp init";
276 			goto err_out;
277 		}
278 
279 		io_region->ret_code = cp_prefetch(&private->cp);
280 		if (io_region->ret_code) {
281 			VFIO_CCW_MSG_EVENT(2,
282 					   "sch %x.%x.%04x: cp_prefetch=%d\n",
283 					   schid.cssid,
284 					   schid.ssid, schid.sch_no,
285 					   io_region->ret_code);
286 			errstr = "cp prefetch";
287 			cp_free(&private->cp);
288 			goto err_out;
289 		}
290 
291 		/* Start channel program and wait for I/O interrupt. */
292 		io_region->ret_code = fsm_io_helper(private);
293 		if (io_region->ret_code) {
294 			VFIO_CCW_MSG_EVENT(2,
295 					   "sch %x.%x.%04x: fsm_io_helper=%d\n",
296 					   schid.cssid,
297 					   schid.ssid, schid.sch_no,
298 					   io_region->ret_code);
299 			errstr = "cp fsm_io_helper";
300 			cp_free(&private->cp);
301 			goto err_out;
302 		}
303 		return;
304 	} else if (scsw->cmd.fctl & SCSW_FCTL_HALT_FUNC) {
305 		VFIO_CCW_MSG_EVENT(2,
306 				   "sch %x.%x.%04x: halt on io_region\n",
307 				   schid.cssid,
308 				   schid.ssid, schid.sch_no);
309 		/* halt is handled via the async cmd region */
310 		io_region->ret_code = -EOPNOTSUPP;
311 		goto err_out;
312 	} else if (scsw->cmd.fctl & SCSW_FCTL_CLEAR_FUNC) {
313 		VFIO_CCW_MSG_EVENT(2,
314 				   "sch %x.%x.%04x: clear on io_region\n",
315 				   schid.cssid,
316 				   schid.ssid, schid.sch_no);
317 		/* clear is handled via the async cmd region */
318 		io_region->ret_code = -EOPNOTSUPP;
319 		goto err_out;
320 	}
321 
322 err_out:
323 	private->state = VFIO_CCW_STATE_IDLE;
324 	trace_vfio_ccw_fsm_io_request(scsw->cmd.fctl, schid,
325 				      io_region->ret_code, errstr);
326 }
327 
328 /*
329  * Deal with an async request from userspace.
330  */
331 static void fsm_async_request(struct vfio_ccw_private *private,
332 			      enum vfio_ccw_event event)
333 {
334 	struct ccw_cmd_region *cmd_region = private->cmd_region;
335 
336 	switch (cmd_region->command) {
337 	case VFIO_CCW_ASYNC_CMD_HSCH:
338 		cmd_region->ret_code = fsm_do_halt(private);
339 		break;
340 	case VFIO_CCW_ASYNC_CMD_CSCH:
341 		cmd_region->ret_code = fsm_do_clear(private);
342 		break;
343 	default:
344 		/* should not happen? */
345 		cmd_region->ret_code = -EINVAL;
346 	}
347 
348 	trace_vfio_ccw_fsm_async_request(get_schid(private),
349 					 cmd_region->command,
350 					 cmd_region->ret_code);
351 }
352 
353 /*
354  * Got an interrupt for a normal io (state busy).
355  */
356 static void fsm_irq(struct vfio_ccw_private *private,
357 		    enum vfio_ccw_event event)
358 {
359 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
360 	struct irb *irb = this_cpu_ptr(&cio_irb);
361 
362 	VFIO_CCW_TRACE_EVENT(6, "IRQ");
363 	VFIO_CCW_TRACE_EVENT(6, dev_name(&sch->dev));
364 
365 	memcpy(&private->irb, irb, sizeof(*irb));
366 
367 	queue_work(vfio_ccw_work_q, &private->io_work);
368 
369 	if (private->completion)
370 		complete(private->completion);
371 }
372 
373 static void fsm_open(struct vfio_ccw_private *private,
374 		     enum vfio_ccw_event event)
375 {
376 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
377 	int ret;
378 
379 	spin_lock_irq(sch->lock);
380 	sch->isc = VFIO_CCW_ISC;
381 	ret = cio_enable_subchannel(sch, (u32)(unsigned long)sch);
382 	if (ret)
383 		goto err_unlock;
384 
385 	private->state = VFIO_CCW_STATE_IDLE;
386 	spin_unlock_irq(sch->lock);
387 	return;
388 
389 err_unlock:
390 	spin_unlock_irq(sch->lock);
391 	vfio_ccw_fsm_event(private, VFIO_CCW_EVENT_NOT_OPER);
392 }
393 
394 static void fsm_close(struct vfio_ccw_private *private,
395 		      enum vfio_ccw_event event)
396 {
397 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
398 	int ret;
399 
400 	spin_lock_irq(sch->lock);
401 
402 	if (!sch->schib.pmcw.ena)
403 		goto err_unlock;
404 
405 	ret = cio_disable_subchannel(sch);
406 	if (ret == -EBUSY)
407 		ret = vfio_ccw_sch_quiesce(sch);
408 	if (ret)
409 		goto err_unlock;
410 
411 	private->state = VFIO_CCW_STATE_STANDBY;
412 	spin_unlock_irq(sch->lock);
413 	cp_free(&private->cp);
414 	return;
415 
416 err_unlock:
417 	spin_unlock_irq(sch->lock);
418 	vfio_ccw_fsm_event(private, VFIO_CCW_EVENT_NOT_OPER);
419 }
420 
421 /*
422  * Device statemachine
423  */
424 fsm_func_t *vfio_ccw_jumptable[NR_VFIO_CCW_STATES][NR_VFIO_CCW_EVENTS] = {
425 	[VFIO_CCW_STATE_NOT_OPER] = {
426 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_nop,
427 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_error,
428 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_error,
429 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_disabled_irq,
430 		[VFIO_CCW_EVENT_OPEN]		= fsm_nop,
431 		[VFIO_CCW_EVENT_CLOSE]		= fsm_nop,
432 	},
433 	[VFIO_CCW_STATE_STANDBY] = {
434 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
435 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_error,
436 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_error,
437 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_disabled_irq,
438 		[VFIO_CCW_EVENT_OPEN]		= fsm_open,
439 		[VFIO_CCW_EVENT_CLOSE]		= fsm_notoper,
440 	},
441 	[VFIO_CCW_STATE_IDLE] = {
442 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
443 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_request,
444 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_request,
445 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
446 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
447 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
448 	},
449 	[VFIO_CCW_STATE_CP_PROCESSING] = {
450 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
451 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_retry,
452 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_retry,
453 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
454 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
455 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
456 	},
457 	[VFIO_CCW_STATE_CP_PENDING] = {
458 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
459 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_busy,
460 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_request,
461 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
462 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
463 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
464 	},
465 };
466