466 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			466 lines
		
	
	
		
			11 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Finite state machine for vfio-ccw device handling
 | |
|  *
 | |
|  * Copyright IBM Corp. 2017
 | |
|  * Copyright Red Hat, Inc. 2019
 | |
|  *
 | |
|  * Author(s): Dong Jia Shi <bjsdjshi@linux.vnet.ibm.com>
 | |
|  *            Cornelia Huck <cohuck@redhat.com>
 | |
|  */
 | |
| 
 | |
| #include <linux/vfio.h>
 | |
| 
 | |
| #include <asm/isc.h>
 | |
| 
 | |
| #include "ioasm.h"
 | |
| #include "vfio_ccw_private.h"
 | |
| 
 | |
| static int fsm_io_helper(struct vfio_ccw_private *private)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	union orb *orb;
 | |
| 	int ccode;
 | |
| 	__u8 lpm;
 | |
| 	unsigned long flags;
 | |
| 	int ret;
 | |
| 
 | |
| 	spin_lock_irqsave(sch->lock, flags);
 | |
| 
 | |
| 	orb = cp_get_orb(&private->cp, sch);
 | |
| 	if (!orb) {
 | |
| 		ret = -EIO;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	VFIO_CCW_TRACE_EVENT(5, "stIO");
 | |
| 	VFIO_CCW_TRACE_EVENT(5, dev_name(&sch->dev));
 | |
| 
 | |
| 	/* Issue "Start Subchannel" */
 | |
| 	ccode = ssch(sch->schid, orb);
 | |
| 
 | |
| 	VFIO_CCW_HEX_EVENT(5, &ccode, sizeof(ccode));
 | |
| 
 | |
| 	switch (ccode) {
 | |
| 	case 0:
 | |
| 		/*
 | |
| 		 * Initialize device status information
 | |
| 		 */
 | |
| 		sch->schib.scsw.cmd.actl |= SCSW_ACTL_START_PEND;
 | |
| 		ret = 0;
 | |
| 		private->state = VFIO_CCW_STATE_CP_PENDING;
 | |
| 		break;
 | |
| 	case 1:		/* Status pending */
 | |
| 	case 2:		/* Busy */
 | |
| 		ret = -EBUSY;
 | |
| 		break;
 | |
| 	case 3:		/* Device/path not operational */
 | |
| 	{
 | |
| 		lpm = orb->cmd.lpm;
 | |
| 		if (lpm != 0)
 | |
| 			sch->lpm &= ~lpm;
 | |
| 		else
 | |
| 			sch->lpm = 0;
 | |
| 
 | |
| 		if (cio_update_schib(sch))
 | |
| 			ret = -ENODEV;
 | |
| 		else
 | |
| 			ret = sch->lpm ? -EACCES : -ENODEV;
 | |
| 		break;
 | |
| 	}
 | |
| 	default:
 | |
| 		ret = ccode;
 | |
| 	}
 | |
| out:
 | |
| 	spin_unlock_irqrestore(sch->lock, flags);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int fsm_do_halt(struct vfio_ccw_private *private)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	unsigned long flags;
 | |
| 	int ccode;
 | |
| 	int ret;
 | |
| 
 | |
| 	spin_lock_irqsave(sch->lock, flags);
 | |
| 
 | |
| 	VFIO_CCW_TRACE_EVENT(2, "haltIO");
 | |
| 	VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
 | |
| 
 | |
| 	/* Issue "Halt Subchannel" */
 | |
| 	ccode = hsch(sch->schid);
 | |
| 
 | |
| 	VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
 | |
| 
 | |
| 	switch (ccode) {
 | |
| 	case 0:
 | |
| 		/*
 | |
| 		 * Initialize device status information
 | |
| 		 */
 | |
| 		sch->schib.scsw.cmd.actl |= SCSW_ACTL_HALT_PEND;
 | |
| 		ret = 0;
 | |
| 		break;
 | |
| 	case 1:		/* Status pending */
 | |
| 	case 2:		/* Busy */
 | |
| 		ret = -EBUSY;
 | |
| 		break;
 | |
| 	case 3:		/* Device not operational */
 | |
| 		ret = -ENODEV;
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = ccode;
 | |
| 	}
 | |
| 	spin_unlock_irqrestore(sch->lock, flags);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static int fsm_do_clear(struct vfio_ccw_private *private)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	unsigned long flags;
 | |
| 	int ccode;
 | |
| 	int ret;
 | |
| 
 | |
| 	spin_lock_irqsave(sch->lock, flags);
 | |
| 
 | |
| 	VFIO_CCW_TRACE_EVENT(2, "clearIO");
 | |
| 	VFIO_CCW_TRACE_EVENT(2, dev_name(&sch->dev));
 | |
| 
 | |
| 	/* Issue "Clear Subchannel" */
 | |
| 	ccode = csch(sch->schid);
 | |
| 
 | |
| 	VFIO_CCW_HEX_EVENT(2, &ccode, sizeof(ccode));
 | |
| 
 | |
| 	switch (ccode) {
 | |
| 	case 0:
 | |
| 		/*
 | |
| 		 * Initialize device status information
 | |
| 		 */
 | |
| 		sch->schib.scsw.cmd.actl = SCSW_ACTL_CLEAR_PEND;
 | |
| 		/* TODO: check what else we might need to clear */
 | |
| 		ret = 0;
 | |
| 		break;
 | |
| 	case 3:		/* Device not operational */
 | |
| 		ret = -ENODEV;
 | |
| 		break;
 | |
| 	default:
 | |
| 		ret = ccode;
 | |
| 	}
 | |
| 	spin_unlock_irqrestore(sch->lock, flags);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void fsm_notoper(struct vfio_ccw_private *private,
 | |
| 			enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 
 | |
| 	VFIO_CCW_MSG_EVENT(2, "sch %x.%x.%04x: notoper event %x state %x\n",
 | |
| 			   sch->schid.cssid,
 | |
| 			   sch->schid.ssid,
 | |
| 			   sch->schid.sch_no,
 | |
| 			   event,
 | |
| 			   private->state);
 | |
| 
 | |
| 	/*
 | |
| 	 * TODO:
 | |
| 	 * Probably we should send the machine check to the guest.
 | |
| 	 */
 | |
| 	css_sched_sch_todo(sch, SCH_TODO_UNREG);
 | |
| 	private->state = VFIO_CCW_STATE_NOT_OPER;
 | |
| 
 | |
| 	/* This is usually handled during CLOSE event */
 | |
| 	cp_free(&private->cp);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * No operation action.
 | |
|  */
 | |
| static void fsm_nop(struct vfio_ccw_private *private,
 | |
| 		    enum vfio_ccw_event event)
 | |
| {
 | |
| }
 | |
| 
 | |
| static void fsm_io_error(struct vfio_ccw_private *private,
 | |
| 			 enum vfio_ccw_event event)
 | |
| {
 | |
| 	pr_err("vfio-ccw: FSM: I/O request from state:%d\n", private->state);
 | |
| 	private->io_region->ret_code = -EIO;
 | |
| }
 | |
| 
 | |
| static void fsm_io_busy(struct vfio_ccw_private *private,
 | |
| 			enum vfio_ccw_event event)
 | |
| {
 | |
| 	private->io_region->ret_code = -EBUSY;
 | |
| }
 | |
| 
 | |
| static void fsm_io_retry(struct vfio_ccw_private *private,
 | |
| 			 enum vfio_ccw_event event)
 | |
| {
 | |
| 	private->io_region->ret_code = -EAGAIN;
 | |
| }
 | |
| 
 | |
| static void fsm_async_error(struct vfio_ccw_private *private,
 | |
| 			    enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct ccw_cmd_region *cmd_region = private->cmd_region;
 | |
| 
 | |
| 	pr_err("vfio-ccw: FSM: %s request from state:%d\n",
 | |
| 	       cmd_region->command == VFIO_CCW_ASYNC_CMD_HSCH ? "halt" :
 | |
| 	       cmd_region->command == VFIO_CCW_ASYNC_CMD_CSCH ? "clear" :
 | |
| 	       "<unknown>", private->state);
 | |
| 	cmd_region->ret_code = -EIO;
 | |
| }
 | |
| 
 | |
| static void fsm_async_retry(struct vfio_ccw_private *private,
 | |
| 			    enum vfio_ccw_event event)
 | |
| {
 | |
| 	private->cmd_region->ret_code = -EAGAIN;
 | |
| }
 | |
| 
 | |
| static void fsm_disabled_irq(struct vfio_ccw_private *private,
 | |
| 			     enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 
 | |
| 	/*
 | |
| 	 * An interrupt in a disabled state means a previous disable was not
 | |
| 	 * successful - should not happen, but we try to disable again.
 | |
| 	 */
 | |
| 	cio_disable_subchannel(sch);
 | |
| }
 | |
| inline struct subchannel_id get_schid(struct vfio_ccw_private *p)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(p->vdev.dev->parent);
 | |
| 
 | |
| 	return sch->schid;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Deal with the ccw command request from the userspace.
 | |
|  */
 | |
| static void fsm_io_request(struct vfio_ccw_private *private,
 | |
| 			   enum vfio_ccw_event event)
 | |
| {
 | |
| 	union orb *orb;
 | |
| 	union scsw *scsw = &private->scsw;
 | |
| 	struct ccw_io_region *io_region = private->io_region;
 | |
| 	char *errstr = "request";
 | |
| 	struct subchannel_id schid = get_schid(private);
 | |
| 
 | |
| 	private->state = VFIO_CCW_STATE_CP_PROCESSING;
 | |
| 	memcpy(scsw, io_region->scsw_area, sizeof(*scsw));
 | |
| 
 | |
| 	if (scsw->cmd.fctl & SCSW_FCTL_START_FUNC) {
 | |
| 		orb = (union orb *)io_region->orb_area;
 | |
| 
 | |
| 		/* Don't try to build a cp if transport mode is specified. */
 | |
| 		if (orb->tm.b) {
 | |
| 			io_region->ret_code = -EOPNOTSUPP;
 | |
| 			VFIO_CCW_MSG_EVENT(2,
 | |
| 					   "sch %x.%x.%04x: transport mode\n",
 | |
| 					   schid.cssid,
 | |
| 					   schid.ssid, schid.sch_no);
 | |
| 			errstr = "transport mode";
 | |
| 			goto err_out;
 | |
| 		}
 | |
| 		io_region->ret_code = cp_init(&private->cp, orb);
 | |
| 		if (io_region->ret_code) {
 | |
| 			VFIO_CCW_MSG_EVENT(2,
 | |
| 					   "sch %x.%x.%04x: cp_init=%d\n",
 | |
| 					   schid.cssid,
 | |
| 					   schid.ssid, schid.sch_no,
 | |
| 					   io_region->ret_code);
 | |
| 			errstr = "cp init";
 | |
| 			goto err_out;
 | |
| 		}
 | |
| 
 | |
| 		io_region->ret_code = cp_prefetch(&private->cp);
 | |
| 		if (io_region->ret_code) {
 | |
| 			VFIO_CCW_MSG_EVENT(2,
 | |
| 					   "sch %x.%x.%04x: cp_prefetch=%d\n",
 | |
| 					   schid.cssid,
 | |
| 					   schid.ssid, schid.sch_no,
 | |
| 					   io_region->ret_code);
 | |
| 			errstr = "cp prefetch";
 | |
| 			cp_free(&private->cp);
 | |
| 			goto err_out;
 | |
| 		}
 | |
| 
 | |
| 		/* Start channel program and wait for I/O interrupt. */
 | |
| 		io_region->ret_code = fsm_io_helper(private);
 | |
| 		if (io_region->ret_code) {
 | |
| 			VFIO_CCW_MSG_EVENT(2,
 | |
| 					   "sch %x.%x.%04x: fsm_io_helper=%d\n",
 | |
| 					   schid.cssid,
 | |
| 					   schid.ssid, schid.sch_no,
 | |
| 					   io_region->ret_code);
 | |
| 			errstr = "cp fsm_io_helper";
 | |
| 			cp_free(&private->cp);
 | |
| 			goto err_out;
 | |
| 		}
 | |
| 		return;
 | |
| 	} else if (scsw->cmd.fctl & SCSW_FCTL_HALT_FUNC) {
 | |
| 		VFIO_CCW_MSG_EVENT(2,
 | |
| 				   "sch %x.%x.%04x: halt on io_region\n",
 | |
| 				   schid.cssid,
 | |
| 				   schid.ssid, schid.sch_no);
 | |
| 		/* halt is handled via the async cmd region */
 | |
| 		io_region->ret_code = -EOPNOTSUPP;
 | |
| 		goto err_out;
 | |
| 	} else if (scsw->cmd.fctl & SCSW_FCTL_CLEAR_FUNC) {
 | |
| 		VFIO_CCW_MSG_EVENT(2,
 | |
| 				   "sch %x.%x.%04x: clear on io_region\n",
 | |
| 				   schid.cssid,
 | |
| 				   schid.ssid, schid.sch_no);
 | |
| 		/* clear is handled via the async cmd region */
 | |
| 		io_region->ret_code = -EOPNOTSUPP;
 | |
| 		goto err_out;
 | |
| 	}
 | |
| 
 | |
| err_out:
 | |
| 	private->state = VFIO_CCW_STATE_IDLE;
 | |
| 	trace_vfio_ccw_fsm_io_request(scsw->cmd.fctl, schid,
 | |
| 				      io_region->ret_code, errstr);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Deal with an async request from userspace.
 | |
|  */
 | |
| static void fsm_async_request(struct vfio_ccw_private *private,
 | |
| 			      enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct ccw_cmd_region *cmd_region = private->cmd_region;
 | |
| 
 | |
| 	switch (cmd_region->command) {
 | |
| 	case VFIO_CCW_ASYNC_CMD_HSCH:
 | |
| 		cmd_region->ret_code = fsm_do_halt(private);
 | |
| 		break;
 | |
| 	case VFIO_CCW_ASYNC_CMD_CSCH:
 | |
| 		cmd_region->ret_code = fsm_do_clear(private);
 | |
| 		break;
 | |
| 	default:
 | |
| 		/* should not happen? */
 | |
| 		cmd_region->ret_code = -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	trace_vfio_ccw_fsm_async_request(get_schid(private),
 | |
| 					 cmd_region->command,
 | |
| 					 cmd_region->ret_code);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Got an interrupt for a normal io (state busy).
 | |
|  */
 | |
| static void fsm_irq(struct vfio_ccw_private *private,
 | |
| 		    enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	struct irb *irb = this_cpu_ptr(&cio_irb);
 | |
| 
 | |
| 	VFIO_CCW_TRACE_EVENT(6, "IRQ");
 | |
| 	VFIO_CCW_TRACE_EVENT(6, dev_name(&sch->dev));
 | |
| 
 | |
| 	memcpy(&private->irb, irb, sizeof(*irb));
 | |
| 
 | |
| 	queue_work(vfio_ccw_work_q, &private->io_work);
 | |
| 
 | |
| 	if (private->completion)
 | |
| 		complete(private->completion);
 | |
| }
 | |
| 
 | |
| static void fsm_open(struct vfio_ccw_private *private,
 | |
| 		     enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	int ret;
 | |
| 
 | |
| 	spin_lock_irq(sch->lock);
 | |
| 	sch->isc = VFIO_CCW_ISC;
 | |
| 	ret = cio_enable_subchannel(sch, (u32)virt_to_phys(sch));
 | |
| 	if (ret)
 | |
| 		goto err_unlock;
 | |
| 
 | |
| 	private->state = VFIO_CCW_STATE_IDLE;
 | |
| 	spin_unlock_irq(sch->lock);
 | |
| 	return;
 | |
| 
 | |
| err_unlock:
 | |
| 	spin_unlock_irq(sch->lock);
 | |
| 	vfio_ccw_fsm_event(private, VFIO_CCW_EVENT_NOT_OPER);
 | |
| }
 | |
| 
 | |
| static void fsm_close(struct vfio_ccw_private *private,
 | |
| 		      enum vfio_ccw_event event)
 | |
| {
 | |
| 	struct subchannel *sch = to_subchannel(private->vdev.dev->parent);
 | |
| 	int ret;
 | |
| 
 | |
| 	spin_lock_irq(sch->lock);
 | |
| 
 | |
| 	if (!sch->schib.pmcw.ena)
 | |
| 		goto err_unlock;
 | |
| 
 | |
| 	ret = cio_disable_subchannel(sch);
 | |
| 	if (ret == -EBUSY)
 | |
| 		ret = vfio_ccw_sch_quiesce(sch);
 | |
| 	if (ret)
 | |
| 		goto err_unlock;
 | |
| 
 | |
| 	private->state = VFIO_CCW_STATE_STANDBY;
 | |
| 	spin_unlock_irq(sch->lock);
 | |
| 	cp_free(&private->cp);
 | |
| 	return;
 | |
| 
 | |
| err_unlock:
 | |
| 	spin_unlock_irq(sch->lock);
 | |
| 	vfio_ccw_fsm_event(private, VFIO_CCW_EVENT_NOT_OPER);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Device statemachine
 | |
|  */
 | |
| fsm_func_t *vfio_ccw_jumptable[NR_VFIO_CCW_STATES][NR_VFIO_CCW_EVENTS] = {
 | |
| 	[VFIO_CCW_STATE_NOT_OPER] = {
 | |
| 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_nop,
 | |
| 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_error,
 | |
| 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_error,
 | |
| 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_disabled_irq,
 | |
| 		[VFIO_CCW_EVENT_OPEN]		= fsm_nop,
 | |
| 		[VFIO_CCW_EVENT_CLOSE]		= fsm_nop,
 | |
| 	},
 | |
| 	[VFIO_CCW_STATE_STANDBY] = {
 | |
| 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_error,
 | |
| 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_error,
 | |
| 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_disabled_irq,
 | |
| 		[VFIO_CCW_EVENT_OPEN]		= fsm_open,
 | |
| 		[VFIO_CCW_EVENT_CLOSE]		= fsm_notoper,
 | |
| 	},
 | |
| 	[VFIO_CCW_STATE_IDLE] = {
 | |
| 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_request,
 | |
| 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_request,
 | |
| 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
 | |
| 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
 | |
| 	},
 | |
| 	[VFIO_CCW_STATE_CP_PROCESSING] = {
 | |
| 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_retry,
 | |
| 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_retry,
 | |
| 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
 | |
| 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
 | |
| 	},
 | |
| 	[VFIO_CCW_STATE_CP_PENDING] = {
 | |
| 		[VFIO_CCW_EVENT_NOT_OPER]	= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_IO_REQ]		= fsm_io_busy,
 | |
| 		[VFIO_CCW_EVENT_ASYNC_REQ]	= fsm_async_request,
 | |
| 		[VFIO_CCW_EVENT_INTERRUPT]	= fsm_irq,
 | |
| 		[VFIO_CCW_EVENT_OPEN]		= fsm_notoper,
 | |
| 		[VFIO_CCW_EVENT_CLOSE]		= fsm_close,
 | |
| 	},
 | |
| };
 |