296 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			296 lines
		
	
	
		
			7.9 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * SPDX-License-Identifier: MIT
 | |
|  *
 | |
|  * Copyright © 2016 Intel Corporation
 | |
|  */
 | |
| 
 | |
| #include <linux/dma-fence-array.h>
 | |
| #include <linux/dma-fence-chain.h>
 | |
| #include <linux/jiffies.h>
 | |
| 
 | |
| #include "gt/intel_engine.h"
 | |
| #include "gt/intel_rps.h"
 | |
| 
 | |
| #include "i915_gem_ioctls.h"
 | |
| #include "i915_gem_object.h"
 | |
| 
 | |
| static long
 | |
| i915_gem_object_wait_fence(struct dma_fence *fence,
 | |
| 			   unsigned int flags,
 | |
| 			   long timeout)
 | |
| {
 | |
| 	BUILD_BUG_ON(I915_WAIT_INTERRUPTIBLE != 0x1);
 | |
| 
 | |
| 	if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
 | |
| 		return timeout;
 | |
| 
 | |
| 	if (dma_fence_is_i915(fence))
 | |
| 		return i915_request_wait_timeout(to_request(fence), flags, timeout);
 | |
| 
 | |
| 	return dma_fence_wait_timeout(fence,
 | |
| 				      flags & I915_WAIT_INTERRUPTIBLE,
 | |
| 				      timeout);
 | |
| }
 | |
| 
 | |
| static void
 | |
| i915_gem_object_boost(struct dma_resv *resv, unsigned int flags)
 | |
| {
 | |
| 	struct dma_resv_iter cursor;
 | |
| 	struct dma_fence *fence;
 | |
| 
 | |
| 	/*
 | |
| 	 * Prescan all fences for potential boosting before we begin waiting.
 | |
| 	 *
 | |
| 	 * When we wait, we wait on outstanding fences serially. If the
 | |
| 	 * dma-resv contains a sequence such as 1:1, 1:2 instead of a reduced
 | |
| 	 * form 1:2, then as we look at each wait in turn we see that each
 | |
| 	 * request is currently executing and not worthy of boosting. But if
 | |
| 	 * we only happen to look at the final fence in the sequence (because
 | |
| 	 * of request coalescing or splitting between read/write arrays by
 | |
| 	 * the iterator), then we would boost. As such our decision to boost
 | |
| 	 * or not is delicately balanced on the order we wait on fences.
 | |
| 	 *
 | |
| 	 * So instead of looking for boosts sequentially, look for all boosts
 | |
| 	 * upfront and then wait on the outstanding fences.
 | |
| 	 */
 | |
| 
 | |
| 	dma_resv_iter_begin(&cursor, resv,
 | |
| 			    dma_resv_usage_rw(flags & I915_WAIT_ALL));
 | |
| 	dma_resv_for_each_fence_unlocked(&cursor, fence)
 | |
| 		if (dma_fence_is_i915(fence) &&
 | |
| 		    !i915_request_started(to_request(fence)))
 | |
| 			intel_rps_boost(to_request(fence));
 | |
| 	dma_resv_iter_end(&cursor);
 | |
| }
 | |
| 
 | |
| static long
 | |
| i915_gem_object_wait_reservation(struct dma_resv *resv,
 | |
| 				 unsigned int flags,
 | |
| 				 long timeout)
 | |
| {
 | |
| 	struct dma_resv_iter cursor;
 | |
| 	struct dma_fence *fence;
 | |
| 	long ret = timeout ?: 1;
 | |
| 
 | |
| 	i915_gem_object_boost(resv, flags);
 | |
| 
 | |
| 	dma_resv_iter_begin(&cursor, resv,
 | |
| 			    dma_resv_usage_rw(flags & I915_WAIT_ALL));
 | |
| 	dma_resv_for_each_fence_unlocked(&cursor, fence) {
 | |
| 		ret = i915_gem_object_wait_fence(fence, flags, timeout);
 | |
| 		if (ret <= 0)
 | |
| 			break;
 | |
| 
 | |
| 		if (timeout)
 | |
| 			timeout = ret;
 | |
| 	}
 | |
| 	dma_resv_iter_end(&cursor);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void fence_set_priority(struct dma_fence *fence,
 | |
| 			       const struct i915_sched_attr *attr)
 | |
| {
 | |
| 	struct i915_request *rq;
 | |
| 	struct intel_engine_cs *engine;
 | |
| 
 | |
| 	if (dma_fence_is_signaled(fence) || !dma_fence_is_i915(fence))
 | |
| 		return;
 | |
| 
 | |
| 	rq = to_request(fence);
 | |
| 	engine = rq->engine;
 | |
| 
 | |
| 	rcu_read_lock(); /* RCU serialisation for set-wedged protection */
 | |
| 	if (engine->sched_engine->schedule)
 | |
| 		engine->sched_engine->schedule(rq, attr);
 | |
| 	rcu_read_unlock();
 | |
| }
 | |
| 
 | |
| static inline bool __dma_fence_is_chain(const struct dma_fence *fence)
 | |
| {
 | |
| 	return fence->ops == &dma_fence_chain_ops;
 | |
| }
 | |
| 
 | |
| void i915_gem_fence_wait_priority(struct dma_fence *fence,
 | |
| 				  const struct i915_sched_attr *attr)
 | |
| {
 | |
| 	if (dma_fence_is_signaled(fence))
 | |
| 		return;
 | |
| 
 | |
| 	local_bh_disable();
 | |
| 
 | |
| 	/* Recurse once into a fence-array */
 | |
| 	if (dma_fence_is_array(fence)) {
 | |
| 		struct dma_fence_array *array = to_dma_fence_array(fence);
 | |
| 		int i;
 | |
| 
 | |
| 		for (i = 0; i < array->num_fences; i++)
 | |
| 			fence_set_priority(array->fences[i], attr);
 | |
| 	} else if (__dma_fence_is_chain(fence)) {
 | |
| 		struct dma_fence *iter;
 | |
| 
 | |
| 		/* The chain is ordered; if we boost the last, we boost all */
 | |
| 		dma_fence_chain_for_each(iter, fence) {
 | |
| 			fence_set_priority(to_dma_fence_chain(iter)->fence,
 | |
| 					   attr);
 | |
| 			break;
 | |
| 		}
 | |
| 		dma_fence_put(iter);
 | |
| 	} else {
 | |
| 		fence_set_priority(fence, attr);
 | |
| 	}
 | |
| 
 | |
| 	local_bh_enable(); /* kick the tasklets if queues were reprioritised */
 | |
| }
 | |
| 
 | |
| int
 | |
| i915_gem_object_wait_priority(struct drm_i915_gem_object *obj,
 | |
| 			      unsigned int flags,
 | |
| 			      const struct i915_sched_attr *attr)
 | |
| {
 | |
| 	struct dma_resv_iter cursor;
 | |
| 	struct dma_fence *fence;
 | |
| 
 | |
| 	dma_resv_iter_begin(&cursor, obj->base.resv,
 | |
| 			    dma_resv_usage_rw(flags & I915_WAIT_ALL));
 | |
| 	dma_resv_for_each_fence_unlocked(&cursor, fence)
 | |
| 		i915_gem_fence_wait_priority(fence, attr);
 | |
| 	dma_resv_iter_end(&cursor);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * Waits for rendering to the object to be completed
 | |
|  * @obj: i915 gem object
 | |
|  * @flags: how to wait (under a lock, for all rendering or just for writes etc)
 | |
|  * @timeout: how long to wait
 | |
|  */
 | |
| int
 | |
| i915_gem_object_wait(struct drm_i915_gem_object *obj,
 | |
| 		     unsigned int flags,
 | |
| 		     long timeout)
 | |
| {
 | |
| 	might_sleep();
 | |
| 	GEM_BUG_ON(timeout < 0);
 | |
| 
 | |
| 	timeout = i915_gem_object_wait_reservation(obj->base.resv,
 | |
| 						   flags, timeout);
 | |
| 
 | |
| 	if (timeout < 0)
 | |
| 		return timeout;
 | |
| 
 | |
| 	return !timeout ? -ETIME : 0;
 | |
| }
 | |
| 
 | |
| static inline unsigned long nsecs_to_jiffies_timeout(const u64 n)
 | |
| {
 | |
| 	/* nsecs_to_jiffies64() does not guard against overflow */
 | |
| 	if (NSEC_PER_SEC % HZ &&
 | |
| 	    div_u64(n, NSEC_PER_SEC) >= MAX_JIFFY_OFFSET / HZ)
 | |
| 		return MAX_JIFFY_OFFSET;
 | |
| 
 | |
| 	return min_t(u64, MAX_JIFFY_OFFSET, nsecs_to_jiffies64(n) + 1);
 | |
| }
 | |
| 
 | |
| static unsigned long to_wait_timeout(s64 timeout_ns)
 | |
| {
 | |
| 	if (timeout_ns < 0)
 | |
| 		return MAX_SCHEDULE_TIMEOUT;
 | |
| 
 | |
| 	if (timeout_ns == 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	return nsecs_to_jiffies_timeout(timeout_ns);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * i915_gem_wait_ioctl - implements DRM_IOCTL_I915_GEM_WAIT
 | |
|  * @dev: drm device pointer
 | |
|  * @data: ioctl data blob
 | |
|  * @file: drm file pointer
 | |
|  *
 | |
|  * Returns 0 if successful, else an error is returned with the remaining time in
 | |
|  * the timeout parameter.
 | |
|  *  -ETIME: object is still busy after timeout
 | |
|  *  -ERESTARTSYS: signal interrupted the wait
 | |
|  *  -ENONENT: object doesn't exist
 | |
|  * Also possible, but rare:
 | |
|  *  -EAGAIN: incomplete, restart syscall
 | |
|  *  -ENOMEM: damn
 | |
|  *  -ENODEV: Internal IRQ fail
 | |
|  *  -E?: The add request failed
 | |
|  *
 | |
|  * The wait ioctl with a timeout of 0 reimplements the busy ioctl. With any
 | |
|  * non-zero timeout parameter the wait ioctl will wait for the given number of
 | |
|  * nanoseconds on an object becoming unbusy. Since the wait itself does so
 | |
|  * without holding struct_mutex the object may become re-busied before this
 | |
|  * function completes. A similar but shorter * race condition exists in the busy
 | |
|  * ioctl
 | |
|  */
 | |
| int
 | |
| i915_gem_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file)
 | |
| {
 | |
| 	struct drm_i915_gem_wait *args = data;
 | |
| 	struct drm_i915_gem_object *obj;
 | |
| 	ktime_t start;
 | |
| 	long ret;
 | |
| 
 | |
| 	if (args->flags != 0)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	obj = i915_gem_object_lookup(file, args->bo_handle);
 | |
| 	if (!obj)
 | |
| 		return -ENOENT;
 | |
| 
 | |
| 	start = ktime_get();
 | |
| 
 | |
| 	ret = i915_gem_object_wait(obj,
 | |
| 				   I915_WAIT_INTERRUPTIBLE |
 | |
| 				   I915_WAIT_PRIORITY |
 | |
| 				   I915_WAIT_ALL,
 | |
| 				   to_wait_timeout(args->timeout_ns));
 | |
| 
 | |
| 	if (args->timeout_ns > 0) {
 | |
| 		args->timeout_ns -= ktime_to_ns(ktime_sub(ktime_get(), start));
 | |
| 		if (args->timeout_ns < 0)
 | |
| 			args->timeout_ns = 0;
 | |
| 
 | |
| 		/*
 | |
| 		 * Apparently ktime isn't accurate enough and occasionally has a
 | |
| 		 * bit of mismatch in the jiffies<->nsecs<->ktime loop. So patch
 | |
| 		 * things up to make the test happy. We allow up to 1 jiffy.
 | |
| 		 *
 | |
| 		 * This is a regression from the timespec->ktime conversion.
 | |
| 		 */
 | |
| 		if (ret == -ETIME && !nsecs_to_jiffies(args->timeout_ns))
 | |
| 			args->timeout_ns = 0;
 | |
| 
 | |
| 		/* Asked to wait beyond the jiffie/scheduler precision? */
 | |
| 		if (ret == -ETIME && args->timeout_ns)
 | |
| 			ret = -EAGAIN;
 | |
| 	}
 | |
| 
 | |
| 	i915_gem_object_put(obj);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * i915_gem_object_wait_migration - Sync an accelerated migration operation
 | |
|  * @obj: The migrating object.
 | |
|  * @flags: waiting flags. Currently supports only I915_WAIT_INTERRUPTIBLE.
 | |
|  *
 | |
|  * Wait for any pending async migration operation on the object,
 | |
|  * whether it's explicitly (i915_gem_object_migrate()) or implicitly
 | |
|  * (swapin, initial clearing) initiated.
 | |
|  *
 | |
|  * Return: 0 if successful, -ERESTARTSYS if a signal was hit during waiting.
 | |
|  */
 | |
| int i915_gem_object_wait_migration(struct drm_i915_gem_object *obj,
 | |
| 				   unsigned int flags)
 | |
| {
 | |
| 	might_sleep();
 | |
| 
 | |
| 	return i915_gem_object_wait_moving_fence(obj, !!(flags & I915_WAIT_INTERRUPTIBLE));
 | |
| }
 |