execlists_submit_ports(engine);
}
+static void execlist_cancel_port_requests(struct intel_engine_execlists *execlists)
+{
+ unsigned int i;
+
+ for (i = 0; i < ARRAY_SIZE(execlists->port); i++)
+ i915_gem_request_put(port_request(&execlists->port[i]));
+
+ memset(execlists->port, 0, sizeof(execlists->port));
+}
+
static void execlists_cancel_requests(struct intel_engine_cs *engine)
{
struct intel_engine_execlists * const execlists = &engine->execlists;
- struct execlist_port *port = execlists->port;
struct drm_i915_gem_request *rq, *rn;
struct rb_node *rb;
unsigned long flags;
- unsigned long n;
spin_lock_irqsave(&engine->timeline->lock, flags);
/* Cancel the requests on the HW and clear the ELSP tracker. */
- for (n = 0; n < ARRAY_SIZE(execlists->port); n++)
- i915_gem_request_put(port_request(&port[n]));
- memset(execlists->port, 0, sizeof(execlists->port));
+ execlist_cancel_port_requests(execlists);
/* Mark all executing requests as skipped. */
list_for_each_entry(rq, &engine->timeline->requests, link) {
/* Remaining _unready_ requests will be nop'ed when submitted */
+
execlists->queue = RB_ROOT;
execlists->first = NULL;
- GEM_BUG_ON(port_isset(&port[0]));
+ GEM_BUG_ON(port_isset(&execlists->port[0]));
/*
* The port is checked prior to scheduling a tasklet, but
struct drm_i915_gem_request *request)
{
struct intel_engine_execlists * const execlists = &engine->execlists;
- struct execlist_port *port = execlists->port;
struct drm_i915_gem_request *rq, *rn;
struct intel_context *ce;
unsigned long flags;
- unsigned int n;
spin_lock_irqsave(&engine->timeline->lock, flags);
* guessing the missed context-switch events by looking at what
* requests were completed.
*/
- for (n = 0; n < ARRAY_SIZE(execlists->port); n++)
- i915_gem_request_put(port_request(&port[n]));
- memset(execlists->port, 0, sizeof(execlists->port));
+ execlist_cancel_port_requests(execlists);
/* Push back any incomplete requests for replay after the reset. */
list_for_each_entry_safe_reverse(rq, rn,