drm/i915: Deminish contribution of wait-boosting from clients
authorChris Wilson <chris@chris-wilson.co.uk>
Tue, 7 Apr 2015 15:20:32 +0000 (16:20 +0100)
committerDaniel Vetter <daniel.vetter@ffwll.ch>
Fri, 10 Apr 2015 06:56:02 +0000 (08:56 +0200)
With boosting for missed pageflips, we have a much stronger indication
of when we need to (temporarily) boost GPU frequency to ensure smooth
delivery of frames. So now only allow each client to perform one RPS boost
in each period of GPU activity due to stalling on results.

Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Deepak S <deepak.s@linux.intel.com>
Reviewed-by: Deepak S <deepak.s@linux.intel.com>
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
drivers/gpu/drm/i915/i915_debugfs.c
drivers/gpu/drm/i915/i915_drv.h
drivers/gpu/drm/i915/i915_gem.c
drivers/gpu/drm/i915/intel_drv.h
drivers/gpu/drm/i915/intel_pm.c

index 10ca5117fceeee42d11bb6a300f9d7f1bd77da85..9c23eec3277e2af5c8886495232ef0979943bc3e 100644 (file)
@@ -2239,6 +2239,44 @@ static int i915_ppgtt_info(struct seq_file *m, void *data)
        return 0;
 }
 
+static int i915_rps_boost_info(struct seq_file *m, void *data)
+{
+       struct drm_info_node *node = m->private;
+       struct drm_device *dev = node->minor->dev;
+       struct drm_i915_private *dev_priv = dev->dev_private;
+       struct drm_file *file;
+       int ret;
+
+       ret = mutex_lock_interruptible(&dev->struct_mutex);
+       if (ret)
+               return ret;
+
+       ret = mutex_lock_interruptible(&dev_priv->rps.hw_lock);
+       if (ret)
+               goto unlock;
+
+       list_for_each_entry_reverse(file, &dev->filelist, lhead) {
+               struct drm_i915_file_private *file_priv = file->driver_priv;
+               struct task_struct *task;
+
+               rcu_read_lock();
+               task = pid_task(file->pid, PIDTYPE_PID);
+               seq_printf(m, "%s [%d]: %d boosts%s\n",
+                          task ? task->comm : "<unknown>",
+                          task ? task->pid : -1,
+                          file_priv->rps_boosts,
+                          list_empty(&file_priv->rps_boost) ? "" : ", active");
+               rcu_read_unlock();
+       }
+       seq_printf(m, "Kernel boosts: %d\n", dev_priv->rps.boosts);
+
+       mutex_unlock(&dev_priv->rps.hw_lock);
+unlock:
+       mutex_unlock(&dev->struct_mutex);
+
+       return ret;
+}
+
 static int i915_llc(struct seq_file *m, void *data)
 {
        struct drm_info_node *node = m->private;
@@ -4704,6 +4742,7 @@ static const struct drm_info_list i915_debugfs_list[] = {
        {"i915_ddb_info", i915_ddb_info, 0},
        {"i915_sseu_status", i915_sseu_status, 0},
        {"i915_drrs_status", i915_drrs_status, 0},
+       {"i915_rps_boost_info", i915_rps_boost_info, 0},
 };
 #define I915_DEBUGFS_ENTRIES ARRAY_SIZE(i915_debugfs_list)
 
index b4501647795ea24cae015cf5976ab219c084fffc..3b6ddd74cd69716a99fa58862ddd28d581ea9f65 100644 (file)
@@ -1043,6 +1043,8 @@ struct intel_gen6_power_mgmt {
 
        bool enabled;
        struct delayed_work delayed_resume_work;
+       struct list_head clients;
+       unsigned boosts;
 
        /* manual wa residency calculations */
        struct intel_rps_ei up_ei, down_ei;
@@ -2190,12 +2192,13 @@ struct drm_i915_file_private {
        struct {
                spinlock_t lock;
                struct list_head request_list;
-               struct delayed_work idle_work;
        } mm;
        struct idr context_idr;
 
-       atomic_t rps_wait_boost;
-       struct  intel_engine_cs *bsd_ring;
+       struct list_head rps_boost;
+       struct intel_engine_cs *bsd_ring;
+
+       unsigned rps_boosts;
 };
 
 /*
index 976d27a191925d2191adb9de29a8ed90938602c9..4061b4ffe81504fe1d6080cfd37322142759d74d 100644 (file)
@@ -1181,14 +1181,6 @@ static bool missed_irq(struct drm_i915_private *dev_priv,
        return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings);
 }
 
-static bool can_wait_boost(struct drm_i915_file_private *file_priv)
-{
-       if (file_priv == NULL)
-               return true;
-
-       return !atomic_xchg(&file_priv->rps_wait_boost, true);
-}
-
 /**
  * __i915_wait_request - wait until execution of request has finished
  * @req: duh!
@@ -1230,13 +1222,8 @@ int __i915_wait_request(struct drm_i915_gem_request *req,
        timeout_expire = timeout ?
                jiffies + nsecs_to_jiffies_timeout((u64)*timeout) : 0;
 
-       if (INTEL_INFO(dev)->gen >= 6 && ring->id == RCS && can_wait_boost(file_priv)) {
-               gen6_rps_boost(dev_priv);
-               if (file_priv)
-                       mod_delayed_work(dev_priv->wq,
-                                        &file_priv->mm.idle_work,
-                                        msecs_to_jiffies(100));
-       }
+       if (ring->id == RCS && INTEL_INFO(dev)->gen >= 6)
+               gen6_rps_boost(dev_priv, file_priv);
 
        if (!irq_test_in_progress && WARN_ON(!ring->irq_get(ring)))
                return -ENODEV;
@@ -5043,8 +5030,6 @@ void i915_gem_release(struct drm_device *dev, struct drm_file *file)
 {
        struct drm_i915_file_private *file_priv = file->driver_priv;
 
-       cancel_delayed_work_sync(&file_priv->mm.idle_work);
-
        /* Clean up our request list when the client is going away, so that
         * later retire_requests won't dereference our soon-to-be-gone
         * file_priv.
@@ -5060,15 +5045,12 @@ void i915_gem_release(struct drm_device *dev, struct drm_file *file)
                request->file_priv = NULL;
        }
        spin_unlock(&file_priv->mm.lock);
-}
-
-static void
-i915_gem_file_idle_work_handler(struct work_struct *work)
-{
-       struct drm_i915_file_private *file_priv =
-               container_of(work, typeof(*file_priv), mm.idle_work.work);
 
-       atomic_set(&file_priv->rps_wait_boost, false);
+       if (!list_empty(&file_priv->rps_boost)) {
+               mutex_lock(&to_i915(dev)->rps.hw_lock);
+               list_del(&file_priv->rps_boost);
+               mutex_unlock(&to_i915(dev)->rps.hw_lock);
+       }
 }
 
 int i915_gem_open(struct drm_device *dev, struct drm_file *file)
@@ -5085,11 +5067,10 @@ int i915_gem_open(struct drm_device *dev, struct drm_file *file)
        file->driver_priv = file_priv;
        file_priv->dev_priv = dev->dev_private;
        file_priv->file = file;
+       INIT_LIST_HEAD(&file_priv->rps_boost);
 
        spin_lock_init(&file_priv->mm.lock);
        INIT_LIST_HEAD(&file_priv->mm.request_list);
-       INIT_DELAYED_WORK(&file_priv->mm.idle_work,
-                         i915_gem_file_idle_work_handler);
 
        ret = i915_gem_context_open(dev, file);
        if (ret)
index 4771d319ac7bc58cb087fa4a7bf4adb20745cf29..efa53d57dd37a90987bb0d4c6310adf9b54740d2 100644 (file)
@@ -1262,7 +1262,8 @@ void gen6_update_ring_freq(struct drm_device *dev);
 void gen6_rps_busy(struct drm_i915_private *dev_priv);
 void gen6_rps_reset_ei(struct drm_i915_private *dev_priv);
 void gen6_rps_idle(struct drm_i915_private *dev_priv);
-void gen6_rps_boost(struct drm_i915_private *dev_priv);
+void gen6_rps_boost(struct drm_i915_private *dev_priv,
+                   struct drm_i915_file_private *file_priv);
 void intel_queue_rps_boost_for_request(struct drm_device *dev,
                                       struct drm_i915_gem_request *rq);
 void ilk_wm_get_hw_state(struct drm_device *dev);
index acf1a318fda9000e9ec102784254310ed735ef2f..1ab9e897994aa5168572a821c5c8e14c08efa8f0 100644 (file)
@@ -4091,10 +4091,14 @@ void gen6_rps_idle(struct drm_i915_private *dev_priv)
                dev_priv->rps.last_adj = 0;
                I915_WRITE(GEN6_PMINTRMSK, 0xffffffff);
        }
+
+       while (!list_empty(&dev_priv->rps.clients))
+               list_del_init(dev_priv->rps.clients.next);
        mutex_unlock(&dev_priv->rps.hw_lock);
 }
 
-void gen6_rps_boost(struct drm_i915_private *dev_priv)
+void gen6_rps_boost(struct drm_i915_private *dev_priv,
+                   struct drm_i915_file_private *file_priv)
 {
        u32 val;
 
@@ -4102,9 +4106,16 @@ void gen6_rps_boost(struct drm_i915_private *dev_priv)
        val = dev_priv->rps.max_freq_softlimit;
        if (dev_priv->rps.enabled &&
            dev_priv->mm.busy &&
-           dev_priv->rps.cur_freq < val) {
+           dev_priv->rps.cur_freq < val &&
+           (file_priv == NULL || list_empty(&file_priv->rps_boost))) {
                intel_set_rps(dev_priv->dev, val);
                dev_priv->rps.last_adj = 0;
+
+               if (file_priv != NULL) {
+                       list_add(&file_priv->rps_boost, &dev_priv->rps.clients);
+                       file_priv->rps_boosts++;
+               } else
+                       dev_priv->rps.boosts++;
        }
        mutex_unlock(&dev_priv->rps.hw_lock);
 }
@@ -6782,7 +6793,7 @@ static void __intel_rps_boost_work(struct work_struct *work)
        struct request_boost *boost = container_of(work, struct request_boost, work);
 
        if (!i915_gem_request_completed(boost->rq, true))
-               gen6_rps_boost(to_i915(boost->rq->ring->dev));
+               gen6_rps_boost(to_i915(boost->rq->ring->dev), NULL);
 
        i915_gem_request_unreference__unlocked(boost->rq);
        kfree(boost);
@@ -6815,6 +6826,7 @@ void intel_pm_setup(struct drm_device *dev)
 
        INIT_DELAYED_WORK(&dev_priv->rps.delayed_resume_work,
                          intel_gen6_powersave_work);
+       INIT_LIST_HEAD(&dev_priv->rps.clients);
 
        dev_priv->pm.suspended = false;
 }