habanalabs: Rename hw_queues_mirror to cs_mirror
authorTomer Tayar <ttayar@habana.ai>
Fri, 30 Oct 2020 09:16:23 +0000 (11:16 +0200)
committerOded Gabbay <ogabbay@kernel.org>
Mon, 30 Nov 2020 08:47:34 +0000 (10:47 +0200)
Future command submission types might be submitted to HW not via the
QMAN queues path. However, it would be still required to have the TDR
mechanism for these CS, and thus the patch renames the TDR fields and
replaces the hw_queues_ prefix with cs_.

Signed-off-by: Tomer Tayar <ttayar@habana.ai>
Reviewed-by: Oded Gabbay <ogabbay@kernel.org>
Signed-off-by: Oded Gabbay <ogabbay@kernel.org>
drivers/misc/habanalabs/common/command_submission.c
drivers/misc/habanalabs/common/device.c
drivers/misc/habanalabs/common/habanalabs.h
drivers/misc/habanalabs/common/hw_queue.c

index 20b34fb054ee27d19ebbac6845eca4ba269adaf7..749ec6c95fc23d31ac09b614cc2d5bddb130d9a2 100644 (file)
@@ -343,25 +343,24 @@ static void cs_do_release(struct kref *ref)
        /* Need to update CI for internal queues */
        hl_int_hw_queue_update_ci(cs);
 
-       spin_lock(&hdev->hw_queues_mirror_lock);
-       /* remove CS from hw_queues mirror list */
+       /* remove CS from CS mirror list */
+       spin_lock(&hdev->cs_mirror_lock);
        list_del_init(&cs->mirror_node);
-       spin_unlock(&hdev->hw_queues_mirror_lock);
+       spin_unlock(&hdev->cs_mirror_lock);
 
        /* Don't cancel TDR in case this CS was timedout because we might be
         * running from the TDR context
         */
-       if (!cs->timedout &&
-                       hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) {
+       if (!cs->timedout && hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) {
                struct hl_cs *next;
 
                if (cs->tdr_active)
                        cancel_delayed_work_sync(&cs->work_tdr);
 
-               spin_lock(&hdev->hw_queues_mirror_lock);
+               spin_lock(&hdev->cs_mirror_lock);
 
                /* queue TDR for next CS */
-               next = list_first_entry_or_null(&hdev->hw_queues_mirror_list,
+               next = list_first_entry_or_null(&hdev->cs_mirror_list,
                                                struct hl_cs, mirror_node);
 
                if (next && !next->tdr_active) {
@@ -370,7 +369,7 @@ static void cs_do_release(struct kref *ref)
                                                hdev->timeout_jiffies);
                }
 
-               spin_unlock(&hdev->hw_queues_mirror_lock);
+               spin_unlock(&hdev->cs_mirror_lock);
        }
 
 out:
@@ -534,8 +533,7 @@ void hl_cs_rollback_all(struct hl_device *hdev)
                flush_workqueue(hdev->cq_wq[i]);
 
        /* Make sure we don't have leftovers in the H/W queues mirror list */
-       list_for_each_entry_safe(cs, tmp, &hdev->hw_queues_mirror_list,
-                               mirror_node) {
+       list_for_each_entry_safe(cs, tmp, &hdev->cs_mirror_list, mirror_node) {
                cs_get(cs);
                cs->aborted = true;
                dev_warn_ratelimited(hdev->dev, "Killing CS %d.%llu\n",
index 348faf31668bc980a7345879a53fca3a86d6c437..ce0a1270e5ff2c75e3ce495d6321a24601c5a57f 100644 (file)
@@ -343,8 +343,8 @@ static int device_early_init(struct hl_device *hdev)
        mutex_init(&hdev->send_cpu_message_lock);
        mutex_init(&hdev->debug_lock);
        mutex_init(&hdev->mmu_cache_lock);
-       INIT_LIST_HEAD(&hdev->hw_queues_mirror_list);
-       spin_lock_init(&hdev->hw_queues_mirror_lock);
+       INIT_LIST_HEAD(&hdev->cs_mirror_list);
+       spin_lock_init(&hdev->cs_mirror_lock);
        INIT_LIST_HEAD(&hdev->fpriv_list);
        mutex_init(&hdev->fpriv_list_lock);
        atomic_set(&hdev->in_reset, 0);
index 40b566c4b791a9b3509d16f8494002a53fa5b417..eb968c30adb952921d4760802f53fdaa6d5aaae2 100644 (file)
@@ -1699,8 +1699,8 @@ struct hl_mmu_funcs {
  * @eq_wq: work queue of event queue for executing work in process context.
  * @kernel_ctx: Kernel driver context structure.
  * @kernel_queues: array of hl_hw_queue.
- * @hw_queues_mirror_list: CS mirror list for TDR.
- * @hw_queues_mirror_lock: protects hw_queues_mirror_list.
+ * @cs_mirror_list: CS mirror list for TDR.
+ * @cs_mirror_lock: protects cs_mirror_list.
  * @kernel_cb_mgr: command buffer manager for creating/destroying/handling CGs.
  * @event_queue: event queue for IRQ from CPU-CP.
  * @dma_pool: DMA pool for small allocations.
@@ -1810,8 +1810,8 @@ struct hl_device {
        struct workqueue_struct         *eq_wq;
        struct hl_ctx                   *kernel_ctx;
        struct hl_hw_queue              *kernel_queues;
-       struct list_head                hw_queues_mirror_list;
-       spinlock_t                      hw_queues_mirror_lock;
+       struct list_head                cs_mirror_list;
+       spinlock_t                      cs_mirror_lock;
        struct hl_cb_mgr                kernel_cb_mgr;
        struct hl_eq                    event_queue;
        struct dma_pool                 *dma_pool;
index f9550fcf550000497588f776afaaa0db7c45d6c9..d1d30fb3641039d3d4c7a2b63e7615e85ebe2ea2 100644 (file)
@@ -578,20 +578,20 @@ int hl_hw_queue_schedule_cs(struct hl_cs *cs)
        else if (cs->type == CS_TYPE_COLLECTIVE_WAIT)
                hdev->asic_funcs->collective_wait_init_cs(cs);
 
-       spin_lock(&hdev->hw_queues_mirror_lock);
-       list_add_tail(&cs->mirror_node, &hdev->hw_queues_mirror_list);
+       spin_lock(&hdev->cs_mirror_lock);
+       list_add_tail(&cs->mirror_node, &hdev->cs_mirror_list);
 
        /* Queue TDR if the CS is the first entry and if timeout is wanted */
        if ((hdev->timeout_jiffies != MAX_SCHEDULE_TIMEOUT) &&
-                       (list_first_entry(&hdev->hw_queues_mirror_list,
+                       (list_first_entry(&hdev->cs_mirror_list,
                                        struct hl_cs, mirror_node) == cs)) {
                cs->tdr_active = true;
                schedule_delayed_work(&cs->work_tdr, hdev->timeout_jiffies);
-               spin_unlock(&hdev->hw_queues_mirror_lock);
-       } else {
-               spin_unlock(&hdev->hw_queues_mirror_lock);
+
        }
 
+       spin_unlock(&hdev->cs_mirror_lock);
+
        if (!hdev->cs_active_cnt++) {
                struct hl_device_idle_busy_ts *ts;