habanalabs: Modify the cs_cnt of a CB to be atomic
authorTomer Tayar <ttayar@habana.ai>
Sun, 2 Aug 2020 19:51:31 +0000 (22:51 +0300)
committerOded Gabbay <ogabbay@kernel.org>
Mon, 30 Nov 2020 08:47:38 +0000 (10:47 +0200)
Modify the CS counter of a CB to be atomic, so no locking is required
when it is being modified or read.

Signed-off-by: Tomer Tayar <ttayar@habana.ai>
Reviewed-by: Oded Gabbay <ogabbay@kernel.org>
Signed-off-by: Oded Gabbay <ogabbay@kernel.org>
drivers/misc/habanalabs/common/command_submission.c
drivers/misc/habanalabs/common/debugfs.c
drivers/misc/habanalabs/common/habanalabs.h
drivers/misc/habanalabs/gaudi/gaudi.c
drivers/misc/habanalabs/goya/goya.c

index dc85ed6ab525f34281fba984d8eaf5f0c9575741..beb482310a58640a255ad9c8a4c815326ad1411e 100644 (file)
@@ -233,10 +233,7 @@ static int cs_parser(struct hl_fpriv *hpriv, struct hl_cs_job *job)
                        job->patched_cb = parser.patched_cb;
                        job->job_cb_size = parser.patched_cb_size;
                        job->contains_dma_pkt = parser.contains_dma_pkt;
-
-                       spin_lock(&job->patched_cb->lock);
-                       job->patched_cb->cs_cnt++;
-                       spin_unlock(&job->patched_cb->lock);
+                       atomic_inc(&job->patched_cb->cs_cnt);
                }
 
                /*
@@ -244,9 +241,7 @@ static int cs_parser(struct hl_fpriv *hpriv, struct hl_cs_job *job)
                 * original CB anymore because it was already parsed and
                 * won't be accessed again for this CS
                 */
-               spin_lock(&job->user_cb->lock);
-               job->user_cb->cs_cnt--;
-               spin_unlock(&job->user_cb->lock);
+               atomic_dec(&job->user_cb->cs_cnt);
                hl_cb_put(job->user_cb);
                job->user_cb = NULL;
        } else if (!rc) {
@@ -268,10 +263,7 @@ static void complete_job(struct hl_device *hdev, struct hl_cs_job *job)
                 * created, so we need to check it's not NULL
                 */
                if (job->patched_cb) {
-                       spin_lock(&job->patched_cb->lock);
-                       job->patched_cb->cs_cnt--;
-                       spin_unlock(&job->patched_cb->lock);
-
+                       atomic_dec(&job->patched_cb->cs_cnt);
                        hl_cb_put(job->patched_cb);
                }
        }
@@ -284,10 +276,7 @@ static void complete_job(struct hl_device *hdev, struct hl_cs_job *job)
        if (job->is_kernel_allocated_cb &&
                ((job->queue_type == QUEUE_TYPE_HW && hdev->mmu_enable) ||
                                job->queue_type == QUEUE_TYPE_INT)) {
-               spin_lock(&job->user_cb->lock);
-               job->user_cb->cs_cnt--;
-               spin_unlock(&job->user_cb->lock);
-
+               atomic_dec(&job->user_cb->cs_cnt);
                hl_cb_put(job->user_cb);
        }
 
@@ -680,9 +669,7 @@ static struct hl_cb *get_cb_from_cs_chunk(struct hl_device *hdev,
                goto release_cb;
        }
 
-       spin_lock(&cb->lock);
-       cb->cs_cnt++;
-       spin_unlock(&cb->lock);
+       atomic_inc(&cb->cs_cnt);
 
        return cb;
 
@@ -936,9 +923,7 @@ static int cs_ioctl_default(struct hl_fpriv *hpriv, void __user *chunks,
        goto put_cs;
 
 release_cb:
-       spin_lock(&cb->lock);
-       cb->cs_cnt--;
-       spin_unlock(&cb->lock);
+       atomic_dec(&cb->cs_cnt);
        hl_cb_put(cb);
 free_cs_object:
        cs_rollback(hdev, cs);
@@ -1133,7 +1118,7 @@ static int cs_ioctl_signal_wait_create_jobs(struct hl_device *hdev,
        job->id = 0;
        job->cs = cs;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = q_idx;
 
index b47a62da0b41e7441d9ab03fb60d8e76fdd3900d..cef716643979d738d1d92f11f1a046a07f9992ed 100644 (file)
@@ -116,7 +116,7 @@ static int command_buffers_show(struct seq_file *s, void *data)
                        "   %03llu        %d    0x%08x      %d          %d          %d\n",
                        cb->id, cb->ctx->asid, cb->size,
                        kref_read(&cb->refcount),
-                       cb->mmap, cb->cs_cnt);
+                       cb->mmap, atomic_read(&cb->cs_cnt));
        }
 
        spin_unlock(&dev_entry->cb_spinlock);
index 8e2d164d97e8bb40f6a0852bbb87400d09f8a9fa..571eda6ef5ab09b77e2ce2a7a43c609efd82397d 100644 (file)
@@ -532,7 +532,7 @@ struct hl_cb_mgr {
  * @refcount: reference counter for usage of the CB.
  * @hdev: pointer to device this CB belongs to.
  * @ctx: pointer to the CB owner's context.
- * @lock: spinlock to protect mmap/cs flows.
+ * @lock: spinlock to protect mmap flows.
  * @debugfs_list: node in debugfs list of command buffers.
  * @pool_list: node in pool list of command buffers.
  * @va_block_list: list of virtual addresses blocks of the CB if it is mapped to
@@ -561,7 +561,7 @@ struct hl_cb {
        dma_addr_t              bus_address;
        u32                     mmap_size;
        u32                     size;
-       u32                     cs_cnt;
+       atomic_t                cs_cnt;
        u8                      mmap;
        u8                      is_pool;
        u8                      is_internal;
index aac3c9c5a2e035ab1f8e1bc9300e8f1e300c32d9..1f1926607c5e7acf9eb2902fa6aecb0e0ebd38c7 100644 (file)
@@ -838,7 +838,7 @@ static int _gaudi_init_tpc_mem(struct hl_device *hdev,
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GAUDI_QUEUE_ID_DMA_0_0;
        job->patched_cb = job->user_cb;
@@ -861,7 +861,7 @@ free_job:
        hl_userptr_delete_list(hdev, &job->userptr_list);
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
 release_cb:
        hl_cb_put(cb);
@@ -1248,7 +1248,7 @@ static int gaudi_collective_wait_create_job(struct hl_device *hdev,
        job->id = 0;
        job->cs = cs;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = queue_id;
 
@@ -5570,7 +5570,7 @@ static int gaudi_memset_device_memory(struct hl_device *hdev, u64 addr,
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GAUDI_QUEUE_ID_DMA_0_0;
        job->patched_cb = job->user_cb;
@@ -5581,7 +5581,7 @@ static int gaudi_memset_device_memory(struct hl_device *hdev, u64 addr,
        rc = gaudi_send_job_on_qman0(hdev, job);
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
        /* Verify DMA is OK */
        err_cause = RREG32(mmDMA0_CORE_ERR_CAUSE);
index 5e151610c1c850e226a917b054d9241e27cf2fe5..3e5eb9e3d7bd84d4434a744903c231eb6ee538ce 100644 (file)
@@ -4811,7 +4811,7 @@ static int goya_memset_device_memory(struct hl_device *hdev, u64 addr, u64 size,
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GOYA_QUEUE_ID_DMA_0;
        job->patched_cb = job->user_cb;
@@ -4823,7 +4823,7 @@ static int goya_memset_device_memory(struct hl_device *hdev, u64 addr, u64 size,
 
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
 release_cb:
        hl_cb_put(cb);