job->patched_cb = parser.patched_cb;
                        job->job_cb_size = parser.patched_cb_size;
                        job->contains_dma_pkt = parser.contains_dma_pkt;
-
-                       spin_lock(&job->patched_cb->lock);
-                       job->patched_cb->cs_cnt++;
-                       spin_unlock(&job->patched_cb->lock);
+                       atomic_inc(&job->patched_cb->cs_cnt);
                }
 
                /*
                 * original CB anymore because it was already parsed and
                 * won't be accessed again for this CS
                 */
-               spin_lock(&job->user_cb->lock);
-               job->user_cb->cs_cnt--;
-               spin_unlock(&job->user_cb->lock);
+               atomic_dec(&job->user_cb->cs_cnt);
                hl_cb_put(job->user_cb);
                job->user_cb = NULL;
        } else if (!rc) {
                 * created, so we need to check it's not NULL
                 */
                if (job->patched_cb) {
-                       spin_lock(&job->patched_cb->lock);
-                       job->patched_cb->cs_cnt--;
-                       spin_unlock(&job->patched_cb->lock);
-
+                       atomic_dec(&job->patched_cb->cs_cnt);
                        hl_cb_put(job->patched_cb);
                }
        }
        if (job->is_kernel_allocated_cb &&
                ((job->queue_type == QUEUE_TYPE_HW && hdev->mmu_enable) ||
                                job->queue_type == QUEUE_TYPE_INT)) {
-               spin_lock(&job->user_cb->lock);
-               job->user_cb->cs_cnt--;
-               spin_unlock(&job->user_cb->lock);
-
+               atomic_dec(&job->user_cb->cs_cnt);
                hl_cb_put(job->user_cb);
        }
 
                goto release_cb;
        }
 
-       spin_lock(&cb->lock);
-       cb->cs_cnt++;
-       spin_unlock(&cb->lock);
+       atomic_inc(&cb->cs_cnt);
 
        return cb;
 
        goto put_cs;
 
 release_cb:
-       spin_lock(&cb->lock);
-       cb->cs_cnt--;
-       spin_unlock(&cb->lock);
+       atomic_dec(&cb->cs_cnt);
        hl_cb_put(cb);
 free_cs_object:
        cs_rollback(hdev, cs);
        job->id = 0;
        job->cs = cs;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = q_idx;
 
 
                        "   %03llu        %d    0x%08x      %d          %d          %d\n",
                        cb->id, cb->ctx->asid, cb->size,
                        kref_read(&cb->refcount),
-                       cb->mmap, cb->cs_cnt);
+                       cb->mmap, atomic_read(&cb->cs_cnt));
        }
 
        spin_unlock(&dev_entry->cb_spinlock);
 
  * @refcount: reference counter for usage of the CB.
  * @hdev: pointer to device this CB belongs to.
  * @ctx: pointer to the CB owner's context.
- * @lock: spinlock to protect mmap/cs flows.
+ * @lock: spinlock to protect mmap flows.
  * @debugfs_list: node in debugfs list of command buffers.
  * @pool_list: node in pool list of command buffers.
  * @va_block_list: list of virtual addresses blocks of the CB if it is mapped to
        dma_addr_t              bus_address;
        u32                     mmap_size;
        u32                     size;
-       u32                     cs_cnt;
+       atomic_t                cs_cnt;
        u8                      mmap;
        u8                      is_pool;
        u8                      is_internal;
 
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GAUDI_QUEUE_ID_DMA_0_0;
        job->patched_cb = job->user_cb;
        hl_userptr_delete_list(hdev, &job->userptr_list);
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
 release_cb:
        hl_cb_put(cb);
        job->id = 0;
        job->cs = cs;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = queue_id;
 
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GAUDI_QUEUE_ID_DMA_0_0;
        job->patched_cb = job->user_cb;
        rc = gaudi_send_job_on_qman0(hdev, job);
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
        /* Verify DMA is OK */
        err_cause = RREG32(mmDMA0_CORE_ERR_CAUSE);
 
 
        job->id = 0;
        job->user_cb = cb;
-       job->user_cb->cs_cnt++;
+       atomic_inc(&job->user_cb->cs_cnt);
        job->user_cb_size = cb_size;
        job->hw_queue_id = GOYA_QUEUE_ID_DMA_0;
        job->patched_cb = job->user_cb;
 
        hl_debugfs_remove_job(hdev, job);
        kfree(job);
-       cb->cs_cnt--;
+       atomic_dec(&cb->cs_cnt);
 
 release_cb:
        hl_cb_put(cb);