{
        struct kpc_dma_device *ldev = (struct kpc_dma_device *)dev_id;
 
-       if ((GetEngineControl(ldev) & ENG_CTL_IRQ_ACTIVE) || (ldev->desc_completed->MyDMAAddr != GetEngineCompletePtr(ldev)))
+       if ((GetEngineControl(ldev) & ENG_CTL_IRQ_ACTIVE) ||
+           (ldev->desc_completed->MyDMAAddr != GetEngineCompletePtr(ldev)))
                schedule_work(&ldev->irq_work);
 
        return IRQ_HANDLED;
        cur = eng->desc_completed;
        do {
                cur = cur->Next;
-               dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n", cur, cur->acd);
+               dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n",
+                       cur, cur->acd);
                BUG_ON(cur == eng->desc_next); // Ordering failure.
 
                if (cur->DescControlFlags & DMA_DESC_CTL_SOP) {
 
                if (cur->DescControlFlags & DMA_DESC_CTL_EOP) {
                        if (cur->acd)
-                               transfer_complete_cb(cur->acd, eng->accumulated_bytes, eng->accumulated_flags | ACD_FLAG_DONE);
+                               transfer_complete_cb(cur->acd, eng->accumulated_bytes,
+                                                    eng->accumulated_flags | ACD_FLAG_DONE);
                }
 
                eng->desc_completed = cur;
                eng->dir = DMA_TO_DEVICE;
 
        eng->desc_pool_cnt = desc_cnt;
-       eng->desc_pool = dma_pool_create("KPC DMA Descriptors", &eng->pldev->dev, sizeof(struct kpc_dma_descriptor), DMA_DESC_ALIGNMENT, 4096);
+       eng->desc_pool = dma_pool_create("KPC DMA Descriptors", &eng->pldev->dev,
+                                        sizeof(struct kpc_dma_descriptor),
+                                        DMA_DESC_ALIGNMENT, 4096);
 
        eng->desc_pool_first = dma_pool_alloc(eng->desc_pool, GFP_KERNEL | GFP_DMA, &head_handle);
        if (!eng->desc_pool_first) {
        INIT_WORK(&eng->irq_work, ndd_irq_worker);
 
        // Grab IRQ line
-       rv = request_irq(eng->irq, ndd_irq_handler, IRQF_SHARED, KP_DRIVER_NAME_DMA_CONTROLLER, eng);
+       rv = request_irq(eng->irq, ndd_irq_handler, IRQF_SHARED,
+                        KP_DRIVER_NAME_DMA_CONTROLLER, eng);
        if (rv) {
                dev_err(&eng->pldev->dev, "%s: failed to request_irq: %d\n", __func__, rv);
                return rv;
        }
 
        // Clear any persistent bits just to make sure there is no residue from the reset
-       SetClearEngineControl(eng, (ENG_CTL_IRQ_ACTIVE | ENG_CTL_DESC_COMPLETE | ENG_CTL_DESC_ALIGN_ERR | ENG_CTL_DESC_FETCH_ERR | ENG_CTL_SW_ABORT_ERR | ENG_CTL_DESC_CHAIN_END | ENG_CTL_DMA_WAITING_PERSIST), 0);
+       SetClearEngineControl(eng, (ENG_CTL_IRQ_ACTIVE | ENG_CTL_DESC_COMPLETE |
+                                   ENG_CTL_DESC_ALIGN_ERR | ENG_CTL_DESC_FETCH_ERR |
+                                   ENG_CTL_SW_ABORT_ERR | ENG_CTL_DESC_CHAIN_END |
+                                   ENG_CTL_DMA_WAITING_PERSIST), 0);
 
        // Reset performance counters
 
 
 
        // Lock the user buffer pages in memory, and hold on to the page pointers (for the sglist)
        mmap_read_lock(current->mm);      /*  get memory map semaphore */
-       rv = pin_user_pages(iov_base, acd->page_count, FOLL_TOUCH | FOLL_WRITE, acd->user_pages, NULL);
+       rv = pin_user_pages(iov_base, acd->page_count, FOLL_TOUCH | FOLL_WRITE,
+                           acd->user_pages, NULL);
        mmap_read_unlock(current->mm);        /*  release the semaphore */
        if (rv != acd->page_count) {
                nr_pages = rv;
        nr_pages = acd->page_count;
 
        // Allocate and setup the sg_table (scatterlist entries)
-       rv = sg_alloc_table_from_pages(&acd->sgt, acd->user_pages, acd->page_count, iov_base & (PAGE_SIZE - 1), iov_len, GFP_KERNEL);
+       rv = sg_alloc_table_from_pages(&acd->sgt, acd->user_pages, acd->page_count,
+                                      iov_base & (PAGE_SIZE - 1), iov_len, GFP_KERNEL);
        if (rv) {
                dev_err(&priv->ldev->pldev->dev, "Couldn't alloc sg_table (%d)\n", rv);
                goto unpin_pages;
        }
 
        // Setup the DMA mapping for all the sg entries
-       acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents, ldev->dir);
+       acd->mapped_entry_count = dma_map_sg(&ldev->pldev->dev, acd->sgt.sgl, acd->sgt.nents,
+                                            ldev->dir);
        if (acd->mapped_entry_count <= 0) {
-               dev_err(&priv->ldev->pldev->dev, "Couldn't dma_map_sg (%d)\n", acd->mapped_entry_count);
+               dev_err(&priv->ldev->pldev->dev, "Couldn't dma_map_sg (%d)\n",
+                       acd->mapped_entry_count);
                goto free_table;
        }
 
 
        // Figoure out how many descriptors are available and return an error if there aren't enough
        num_descrs_avail = count_descriptors_available(ldev);
-       dev_dbg(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
+       dev_dbg(&priv->ldev->pldev->dev,
+               "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d\n",
+               acd->mapped_entry_count, desc_needed, num_descrs_avail);
+
        if (desc_needed >= ldev->desc_pool_cnt) {
-               dev_warn(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    TOO MANY to ever complete!\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
+               dev_warn(&priv->ldev->pldev->dev,
+                        "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    TOO MANY to ever complete!\n",
+                        acd->mapped_entry_count, desc_needed, num_descrs_avail);
                rv = -EAGAIN;
                goto err_descr_too_many;
        }
        if (desc_needed > num_descrs_avail) {
-               dev_warn(&priv->ldev->pldev->dev, "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    Too many to complete right now.\n", acd->mapped_entry_count, desc_needed, num_descrs_avail);
+               dev_warn(&priv->ldev->pldev->dev,
+                        "    mapped_entry_count = %d    num_descrs_needed = %d    num_descrs_avail = %d    Too many to complete right now.\n",
+                        acd->mapped_entry_count, desc_needed, num_descrs_avail);
                rv = -EMSGSIZE;
                goto err_descr_too_many;
        }
                        if (i == acd->mapped_entry_count - 1 && p == pcnt - 1)
                                desc->acd = acd;
 
-                       dev_dbg(&priv->ldev->pldev->dev, "  Filled descriptor %p (acd = %p)\n", desc, desc->acd);
+                       dev_dbg(&priv->ldev->pldev->dev, "  Filled descriptor %p (acd = %p)\n",
+                               desc, desc->acd);
 
                        ldev->desc_next = desc->Next;
                        desc = desc->Next;
 
 
        // Setup miscdev struct
        dev = MKDEV(assigned_major_num, pldev->id);
-       ldev->kpc_dma_dev = device_create(kpc_dma_class, &pldev->dev, dev, ldev, "kpc_dma%d", pldev->id);
+       ldev->kpc_dma_dev = device_create(kpc_dma_class, &pldev->dev, dev, ldev,
+                                         "kpc_dma%d", pldev->id);
        if (IS_ERR(ldev->kpc_dma_dev)) {
                rv = PTR_ERR(ldev->kpc_dma_dev);
                dev_err(&ldev->pldev->dev, "%s: device_create failed: %d\n", __func__, rv);
 {
        int err;
 
-       err = __register_chrdev(KPC_DMA_CHAR_MAJOR, 0, KPC_DMA_NUM_MINORS, "kpc_dma", &kpc_dma_fops);
+       err = __register_chrdev(KPC_DMA_CHAR_MAJOR, 0, KPC_DMA_NUM_MINORS,
+                               "kpc_dma", &kpc_dma_fops);
        if (err < 0) {
-               pr_err("Can't allocate a major number (%d) for kpc_dma (err = %d)\n", KPC_DMA_CHAR_MAJOR, err);
+               pr_err("Can't allocate a major number (%d) for kpc_dma (err = %d)\n",
+                      KPC_DMA_CHAR_MAJOR, err);
                goto fail_chrdev_register;
        }
        assigned_major_num = err;