clear_bit(WQ_FLAG_DEDICATED, &wq->flags);
        clear_bit(WQ_FLAG_BLOCK_ON_FAULT, &wq->flags);
        memset(wq->name, 0, WQ_NAME_SIZE);
+       wq->max_xfer_bytes = WQ_DEFAULT_MAX_XFER;
+       wq->max_batch_size = WQ_DEFAULT_MAX_BATCH;
 }
 
 static void idxd_wq_ref_release(struct percpu_ref *ref)
                wq->wqcfg->bits[i] = ioread32(idxd->reg_base + wq_offset);
        }
 
+       if (wq->size == 0 && wq->type != IDXD_WQT_NONE)
+               wq->size = WQ_DEFAULT_QUEUE_DEPTH;
+
        /* byte 0-3 */
        wq->wqcfg->wq_size = wq->size;
 
-       if (wq->size == 0) {
-               idxd->cmd_status = IDXD_SCMD_WQ_NO_SIZE;
-               dev_warn(dev, "Incorrect work queue size: 0\n");
-               return -EINVAL;
-       }
-
        /* bytes 4-7 */
        wq->wqcfg->wq_thresh = wq->threshold;
 
 
                if (!wq->group)
                        continue;
-               if (!wq->size)
-                       continue;
 
                if (wq_shared(wq) && !device_swq_supported(idxd)) {
                        idxd->cmd_status = IDXD_SCMD_WQ_NO_SWQ_SUPPORT;
 
 #define WQ_NAME_SIZE   1024
 #define WQ_TYPE_SIZE   10
 
+#define WQ_DEFAULT_QUEUE_DEPTH         16
+#define WQ_DEFAULT_MAX_XFER            SZ_2M
+#define WQ_DEFAULT_MAX_BATCH           32
+
 enum idxd_op_type {
        IDXD_OP_BLOCK = 0,
        IDXD_OP_NONBLOCK = 1,
 
                init_waitqueue_head(&wq->err_queue);
                init_completion(&wq->wq_dead);
                init_completion(&wq->wq_resurrect);
-               wq->max_xfer_bytes = idxd->max_xfer_bytes;
-               wq->max_batch_size = idxd->max_batch_size;
+               wq->max_xfer_bytes = WQ_DEFAULT_MAX_XFER;
+               wq->max_batch_size = WQ_DEFAULT_MAX_BATCH;
                wq->wqcfg = kzalloc_node(idxd->wqcfg_size, GFP_KERNEL, dev_to_node(dev));
                if (!wq->wqcfg) {
                        put_device(conf_dev);