scsi: scsi_debug: Use blk_mq_tagset_busy_iter() in sdebug_blk_mq_poll()
authorJohn Garry <john.g.garry@oracle.com>
Mon, 27 Mar 2023 07:43:06 +0000 (07:43 +0000)
committerMartin K. Petersen <martin.petersen@oracle.com>
Mon, 3 Apr 2023 02:09:22 +0000 (22:09 -0400)
Instead of iterating all deferred commands in the submission queue
structures, use blk_mq_tagset_busy_iter(), which is a standard API for
this.

Signed-off-by: John Garry <john.g.garry@oracle.com>
Link: https://lore.kernel.org/r/20230327074310.1862889-8-john.g.garry@oracle.com
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
drivers/scsi/scsi_debug.c

index 1c85cbd92178565d76fc51c4e3a83bd6d3fe2641..4ec94ba593e84106f9892d741b195850b4016d16 100644 (file)
@@ -7511,123 +7511,124 @@ static void sdebug_map_queues(struct Scsi_Host *shost)
        }
 }
 
-static int sdebug_blk_mq_poll(struct Scsi_Host *shost, unsigned int queue_num)
+struct sdebug_blk_mq_poll_data {
+       unsigned int queue_num;
+       int *num_entries;
+};
+
+/*
+ * We don't handle aborted commands here, but it does not seem possible to have
+ * aborted polled commands from schedule_resp()
+ */
+static bool sdebug_blk_mq_poll_iter(struct request *rq, void *opaque)
 {
-       bool first;
-       bool retiring = false;
-       int num_entries = 0;
-       unsigned int qc_idx = 0;
-       unsigned long iflags;
-       ktime_t kt_from_boot = ktime_get_boottime();
-       struct sdebug_queue *sqp;
-       struct sdebug_queued_cmd *sqcp;
-       struct scsi_cmnd *scp;
+       struct sdebug_blk_mq_poll_data *data = opaque;
+       struct scsi_cmnd *cmd = blk_mq_rq_to_pdu(rq);
+       struct sdebug_scsi_cmd *sdsc = scsi_cmd_priv(cmd);
        struct sdebug_defer *sd_dp;
+       u32 unique_tag = blk_mq_unique_tag(rq);
+       u16 hwq = blk_mq_unique_tag_to_hwq(unique_tag);
+       struct sdebug_queued_cmd *sqcp;
+       struct sdebug_queue *sqp;
+       unsigned long flags;
+       int queue_num = data->queue_num;
+       bool retiring = false;
+       int qc_idx;
+       ktime_t time;
 
-       sqp = sdebug_q_arr + queue_num;
+       /* We're only interested in one queue for this iteration */
+       if (hwq != queue_num)
+               return true;
 
-       spin_lock_irqsave(&sqp->qc_lock, iflags);
+       /* Subsequent checks would fail if this failed, but check anyway */
+       if (!test_bit(SCMD_STATE_INFLIGHT, &cmd->state))
+               return true;
 
-       qc_idx = find_first_bit(sqp->in_use_bm, sdebug_max_queue);
-       if (qc_idx >= sdebug_max_queue)
-               goto unlock;
+       time = ktime_get_boottime();
 
-       for (first = true; first || qc_idx + 1 < sdebug_max_queue; )   {
-               unsigned long flags;
-               struct sdebug_scsi_cmd *sdsc;
-               if (first) {
-                       first = false;
-                       if (!test_bit(qc_idx, sqp->in_use_bm))
-                               continue;
-               } else {
-                       qc_idx = find_next_bit(sqp->in_use_bm, sdebug_max_queue, qc_idx + 1);
-               }
-               if (qc_idx >= sdebug_max_queue)
-                       break;
+       spin_lock_irqsave(&sdsc->lock, flags);
+       sqcp = TO_QUEUED_CMD(cmd);
+       if (!sqcp) {
+               spin_unlock_irqrestore(&sdsc->lock, flags);
+               return true;
+       }
 
-               sqcp = sqp->qc_arr[qc_idx];
-               if (!sqcp) {
-                       pr_err("sqcp is NULL, queue_num=%d, qc_idx=%u from %s\n",
-                              queue_num, qc_idx, __func__);
-                       break;
-               }
-               sd_dp = &sqcp->sd_dp;
+       sqp = sdebug_q_arr + queue_num;
+       sd_dp = &sqcp->sd_dp;
 
-               scp = sqcp->scmd;
-               if (unlikely(scp == NULL)) {
-                       pr_err("scp is NULL, queue_num=%d, qc_idx=%u from %s\n",
-                              queue_num, qc_idx, __func__);
-                       break;
-               }
-               sdsc = scsi_cmd_priv(scp);
-               spin_lock_irqsave(&sdsc->lock, flags);
-               if (READ_ONCE(sd_dp->defer_t) == SDEB_DEFER_POLL) {
-                       struct sdebug_queued_cmd *_sqcp = TO_QUEUED_CMD(scp);
-
-                       if (_sqcp != sqcp) {
-                               pr_err("inconsistent queued cmd tag=%#x\n",
-                                      blk_mq_unique_tag(scsi_cmd_to_rq(scp)));
-                               spin_unlock_irqrestore(&sdsc->lock, flags);
-                               continue;
-                       }
+       if (READ_ONCE(sd_dp->defer_t) != SDEB_DEFER_POLL) {
+               spin_unlock_irqrestore(&sdsc->lock, flags);
+               return true;
+       }
 
-                       if (kt_from_boot < sd_dp->cmpl_ts) {
-                               spin_unlock_irqrestore(&sdsc->lock, flags);
-                               continue;
-                       }
+       if (time < sd_dp->cmpl_ts) {
+               spin_unlock_irqrestore(&sdsc->lock, flags);
+               return true;
+       }
 
-               } else          /* ignoring non REQ_POLLED requests */ {
-                       spin_unlock_irqrestore(&sdsc->lock, flags);
-                       continue;
-               }
-               if (unlikely(atomic_read(&retired_max_queue) > 0))
-                       retiring = true;
+       if (unlikely(atomic_read(&retired_max_queue) > 0))
+               retiring = true;
 
-               if (unlikely(!test_and_clear_bit(qc_idx, sqp->in_use_bm))) {
+       qc_idx = sd_dp->sqa_idx;
+       sqp->qc_arr[qc_idx] = NULL;
+       if (unlikely(!test_and_clear_bit(qc_idx, sqp->in_use_bm))) {
+               spin_unlock_irqrestore(&sdsc->lock, flags);
+               pr_err("Unexpected completion sqp %p queue_num=%d qc_idx=%u\n",
+                       sqp, queue_num, qc_idx);
+               sdebug_free_queued_cmd(sqcp);
+               return true;
+       }
+
+       if (unlikely(retiring)) {       /* user has reduced max_queue */
+               int k, retval = atomic_read(&retired_max_queue);
+
+               if (qc_idx >= retval) {
+                       pr_err("index %d too large\n", retval);
                        spin_unlock_irqrestore(&sdsc->lock, flags);
-                       pr_err("Unexpected completion sqp %p queue_num=%d qc_idx=%u from %s\n",
-                               sqp, queue_num, qc_idx, __func__);
                        sdebug_free_queued_cmd(sqcp);
-                       break;
-               }
-               sqp->qc_arr[qc_idx] = NULL;
-               if (unlikely(retiring)) {       /* user has reduced max_queue */
-                       int k, retval;
-
-                       retval = atomic_read(&retired_max_queue);
-                       if (qc_idx >= retval) {
-                               pr_err("index %d too large\n", retval);
-                               spin_unlock_irqrestore(&sdsc->lock, flags);
-                               sdebug_free_queued_cmd(sqcp);
-                               break;
-                       }
-                       k = find_last_bit(sqp->in_use_bm, retval);
-                       if ((k < sdebug_max_queue) || (k == retval))
-                               atomic_set(&retired_max_queue, 0);
-                       else
-                               atomic_set(&retired_max_queue, k + 1);
+                       return true;
                }
-               spin_unlock_irqrestore(&sdsc->lock, flags);
-               spin_unlock_irqrestore(&sqp->qc_lock, iflags);
 
-               if (sdebug_statistics) {
-                       atomic_inc(&sdebug_completions);
-                       if (raw_smp_processor_id() != sd_dp->issuing_cpu)
-                               atomic_inc(&sdebug_miss_cpus);
-               }
+               k = find_last_bit(sqp->in_use_bm, retval);
+               if ((k < sdebug_max_queue) || (k == retval))
+                       atomic_set(&retired_max_queue, 0);
+               else
+                       atomic_set(&retired_max_queue, k + 1);
+       }
 
-               sdebug_free_queued_cmd(sqcp);
+       ASSIGN_QUEUED_CMD(cmd, NULL);
+       spin_unlock_irqrestore(&sdsc->lock, flags);
 
-               scsi_done(scp); /* callback to mid level */
-               num_entries++;
-               spin_lock_irqsave(&sqp->qc_lock, iflags);
-               if (find_first_bit(sqp->in_use_bm, sdebug_max_queue) >= sdebug_max_queue)
-                       break;
+       if (sdebug_statistics) {
+               atomic_inc(&sdebug_completions);
+               if (raw_smp_processor_id() != sd_dp->issuing_cpu)
+                       atomic_inc(&sdebug_miss_cpus);
        }
 
-unlock:
-       spin_unlock_irqrestore(&sqp->qc_lock, iflags);
+       sdebug_free_queued_cmd(sqcp);
 
+       scsi_done(cmd); /* callback to mid level */
+       (*data->num_entries)++;
+       return true;
+}
+
+static int sdebug_blk_mq_poll(struct Scsi_Host *shost, unsigned int queue_num)
+{
+       int num_entries = 0;
+       unsigned long iflags;
+       struct sdebug_queue *sqp;
+       struct sdebug_blk_mq_poll_data data = {
+               .queue_num = queue_num,
+               .num_entries = &num_entries,
+       };
+       sqp = sdebug_q_arr + queue_num;
+
+       spin_lock_irqsave(&sqp->qc_lock, iflags);
+
+       blk_mq_tagset_busy_iter(&shost->tag_set, sdebug_blk_mq_poll_iter,
+                               &data);
+
+       spin_unlock_irqrestore(&sqp->qc_lock, iflags);
        if (num_entries > 0)
                atomic_add(num_entries, &sdeb_mq_poll_count);
        return num_entries;