LCOV - code coverage report
Current view: top level - lib/nvmf - ctrlr.c (source / functions) Hit Total Coverage
Test: ut_cov_unit.info Lines: 1320 2477 53.3 %
Date: 2024-07-11 21:30:05 Functions: 85 154 55.2 %

          Line data    Source code
       1             : /*   SPDX-License-Identifier: BSD-3-Clause
       2             :  *   Copyright (C) 2017 Intel Corporation. All rights reserved.
       3             :  *   Copyright (c) 2019, 2020 Mellanox Technologies LTD. All rights reserved.
       4             :  *   Copyright (c) 2021, 2023 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
       5             :  */
       6             : 
       7             : #include "spdk/stdinc.h"
       8             : 
       9             : #include "nvmf_internal.h"
      10             : #include "transport.h"
      11             : 
      12             : #include "spdk/bdev.h"
      13             : #include "spdk/bdev_zone.h"
      14             : #include "spdk/bit_array.h"
      15             : #include "spdk/endian.h"
      16             : #include "spdk/thread.h"
      17             : #include "spdk/nvme_spec.h"
      18             : #include "spdk/nvmf_cmd.h"
      19             : #include "spdk/string.h"
      20             : #include "spdk/util.h"
      21             : #include "spdk/version.h"
      22             : #include "spdk/log.h"
      23             : #include "spdk_internal/usdt.h"
      24             : 
      25             : #define MIN_KEEP_ALIVE_TIMEOUT_IN_MS 10000
      26             : #define NVMF_DISC_KATO_IN_MS 120000
      27             : #define KAS_TIME_UNIT_IN_MS 100
      28             : #define KAS_DEFAULT_VALUE (MIN_KEEP_ALIVE_TIMEOUT_IN_MS / KAS_TIME_UNIT_IN_MS)
      29             : 
      30             : #define NVMF_CC_RESET_SHN_TIMEOUT_IN_MS 10000
      31             : 
      32             : #define NVMF_CTRLR_RESET_SHN_TIMEOUT_IN_MS      (NVMF_CC_RESET_SHN_TIMEOUT_IN_MS + 5000)
      33             : 
      34             : #define DUPLICATE_QID_RETRY_US 1000
      35             : 
      36             : /*
      37             :  * Report the SPDK version as the firmware revision.
      38             :  * SPDK_VERSION_STRING won't fit into FR (only 8 bytes), so try to fit the most important parts.
      39             :  */
      40             : #define FW_VERSION SPDK_VERSION_MAJOR_STRING SPDK_VERSION_MINOR_STRING SPDK_VERSION_PATCH_STRING
      41             : 
      42             : #define ANA_TRANSITION_TIME_IN_SEC 10
      43             : 
      44             : #define NVMF_ABORT_COMMAND_LIMIT 3
      45             : 
      46             : /*
      47             :  * Support for custom admin command handlers
      48             :  */
      49             : struct spdk_nvmf_custom_admin_cmd {
      50             :         spdk_nvmf_custom_cmd_hdlr hdlr;
      51             :         uint32_t nsid; /* nsid to forward */
      52             : };
      53             : 
      54             : static struct spdk_nvmf_custom_admin_cmd g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_MAX_OPC + 1];
      55             : 
      56             : static void _nvmf_request_complete(void *ctx);
      57             : int nvmf_passthru_admin_cmd_for_ctrlr(struct spdk_nvmf_request *req, struct spdk_nvmf_ctrlr *ctrlr);
      58             : 
      59             : static inline void
      60          12 : nvmf_invalid_connect_response(struct spdk_nvmf_fabric_connect_rsp *rsp,
      61             :                               uint8_t iattr, uint16_t ipo)
      62             : {
      63          12 :         rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
      64          12 :         rsp->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
      65          12 :         rsp->status_code_specific.invalid.iattr = iattr;
      66          12 :         rsp->status_code_specific.invalid.ipo = ipo;
      67          12 : }
      68             : 
      69             : #define SPDK_NVMF_INVALID_CONNECT_CMD(rsp, field)       \
      70             :         nvmf_invalid_connect_response(rsp, 0, offsetof(struct spdk_nvmf_fabric_connect_cmd, field))
      71             : #define SPDK_NVMF_INVALID_CONNECT_DATA(rsp, field)      \
      72             :         nvmf_invalid_connect_response(rsp, 1, offsetof(struct spdk_nvmf_fabric_connect_data, field))
      73             : 
      74             : 
      75             : static void
      76           4 : nvmf_ctrlr_stop_keep_alive_timer(struct spdk_nvmf_ctrlr *ctrlr)
      77             : {
      78           4 :         if (!ctrlr) {
      79           0 :                 SPDK_ERRLOG("Controller is NULL\n");
      80           0 :                 return;
      81             :         }
      82             : 
      83           4 :         if (ctrlr->keep_alive_poller == NULL) {
      84           0 :                 return;
      85             :         }
      86             : 
      87           4 :         SPDK_DEBUGLOG(nvmf, "Stop keep alive poller\n");
      88           4 :         spdk_poller_unregister(&ctrlr->keep_alive_poller);
      89             : }
      90             : 
      91             : static void
      92           1 : nvmf_ctrlr_stop_association_timer(struct spdk_nvmf_ctrlr *ctrlr)
      93             : {
      94           1 :         if (!ctrlr) {
      95           0 :                 SPDK_ERRLOG("Controller is NULL\n");
      96           0 :                 assert(false);
      97             :                 return;
      98             :         }
      99             : 
     100           1 :         if (ctrlr->association_timer == NULL) {
     101           1 :                 return;
     102             :         }
     103             : 
     104           0 :         SPDK_DEBUGLOG(nvmf, "Stop association timer\n");
     105           0 :         spdk_poller_unregister(&ctrlr->association_timer);
     106             : }
     107             : 
     108             : static void
     109           0 : nvmf_ctrlr_disconnect_qpairs_done(struct spdk_io_channel_iter *i, int status)
     110             : {
     111           0 :         if (status == 0) {
     112           0 :                 SPDK_DEBUGLOG(nvmf, "ctrlr disconnect qpairs complete successfully\n");
     113             :         } else {
     114           0 :                 SPDK_ERRLOG("Fail to disconnect ctrlr qpairs\n");
     115             :         }
     116           0 : }
     117             : 
     118             : static int
     119           0 : _nvmf_ctrlr_disconnect_qpairs_on_pg(struct spdk_io_channel_iter *i, bool include_admin)
     120             : {
     121           0 :         int rc = 0;
     122             :         struct spdk_nvmf_ctrlr *ctrlr;
     123             :         struct spdk_nvmf_qpair *qpair, *temp_qpair;
     124             :         struct spdk_io_channel *ch;
     125             :         struct spdk_nvmf_poll_group *group;
     126             : 
     127           0 :         ctrlr = spdk_io_channel_iter_get_ctx(i);
     128           0 :         ch = spdk_io_channel_iter_get_channel(i);
     129           0 :         group = spdk_io_channel_get_ctx(ch);
     130             : 
     131           0 :         TAILQ_FOREACH_SAFE(qpair, &group->qpairs, link, temp_qpair) {
     132           0 :                 if (qpair->ctrlr == ctrlr && (include_admin || !nvmf_qpair_is_admin_queue(qpair))) {
     133           0 :                         rc = spdk_nvmf_qpair_disconnect(qpair, NULL, NULL);
     134           0 :                         if (rc) {
     135           0 :                                 if (rc == -EINPROGRESS) {
     136           0 :                                         rc = 0;
     137             :                                 } else {
     138           0 :                                         SPDK_ERRLOG("Qpair disconnect failed\n");
     139           0 :                                         return rc;
     140             :                                 }
     141             :                         }
     142             :                 }
     143             :         }
     144             : 
     145           0 :         return rc;
     146             : }
     147             : 
     148             : static void
     149           0 : nvmf_ctrlr_disconnect_qpairs_on_pg(struct spdk_io_channel_iter *i)
     150             : {
     151           0 :         spdk_for_each_channel_continue(i, _nvmf_ctrlr_disconnect_qpairs_on_pg(i, true));
     152           0 : }
     153             : 
     154             : static void
     155           0 : nvmf_ctrlr_disconnect_io_qpairs_on_pg(struct spdk_io_channel_iter *i)
     156             : {
     157           0 :         spdk_for_each_channel_continue(i, _nvmf_ctrlr_disconnect_qpairs_on_pg(i, false));
     158           0 : }
     159             : 
     160             : static int
     161           0 : nvmf_ctrlr_keep_alive_poll(void *ctx)
     162             : {
     163             :         uint64_t keep_alive_timeout_tick;
     164           0 :         uint64_t now = spdk_get_ticks();
     165           0 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
     166             : 
     167           0 :         if (ctrlr->in_destruct) {
     168           0 :                 nvmf_ctrlr_stop_keep_alive_timer(ctrlr);
     169           0 :                 return SPDK_POLLER_IDLE;
     170             :         }
     171             : 
     172           0 :         SPDK_DEBUGLOG(nvmf, "Polling ctrlr keep alive timeout\n");
     173             : 
     174             :         /* If the Keep alive feature is in use and the timer expires */
     175           0 :         keep_alive_timeout_tick = ctrlr->last_keep_alive_tick +
     176           0 :                                   ctrlr->feat.keep_alive_timer.bits.kato * spdk_get_ticks_hz() / UINT64_C(1000);
     177           0 :         if (now > keep_alive_timeout_tick) {
     178           0 :                 SPDK_NOTICELOG("Disconnecting host %s from subsystem %s due to keep alive timeout.\n",
     179             :                                ctrlr->hostnqn, ctrlr->subsys->subnqn);
     180             :                 /* set the Controller Fatal Status bit to '1' */
     181           0 :                 if (ctrlr->vcprop.csts.bits.cfs == 0) {
     182           0 :                         nvmf_ctrlr_set_fatal_status(ctrlr);
     183             : 
     184             :                         /*
     185             :                          * disconnect qpairs, terminate Transport connection
     186             :                          * destroy ctrlr, break the host to controller association
     187             :                          * disconnect qpairs with qpair->ctrlr == ctrlr
     188             :                          */
     189           0 :                         spdk_for_each_channel(ctrlr->subsys->tgt,
     190             :                                               nvmf_ctrlr_disconnect_qpairs_on_pg,
     191             :                                               ctrlr,
     192             :                                               nvmf_ctrlr_disconnect_qpairs_done);
     193           0 :                         return SPDK_POLLER_BUSY;
     194             :                 }
     195             :         }
     196             : 
     197           0 :         return SPDK_POLLER_IDLE;
     198             : }
     199             : 
     200             : static void
     201           5 : nvmf_ctrlr_start_keep_alive_timer(struct spdk_nvmf_ctrlr *ctrlr)
     202             : {
     203           5 :         if (!ctrlr) {
     204           0 :                 SPDK_ERRLOG("Controller is NULL\n");
     205           0 :                 return;
     206             :         }
     207             : 
     208             :         /* if cleared to 0 then the Keep Alive Timer is disabled */
     209           5 :         if (ctrlr->feat.keep_alive_timer.bits.kato != 0) {
     210             : 
     211           4 :                 ctrlr->last_keep_alive_tick = spdk_get_ticks();
     212             : 
     213           4 :                 SPDK_DEBUGLOG(nvmf, "Ctrlr add keep alive poller\n");
     214           4 :                 ctrlr->keep_alive_poller = SPDK_POLLER_REGISTER(nvmf_ctrlr_keep_alive_poll, ctrlr,
     215             :                                            ctrlr->feat.keep_alive_timer.bits.kato * 1000);
     216             :         }
     217             : }
     218             : 
     219             : static int _retry_qid_check(void *ctx);
     220             : 
     221             : static void
     222          10 : ctrlr_add_qpair_and_send_rsp(struct spdk_nvmf_qpair *qpair,
     223             :                              struct spdk_nvmf_ctrlr *ctrlr,
     224             :                              struct spdk_nvmf_request *req)
     225             : {
     226          10 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     227             : 
     228          10 :         assert(ctrlr->admin_qpair->group->thread == spdk_get_thread());
     229             : 
     230          10 :         if (spdk_bit_array_get(ctrlr->qpair_mask, qpair->qid)) {
     231           3 :                 if (qpair->connect_req != NULL) {
     232           1 :                         SPDK_ERRLOG("Got I/O connect with duplicate QID %u\n", qpair->qid);
     233           1 :                         rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
     234           1 :                         rsp->status.sc = SPDK_NVME_SC_INVALID_QUEUE_IDENTIFIER;
     235           1 :                         qpair->connect_req = NULL;
     236           1 :                         qpair->ctrlr = NULL;
     237           1 :                         spdk_nvmf_request_complete(req);
     238             :                 } else {
     239           2 :                         SPDK_WARNLOG("Duplicate QID detected, re-check in %dus\n",
     240             :                                      DUPLICATE_QID_RETRY_US);
     241           2 :                         qpair->connect_req = req;
     242             :                         /* Set qpair->ctrlr here so that we'll have it when the poller expires. */
     243           2 :                         qpair->ctrlr = ctrlr;
     244           2 :                         req->poller = SPDK_POLLER_REGISTER(_retry_qid_check, qpair,
     245             :                                                            DUPLICATE_QID_RETRY_US);
     246             :                 }
     247           3 :                 return;
     248             :         }
     249             : 
     250           7 :         qpair->ctrlr = ctrlr;
     251           7 :         spdk_bit_array_set(ctrlr->qpair_mask, qpair->qid);
     252             : 
     253           7 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
     254           7 :         rsp->status_code_specific.success.cntlid = ctrlr->cntlid;
     255           7 :         SPDK_DEBUGLOG(nvmf, "connect capsule response: cntlid = 0x%04x\n",
     256             :                       rsp->status_code_specific.success.cntlid);
     257           7 :         spdk_nvmf_request_complete(req);
     258             : 
     259             :         SPDK_DTRACE_PROBE4_TICKS(nvmf_ctrlr_add_qpair, qpair, qpair->qid, ctrlr->subsys->subnqn,
     260             :                                  ctrlr->hostnqn);
     261             : }
     262             : 
     263             : static int
     264           2 : _retry_qid_check(void *ctx)
     265             : {
     266           2 :         struct spdk_nvmf_qpair *qpair = ctx;
     267           2 :         struct spdk_nvmf_request *req = qpair->connect_req;
     268           2 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
     269             : 
     270           2 :         spdk_poller_unregister(&req->poller);
     271           2 :         ctrlr_add_qpair_and_send_rsp(qpair, ctrlr, req);
     272           2 :         return SPDK_POLLER_BUSY;
     273             : }
     274             : 
     275             : static void
     276           5 : _nvmf_ctrlr_add_admin_qpair(void *ctx)
     277             : {
     278           5 :         struct spdk_nvmf_request *req = ctx;
     279           5 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     280           5 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
     281             : 
     282           5 :         ctrlr->admin_qpair = qpair;
     283           5 :         ctrlr->association_timeout = qpair->transport->opts.association_timeout;
     284           5 :         nvmf_ctrlr_start_keep_alive_timer(ctrlr);
     285           5 :         ctrlr_add_qpair_and_send_rsp(qpair, ctrlr, req);
     286           5 : }
     287             : 
     288             : static void
     289           5 : _nvmf_subsystem_add_ctrlr(void *ctx)
     290             : {
     291           5 :         struct spdk_nvmf_request *req = ctx;
     292           5 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     293           5 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     294           5 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
     295             : 
     296           5 :         if (nvmf_subsystem_add_ctrlr(ctrlr->subsys, ctrlr)) {
     297           0 :                 SPDK_ERRLOG("Unable to add controller to subsystem\n");
     298           0 :                 spdk_bit_array_free(&ctrlr->qpair_mask);
     299           0 :                 free(ctrlr);
     300           0 :                 qpair->ctrlr = NULL;
     301           0 :                 rsp->status.sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
     302           0 :                 spdk_nvmf_request_complete(req);
     303           0 :                 return;
     304             :         }
     305             : 
     306           5 :         spdk_thread_send_msg(ctrlr->thread, _nvmf_ctrlr_add_admin_qpair, req);
     307             : }
     308             : 
     309             : static void
     310           6 : nvmf_ctrlr_cdata_init(struct spdk_nvmf_transport *transport, struct spdk_nvmf_subsystem *subsystem,
     311             :                       struct spdk_nvmf_ctrlr_data *cdata)
     312             : {
     313           6 :         cdata->aerl = SPDK_NVMF_MAX_ASYNC_EVENTS - 1;
     314           6 :         cdata->kas = KAS_DEFAULT_VALUE;
     315           6 :         cdata->vid = SPDK_PCI_VID_INTEL;
     316           6 :         cdata->ssvid = SPDK_PCI_VID_INTEL;
     317             :         /* INTEL OUI */
     318           6 :         cdata->ieee[0] = 0xe4;
     319           6 :         cdata->ieee[1] = 0xd2;
     320           6 :         cdata->ieee[2] = 0x5c;
     321           6 :         cdata->oncs.compare = 1;
     322           6 :         cdata->oncs.reservations = 1;
     323           6 :         cdata->fuses.compare_and_write = 1;
     324           6 :         cdata->oncs.copy = 1;
     325           6 :         cdata->sgls.supported = 1;
     326           6 :         cdata->sgls.keyed_sgl = 1;
     327           6 :         cdata->sgls.sgl_offset = 1;
     328           6 :         cdata->nvmf_specific.ioccsz = sizeof(struct spdk_nvme_cmd) / 16;
     329           6 :         cdata->nvmf_specific.ioccsz += transport->opts.in_capsule_data_size / 16;
     330           6 :         cdata->nvmf_specific.iorcsz = sizeof(struct spdk_nvme_cpl) / 16;
     331           6 :         cdata->nvmf_specific.icdoff = 0; /* offset starts directly after SQE */
     332           6 :         cdata->nvmf_specific.ctrattr.ctrlr_model = SPDK_NVMF_CTRLR_MODEL_DYNAMIC;
     333           6 :         cdata->nvmf_specific.msdbd = 1;
     334             : 
     335           6 :         if (transport->ops->cdata_init) {
     336           0 :                 transport->ops->cdata_init(transport, subsystem, cdata);
     337             :         }
     338           6 : }
     339             : 
     340             : static bool
     341           5 : nvmf_subsys_has_multi_iocs(struct spdk_nvmf_subsystem *subsystem)
     342             : {
     343             :         struct spdk_nvmf_ns *ns;
     344             :         uint32_t i;
     345             : 
     346           5 :         for (i = 0; i < subsystem->max_nsid; i++) {
     347           0 :                 ns = subsystem->ns[i];
     348           0 :                 if (ns && ns->bdev && spdk_bdev_is_zoned(ns->bdev)) {
     349           0 :                         return true;
     350             :                 }
     351             :         }
     352           5 :         return false;
     353             : }
     354             : 
     355             : static struct spdk_nvmf_ctrlr *
     356           5 : nvmf_ctrlr_create(struct spdk_nvmf_subsystem *subsystem,
     357             :                   struct spdk_nvmf_request *req,
     358             :                   struct spdk_nvmf_fabric_connect_cmd *connect_cmd,
     359             :                   struct spdk_nvmf_fabric_connect_data *connect_data)
     360             : {
     361             :         struct spdk_nvmf_ctrlr *ctrlr;
     362           5 :         struct spdk_nvmf_transport *transport = req->qpair->transport;
     363           5 :         struct spdk_nvme_transport_id listen_trid = {};
     364           5 :         bool subsys_has_multi_iocs = false;
     365             : 
     366           5 :         ctrlr = calloc(1, sizeof(*ctrlr));
     367           5 :         if (ctrlr == NULL) {
     368           0 :                 SPDK_ERRLOG("Memory allocation failed\n");
     369           0 :                 return NULL;
     370             :         }
     371             : 
     372           5 :         if (spdk_nvme_trtype_is_fabrics(transport->ops->type)) {
     373           5 :                 ctrlr->dynamic_ctrlr = true;
     374             :         } else {
     375           0 :                 ctrlr->cntlid = connect_data->cntlid;
     376             :         }
     377             : 
     378             :         SPDK_DTRACE_PROBE3_TICKS(nvmf_ctrlr_create, ctrlr, subsystem->subnqn,
     379             :                                  spdk_thread_get_id(req->qpair->group->thread));
     380             : 
     381           5 :         STAILQ_INIT(&ctrlr->async_events);
     382           5 :         TAILQ_INIT(&ctrlr->log_head);
     383           5 :         ctrlr->subsys = subsystem;
     384           5 :         ctrlr->thread = req->qpair->group->thread;
     385           5 :         ctrlr->disconnect_in_progress = false;
     386             : 
     387           5 :         ctrlr->qpair_mask = spdk_bit_array_create(transport->opts.max_qpairs_per_ctrlr);
     388           5 :         if (!ctrlr->qpair_mask) {
     389           0 :                 SPDK_ERRLOG("Failed to allocate controller qpair mask\n");
     390           0 :                 goto err_qpair_mask;
     391             :         }
     392             : 
     393           5 :         nvmf_ctrlr_cdata_init(transport, subsystem, &ctrlr->cdata);
     394             : 
     395             :         /*
     396             :          * KAS: This field indicates the granularity of the Keep Alive Timer in 100ms units.
     397             :          * If this field is cleared to 0h, then Keep Alive is not supported.
     398             :          */
     399           5 :         if (ctrlr->cdata.kas) {
     400           5 :                 ctrlr->feat.keep_alive_timer.bits.kato = spdk_divide_round_up(connect_cmd->kato,
     401             :                                 KAS_DEFAULT_VALUE * KAS_TIME_UNIT_IN_MS) *
     402             :                                 KAS_DEFAULT_VALUE * KAS_TIME_UNIT_IN_MS;
     403             :         }
     404             : 
     405           5 :         ctrlr->feat.async_event_configuration.bits.ns_attr_notice = 1;
     406           5 :         if (ctrlr->subsys->flags.ana_reporting) {
     407           0 :                 ctrlr->feat.async_event_configuration.bits.ana_change_notice = 1;
     408             :         }
     409           5 :         ctrlr->feat.volatile_write_cache.bits.wce = 1;
     410             :         /* Coalescing Disable */
     411           5 :         ctrlr->feat.interrupt_vector_configuration.bits.cd = 1;
     412             : 
     413           5 :         if (spdk_nvmf_subsystem_is_discovery(ctrlr->subsys)) {
     414             :                 /*
     415             :                  * If keep-alive timeout is not set, discovery controllers use some
     416             :                  * arbitrary high value in order to cleanup stale discovery sessions
     417             :                  *
     418             :                  * From the 1.0a nvme-of spec:
     419             :                  * "The Keep Alive command is reserved for
     420             :                  * Discovery controllers. A transport may specify a
     421             :                  * fixed Discovery controller activity timeout value
     422             :                  * (e.g., 2 minutes). If no commands are received
     423             :                  * by a Discovery controller within that time
     424             :                  * period, the controller may perform the
     425             :                  * actions for Keep Alive Timer expiration".
     426             :                  *
     427             :                  * From the 1.1 nvme-of spec:
     428             :                  * "A host requests an explicit persistent connection
     429             :                  * to a Discovery controller and Asynchronous Event Notifications from
     430             :                  * the Discovery controller on that persistent connection by specifying
     431             :                  * a non-zero Keep Alive Timer value in the Connect command."
     432             :                  *
     433             :                  * In case non-zero KATO is used, we enable discovery_log_change_notice
     434             :                  * otherwise we disable it and use default discovery controller KATO.
     435             :                  * KATO is in millisecond.
     436             :                  */
     437           2 :                 if (ctrlr->feat.keep_alive_timer.bits.kato == 0) {
     438           1 :                         ctrlr->feat.keep_alive_timer.bits.kato = NVMF_DISC_KATO_IN_MS;
     439           1 :                         ctrlr->feat.async_event_configuration.bits.discovery_log_change_notice = 0;
     440             :                 } else {
     441           1 :                         ctrlr->feat.async_event_configuration.bits.discovery_log_change_notice = 1;
     442             :                 }
     443             :         }
     444             : 
     445             :         /* Subtract 1 for admin queue, 1 for 0's based */
     446           5 :         ctrlr->feat.number_of_queues.bits.ncqr = transport->opts.max_qpairs_per_ctrlr - 1 -
     447             :                         1;
     448           5 :         ctrlr->feat.number_of_queues.bits.nsqr = transport->opts.max_qpairs_per_ctrlr - 1 -
     449             :                         1;
     450             : 
     451           5 :         spdk_uuid_copy(&ctrlr->hostid, (struct spdk_uuid *)connect_data->hostid);
     452           5 :         memcpy(ctrlr->hostnqn, connect_data->hostnqn, sizeof(ctrlr->hostnqn));
     453             : 
     454           5 :         ctrlr->vcprop.cap.raw = 0;
     455           5 :         ctrlr->vcprop.cap.bits.cqr = 1; /* NVMe-oF specification required */
     456           5 :         ctrlr->vcprop.cap.bits.mqes = transport->opts.max_queue_depth -
     457             :                                       1; /* max queue depth */
     458           5 :         ctrlr->vcprop.cap.bits.ams = 0; /* optional arb mechanisms */
     459             :         /* ready timeout - 500 msec units */
     460           5 :         ctrlr->vcprop.cap.bits.to = NVMF_CTRLR_RESET_SHN_TIMEOUT_IN_MS / 500;
     461           5 :         ctrlr->vcprop.cap.bits.dstrd = 0; /* fixed to 0 for NVMe-oF */
     462           5 :         subsys_has_multi_iocs = nvmf_subsys_has_multi_iocs(subsystem);
     463           5 :         if (subsys_has_multi_iocs) {
     464           0 :                 ctrlr->vcprop.cap.bits.css =
     465             :                         SPDK_NVME_CAP_CSS_IOCS; /* One or more I/O command sets supported */
     466             :         } else {
     467           5 :                 ctrlr->vcprop.cap.bits.css = SPDK_NVME_CAP_CSS_NVM; /* NVM command set */
     468             :         }
     469             : 
     470           5 :         ctrlr->vcprop.cap.bits.mpsmin = 0; /* 2 ^ (12 + mpsmin) == 4k */
     471           5 :         ctrlr->vcprop.cap.bits.mpsmax = 0; /* 2 ^ (12 + mpsmax) == 4k */
     472             : 
     473             :         /* Version Supported: 1.3 */
     474           5 :         ctrlr->vcprop.vs.bits.mjr = 1;
     475           5 :         ctrlr->vcprop.vs.bits.mnr = 3;
     476           5 :         ctrlr->vcprop.vs.bits.ter = 0;
     477             : 
     478           5 :         ctrlr->vcprop.cc.raw = 0;
     479           5 :         ctrlr->vcprop.cc.bits.en = 0; /* Init controller disabled */
     480           5 :         if (subsys_has_multi_iocs) {
     481           0 :                 ctrlr->vcprop.cc.bits.css =
     482             :                         SPDK_NVME_CC_CSS_IOCS; /* All supported I/O Command Sets */
     483             :         }
     484             : 
     485           5 :         ctrlr->vcprop.csts.raw = 0;
     486           5 :         ctrlr->vcprop.csts.bits.rdy = 0; /* Init controller as not ready */
     487             : 
     488           5 :         SPDK_DEBUGLOG(nvmf, "cap 0x%" PRIx64 "\n", ctrlr->vcprop.cap.raw);
     489           5 :         SPDK_DEBUGLOG(nvmf, "vs 0x%x\n", ctrlr->vcprop.vs.raw);
     490           5 :         SPDK_DEBUGLOG(nvmf, "cc 0x%x\n", ctrlr->vcprop.cc.raw);
     491           5 :         SPDK_DEBUGLOG(nvmf, "csts 0x%x\n", ctrlr->vcprop.csts.raw);
     492             : 
     493           5 :         ctrlr->dif_insert_or_strip = transport->opts.dif_insert_or_strip;
     494             : 
     495           5 :         if (ctrlr->subsys->subtype == SPDK_NVMF_SUBTYPE_NVME) {
     496           3 :                 if (spdk_nvmf_qpair_get_listen_trid(req->qpair, &listen_trid) != 0) {
     497           0 :                         SPDK_ERRLOG("Could not get listener transport ID\n");
     498           0 :                         goto err_listener;
     499             :                 }
     500             : 
     501           3 :                 ctrlr->listener = nvmf_subsystem_find_listener(ctrlr->subsys, &listen_trid);
     502           3 :                 if (!ctrlr->listener) {
     503           0 :                         SPDK_ERRLOG("Listener was not found\n");
     504           0 :                         goto err_listener;
     505             :                 }
     506             :         }
     507             : 
     508           5 :         req->qpair->ctrlr = ctrlr;
     509           5 :         spdk_thread_send_msg(subsystem->thread, _nvmf_subsystem_add_ctrlr, req);
     510             : 
     511           5 :         return ctrlr;
     512           0 : err_listener:
     513           0 :         spdk_bit_array_free(&ctrlr->qpair_mask);
     514           0 : err_qpair_mask:
     515           0 :         free(ctrlr);
     516           0 :         return NULL;
     517             : }
     518             : 
     519             : static void
     520           1 : _nvmf_ctrlr_destruct(void *ctx)
     521             : {
     522           1 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
     523             :         struct spdk_nvmf_reservation_log *log, *log_tmp;
     524             :         struct spdk_nvmf_async_event_completion *event, *event_tmp;
     525             : 
     526             :         SPDK_DTRACE_PROBE3_TICKS(nvmf_ctrlr_destruct, ctrlr, ctrlr->subsys->subnqn,
     527             :                                  spdk_thread_get_id(ctrlr->thread));
     528             : 
     529           1 :         assert(spdk_get_thread() == ctrlr->thread);
     530           1 :         assert(ctrlr->in_destruct);
     531             : 
     532           1 :         SPDK_DEBUGLOG(nvmf, "Destroy ctrlr 0x%hx\n", ctrlr->cntlid);
     533           1 :         if (ctrlr->disconnect_in_progress) {
     534           0 :                 SPDK_ERRLOG("freeing ctrlr with disconnect in progress\n");
     535           0 :                 spdk_thread_send_msg(ctrlr->thread, _nvmf_ctrlr_destruct, ctrlr);
     536           0 :                 return;
     537             :         }
     538             : 
     539           1 :         nvmf_ctrlr_stop_keep_alive_timer(ctrlr);
     540           1 :         nvmf_ctrlr_stop_association_timer(ctrlr);
     541           1 :         spdk_bit_array_free(&ctrlr->qpair_mask);
     542             : 
     543           1 :         TAILQ_FOREACH_SAFE(log, &ctrlr->log_head, link, log_tmp) {
     544           0 :                 TAILQ_REMOVE(&ctrlr->log_head, log, link);
     545           0 :                 free(log);
     546             :         }
     547           1 :         STAILQ_FOREACH_SAFE(event, &ctrlr->async_events, link, event_tmp) {
     548           0 :                 STAILQ_REMOVE(&ctrlr->async_events, event, spdk_nvmf_async_event_completion, link);
     549           0 :                 free(event);
     550             :         }
     551           1 :         free(ctrlr);
     552             : }
     553             : 
     554             : void
     555           1 : nvmf_ctrlr_destruct(struct spdk_nvmf_ctrlr *ctrlr)
     556             : {
     557           1 :         nvmf_subsystem_remove_ctrlr(ctrlr->subsys, ctrlr);
     558             : 
     559           1 :         spdk_thread_send_msg(ctrlr->thread, _nvmf_ctrlr_destruct, ctrlr);
     560           1 : }
     561             : 
     562             : static void
     563           8 : nvmf_ctrlr_add_io_qpair(void *ctx)
     564             : {
     565           8 :         struct spdk_nvmf_request *req = ctx;
     566           8 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     567           8 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     568           8 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
     569           8 :         struct spdk_nvmf_qpair *admin_qpair = ctrlr->admin_qpair;
     570             : 
     571             :         SPDK_DTRACE_PROBE4_TICKS(nvmf_ctrlr_add_io_qpair, ctrlr, req->qpair, req->qpair->qid,
     572             :                                  spdk_thread_get_id(ctrlr->thread));
     573             : 
     574             :         /* Unit test will check qpair->ctrlr after calling spdk_nvmf_ctrlr_connect.
     575             :           * For error case, the value should be NULL. So set it to NULL at first.
     576             :           */
     577           8 :         qpair->ctrlr = NULL;
     578             : 
     579             :         /* Make sure the controller is not being destroyed. */
     580           8 :         if (ctrlr->in_destruct) {
     581           0 :                 SPDK_ERRLOG("Got I/O connect while ctrlr was being destroyed.\n");
     582           0 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     583           0 :                 goto end;
     584             :         }
     585             : 
     586           8 :         if (spdk_nvmf_subsystem_is_discovery(ctrlr->subsys)) {
     587           1 :                 SPDK_ERRLOG("I/O connect not allowed on discovery controller\n");
     588           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     589           1 :                 goto end;
     590             :         }
     591             : 
     592           7 :         if (!ctrlr->vcprop.cc.bits.en) {
     593           1 :                 SPDK_ERRLOG("Got I/O connect before ctrlr was enabled\n");
     594           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     595           1 :                 goto end;
     596             :         }
     597             : 
     598           6 :         if (1u << ctrlr->vcprop.cc.bits.iosqes != sizeof(struct spdk_nvme_cmd)) {
     599           1 :                 SPDK_ERRLOG("Got I/O connect with invalid IOSQES %u\n",
     600             :                             ctrlr->vcprop.cc.bits.iosqes);
     601           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     602           1 :                 goto end;
     603             :         }
     604             : 
     605           5 :         if (1u << ctrlr->vcprop.cc.bits.iocqes != sizeof(struct spdk_nvme_cpl)) {
     606           1 :                 SPDK_ERRLOG("Got I/O connect with invalid IOCQES %u\n",
     607             :                             ctrlr->vcprop.cc.bits.iocqes);
     608           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     609           1 :                 goto end;
     610             :         }
     611             : 
     612           4 :         if (admin_qpair->state != SPDK_NVMF_QPAIR_ACTIVE || admin_qpair->group == NULL) {
     613             :                 /* There is a chance that admin qpair is being destroyed at this moment due to e.g.
     614             :                  * expired keep alive timer. Part of the qpair destruction process is change of qpair's
     615             :                  * state to DEACTIVATING and removing it from poll group */
     616           0 :                 SPDK_ERRLOG("Inactive admin qpair (state %d, group %p)\n", admin_qpair->state, admin_qpair->group);
     617           0 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     618           0 :                 goto end;
     619             :         }
     620             : 
     621             :         /* check if we would exceed ctrlr connection limit */
     622           4 :         if (qpair->qid >= spdk_bit_array_capacity(ctrlr->qpair_mask)) {
     623           1 :                 SPDK_ERRLOG("Requested QID %u but Max QID is %u\n",
     624             :                             qpair->qid, spdk_bit_array_capacity(ctrlr->qpair_mask) - 1);
     625           1 :                 rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
     626           1 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_QUEUE_IDENTIFIER;
     627           1 :                 goto end;
     628             :         }
     629             : 
     630           3 :         ctrlr_add_qpair_and_send_rsp(qpair, ctrlr, req);
     631           3 :         return;
     632           5 : end:
     633           5 :         spdk_nvmf_request_complete(req);
     634             : }
     635             : 
     636             : static void
     637          10 : _nvmf_ctrlr_add_io_qpair(void *ctx)
     638             : {
     639          10 :         struct spdk_nvmf_request *req = ctx;
     640          10 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     641             :         struct spdk_nvmf_fabric_connect_data *data;
     642             :         struct spdk_nvmf_ctrlr *ctrlr;
     643          10 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     644             :         struct spdk_nvmf_qpair *admin_qpair;
     645          10 :         struct spdk_nvmf_tgt *tgt = qpair->transport->tgt;
     646             :         struct spdk_nvmf_subsystem *subsystem;
     647          10 :         struct spdk_nvme_transport_id listen_trid = {};
     648             :         const struct spdk_nvmf_subsystem_listener *listener;
     649             : 
     650          10 :         assert(req->iovcnt == 1);
     651             : 
     652          10 :         data = req->iov[0].iov_base;
     653             : 
     654          10 :         SPDK_DEBUGLOG(nvmf, "Connect I/O Queue for controller id 0x%x\n", data->cntlid);
     655             : 
     656          10 :         subsystem = spdk_nvmf_tgt_find_subsystem(tgt, data->subnqn);
     657             :         /* We already checked this in spdk_nvmf_ctrlr_connect */
     658          10 :         assert(subsystem != NULL);
     659             : 
     660          10 :         ctrlr = nvmf_subsystem_get_ctrlr(subsystem, data->cntlid);
     661          10 :         if (ctrlr == NULL) {
     662           1 :                 SPDK_ERRLOG("Unknown controller ID 0x%x\n", data->cntlid);
     663           1 :                 SPDK_NVMF_INVALID_CONNECT_DATA(rsp, cntlid);
     664           1 :                 spdk_nvmf_request_complete(req);
     665           1 :                 return;
     666             :         }
     667             : 
     668             :         /* fail before passing a message to the controller thread. */
     669           9 :         if (ctrlr->in_destruct) {
     670           0 :                 SPDK_ERRLOG("Got I/O connect while ctrlr was being destroyed.\n");
     671           0 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     672           0 :                 spdk_nvmf_request_complete(req);
     673           0 :                 return;
     674             :         }
     675             : 
     676             :         /* If ANA reporting is enabled, check if I/O connect is on the same listener. */
     677           9 :         if (subsystem->flags.ana_reporting) {
     678           0 :                 if (spdk_nvmf_qpair_get_listen_trid(req->qpair, &listen_trid) != 0) {
     679           0 :                         SPDK_ERRLOG("Could not get listener transport ID\n");
     680           0 :                         SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     681           0 :                         spdk_nvmf_request_complete(req);
     682           0 :                         return;
     683             :                 }
     684             : 
     685           0 :                 listener = nvmf_subsystem_find_listener(subsystem, &listen_trid);
     686           0 :                 if (listener != ctrlr->listener) {
     687           0 :                         SPDK_ERRLOG("I/O connect is on a listener different from admin connect\n");
     688           0 :                         SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     689           0 :                         spdk_nvmf_request_complete(req);
     690           0 :                         return;
     691             :                 }
     692             :         }
     693             : 
     694           9 :         admin_qpair = ctrlr->admin_qpair;
     695           9 :         if (admin_qpair->state != SPDK_NVMF_QPAIR_ACTIVE || admin_qpair->group == NULL) {
     696             :                 /* There is a chance that admin qpair is being destroyed at this moment due to e.g.
     697             :                  * expired keep alive timer. Part of the qpair destruction process is change of qpair's
     698             :                  * state to DEACTIVATING and removing it from poll group */
     699           1 :                 SPDK_ERRLOG("Inactive admin qpair (state %d, group %p)\n", admin_qpair->state, admin_qpair->group);
     700           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, qid);
     701           1 :                 spdk_nvmf_request_complete(req);
     702           1 :                 return;
     703             :         }
     704           8 :         qpair->ctrlr = ctrlr;
     705           8 :         spdk_thread_send_msg(admin_qpair->group->thread, nvmf_ctrlr_add_io_qpair, req);
     706             : }
     707             : 
     708             : static bool
     709          20 : nvmf_qpair_access_allowed(struct spdk_nvmf_qpair *qpair, struct spdk_nvmf_subsystem *subsystem,
     710             :                           const char *hostnqn)
     711             : {
     712          20 :         struct spdk_nvme_transport_id listen_trid = {};
     713             : 
     714          20 :         if (!spdk_nvmf_subsystem_host_allowed(subsystem, hostnqn)) {
     715           1 :                 SPDK_ERRLOG("Subsystem '%s' does not allow host '%s'\n", subsystem->subnqn, hostnqn);
     716           1 :                 return false;
     717             :         }
     718             : 
     719          19 :         if (spdk_nvmf_qpair_get_listen_trid(qpair, &listen_trid)) {
     720           0 :                 SPDK_ERRLOG("Subsystem '%s' is unable to enforce access control due to an internal error.\n",
     721             :                             subsystem->subnqn);
     722           0 :                 return false;
     723             :         }
     724             : 
     725          19 :         if (!spdk_nvmf_subsystem_listener_allowed(subsystem, &listen_trid)) {
     726           0 :                 SPDK_ERRLOG("Subsystem '%s' does not allow host '%s' to connect at this address.\n",
     727             :                             subsystem->subnqn, hostnqn);
     728           0 :                 return false;
     729             :         }
     730             : 
     731          19 :         return true;
     732             : }
     733             : 
     734             : static int
     735          19 : _nvmf_ctrlr_connect(struct spdk_nvmf_request *req)
     736             : {
     737          19 :         struct spdk_nvmf_fabric_connect_data *data = req->iov[0].iov_base;
     738          19 :         struct spdk_nvmf_fabric_connect_cmd *cmd = &req->cmd->connect_cmd;
     739          19 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     740          19 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     741          19 :         struct spdk_nvmf_transport *transport = qpair->transport;
     742             :         struct spdk_nvmf_ctrlr *ctrlr;
     743             :         struct spdk_nvmf_subsystem *subsystem;
     744             : 
     745          19 :         SPDK_DEBUGLOG(nvmf, "recfmt 0x%x qid %u sqsize %u\n",
     746             :                       cmd->recfmt, cmd->qid, cmd->sqsize);
     747             : 
     748          19 :         SPDK_DEBUGLOG(nvmf, "Connect data:\n");
     749          19 :         SPDK_DEBUGLOG(nvmf, "  cntlid:  0x%04x\n", data->cntlid);
     750          19 :         SPDK_DEBUGLOG(nvmf, "  hostid: %08x-%04x-%04x-%02x%02x-%04x%08x ***\n",
     751             :                       ntohl(*(uint32_t *)&data->hostid[0]),
     752             :                       ntohs(*(uint16_t *)&data->hostid[4]),
     753             :                       ntohs(*(uint16_t *)&data->hostid[6]),
     754             :                       data->hostid[8],
     755             :                       data->hostid[9],
     756             :                       ntohs(*(uint16_t *)&data->hostid[10]),
     757             :                       ntohl(*(uint32_t *)&data->hostid[12]));
     758          19 :         SPDK_DEBUGLOG(nvmf, "  subnqn: \"%s\"\n", data->subnqn);
     759          19 :         SPDK_DEBUGLOG(nvmf, "  hostnqn: \"%s\"\n", data->hostnqn);
     760             : 
     761          19 :         subsystem = spdk_nvmf_tgt_find_subsystem(transport->tgt, data->subnqn);
     762          19 :         if (!subsystem) {
     763           0 :                 SPDK_NVMF_INVALID_CONNECT_DATA(rsp, subnqn);
     764           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     765             :         }
     766             : 
     767          19 :         if (cmd->recfmt != 0) {
     768           1 :                 SPDK_ERRLOG("Connect command unsupported RECFMT %u\n", cmd->recfmt);
     769           1 :                 rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
     770           1 :                 rsp->status.sc = SPDK_NVMF_FABRIC_SC_INCOMPATIBLE_FORMAT;
     771           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     772             :         }
     773             : 
     774             :         /*
     775             :          * SQSIZE is a 0-based value, so it must be at least 1 (minimum queue depth is 2) and
     776             :          * strictly less than max_aq_depth (admin queues) or max_queue_depth (io queues).
     777             :          */
     778          18 :         if (cmd->sqsize == 0) {
     779           1 :                 SPDK_ERRLOG("Invalid SQSIZE = 0\n");
     780           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, sqsize);
     781           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     782             :         }
     783             : 
     784          17 :         if (cmd->qid == 0) {
     785           6 :                 if (cmd->sqsize >= transport->opts.max_aq_depth) {
     786           1 :                         SPDK_ERRLOG("Invalid SQSIZE for admin queue %u (min 1, max %u)\n",
     787             :                                     cmd->sqsize, transport->opts.max_aq_depth - 1);
     788           1 :                         SPDK_NVMF_INVALID_CONNECT_CMD(rsp, sqsize);
     789           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     790             :                 }
     791          11 :         } else if (cmd->sqsize >= transport->opts.max_queue_depth) {
     792           1 :                 SPDK_ERRLOG("Invalid SQSIZE %u (min 1, max %u)\n",
     793             :                             cmd->sqsize, transport->opts.max_queue_depth - 1);
     794           1 :                 SPDK_NVMF_INVALID_CONNECT_CMD(rsp, sqsize);
     795           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     796             :         }
     797             : 
     798          15 :         qpair->sq_head_max = cmd->sqsize;
     799          15 :         qpair->qid = cmd->qid;
     800          15 :         qpair->connect_received = true;
     801             : 
     802          15 :         pthread_mutex_lock(&qpair->group->mutex);
     803          15 :         qpair->group->current_unassociated_qpairs--;
     804          15 :         pthread_mutex_unlock(&qpair->group->mutex);
     805             : 
     806          15 :         if (0 == qpair->qid) {
     807           5 :                 qpair->group->stat.admin_qpairs++;
     808           5 :                 qpair->group->stat.current_admin_qpairs++;
     809             :         } else {
     810          10 :                 qpair->group->stat.io_qpairs++;
     811          10 :                 qpair->group->stat.current_io_qpairs++;
     812             :         }
     813             : 
     814          15 :         if (cmd->qid == 0) {
     815           5 :                 SPDK_DEBUGLOG(nvmf, "Connect Admin Queue for controller ID 0x%x\n", data->cntlid);
     816             : 
     817           5 :                 if (spdk_nvme_trtype_is_fabrics(transport->ops->type) && data->cntlid != 0xFFFF) {
     818             :                         /* This NVMf target only supports dynamic mode. */
     819           1 :                         SPDK_ERRLOG("The NVMf target only supports dynamic mode (CNTLID = 0x%x).\n", data->cntlid);
     820           1 :                         SPDK_NVMF_INVALID_CONNECT_DATA(rsp, cntlid);
     821           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     822             :                 }
     823             : 
     824             :                 /* Establish a new ctrlr */
     825           4 :                 ctrlr = nvmf_ctrlr_create(subsystem, req, cmd, data);
     826           4 :                 if (!ctrlr) {
     827           0 :                         SPDK_ERRLOG("nvmf_ctrlr_create() failed\n");
     828           0 :                         rsp->status.sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
     829           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     830             :                 } else {
     831           4 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
     832             :                 }
     833             :         } else {
     834          10 :                 spdk_thread_send_msg(subsystem->thread, _nvmf_ctrlr_add_io_qpair, req);
     835          10 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
     836             :         }
     837             : }
     838             : 
     839             : static struct spdk_nvmf_subsystem_poll_group *
     840           8 : nvmf_subsystem_pg_from_connect_cmd(struct spdk_nvmf_request *req)
     841             : {
     842             :         struct spdk_nvmf_fabric_connect_data *data;
     843             :         struct spdk_nvmf_subsystem *subsystem;
     844             :         struct spdk_nvmf_tgt *tgt;
     845             : 
     846           8 :         assert(nvmf_request_is_fabric_connect(req));
     847           8 :         assert(req->qpair->ctrlr == NULL);
     848           8 :         assert(req->iovcnt == 1);
     849             : 
     850           8 :         data = req->iov[0].iov_base;
     851           8 :         tgt = req->qpair->transport->tgt;
     852             : 
     853           8 :         subsystem = spdk_nvmf_tgt_find_subsystem(tgt, data->subnqn);
     854           8 :         if (subsystem == NULL) {
     855           0 :                 return NULL;
     856             :         }
     857             : 
     858           8 :         return &req->qpair->group->sgroups[subsystem->id];
     859             : }
     860             : 
     861             : int
     862           0 : spdk_nvmf_ctrlr_connect(struct spdk_nvmf_request *req)
     863             : {
     864           0 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     865             :         struct spdk_nvmf_subsystem_poll_group *sgroup;
     866           0 :         struct spdk_nvmf_qpair *qpair = req->qpair;
     867             :         enum spdk_nvmf_request_exec_status status;
     868             : 
     869           0 :         if (req->iovcnt > 1) {
     870           0 :                 SPDK_ERRLOG("Connect command invalid iovcnt: %d\n", req->iovcnt);
     871           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
     872           0 :                 status = SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     873           0 :                 goto out;
     874             :         }
     875             : 
     876           0 :         sgroup = nvmf_subsystem_pg_from_connect_cmd(req);
     877           0 :         if (!sgroup) {
     878           0 :                 SPDK_NVMF_INVALID_CONNECT_DATA(rsp, subnqn);
     879           0 :                 status = SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     880           0 :                 goto out;
     881             :         }
     882             : 
     883           0 :         sgroup->mgmt_io_outstanding++;
     884           0 :         TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
     885             : 
     886           0 :         status = _nvmf_ctrlr_connect(req);
     887             : 
     888           0 : out:
     889           0 :         if (status == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
     890           0 :                 _nvmf_request_complete(req);
     891             :         }
     892             : 
     893           0 :         return status;
     894             : }
     895             : 
     896             : static int
     897          23 : nvmf_ctrlr_cmd_connect(struct spdk_nvmf_request *req)
     898             : {
     899          23 :         struct spdk_nvmf_fabric_connect_data *data = req->iov[0].iov_base;
     900          23 :         struct spdk_nvmf_fabric_connect_rsp *rsp = &req->rsp->connect_rsp;
     901          23 :         struct spdk_nvmf_transport *transport = req->qpair->transport;
     902             :         struct spdk_nvmf_subsystem *subsystem;
     903             : 
     904          23 :         if (req->length < sizeof(struct spdk_nvmf_fabric_connect_data)) {
     905           1 :                 SPDK_ERRLOG("Connect command data length 0x%x too small\n", req->length);
     906           1 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
     907           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     908             :         }
     909             : 
     910          22 :         if (req->iovcnt > 1) {
     911           0 :                 SPDK_ERRLOG("Connect command invalid iovcnt: %d\n", req->iovcnt);
     912           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
     913           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     914             :         }
     915             : 
     916          22 :         subsystem = spdk_nvmf_tgt_find_subsystem(transport->tgt, data->subnqn);
     917          22 :         if (!subsystem) {
     918           1 :                 SPDK_NVMF_INVALID_CONNECT_DATA(rsp, subnqn);
     919           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     920             :         }
     921             : 
     922          21 :         if ((subsystem->state == SPDK_NVMF_SUBSYSTEM_INACTIVE) ||
     923          21 :             (subsystem->state == SPDK_NVMF_SUBSYSTEM_PAUSING) ||
     924          21 :             (subsystem->state == SPDK_NVMF_SUBSYSTEM_PAUSED) ||
     925          21 :             (subsystem->state == SPDK_NVMF_SUBSYSTEM_DEACTIVATING)) {
     926             :                 struct spdk_nvmf_subsystem_poll_group *sgroup;
     927             : 
     928             :                 /* Subsystem is not ready to handle a connect. Decrement
     929             :                  * the mgmt_io_outstanding to avoid the subsystem waiting
     930             :                  * for this command to complete before unpausing. Queued
     931             :                  * requests get retried when subsystem resumes.
     932             :                  */
     933           0 :                 sgroup = nvmf_subsystem_pg_from_connect_cmd(req);
     934           0 :                 assert(sgroup != NULL);
     935           0 :                 sgroup->mgmt_io_outstanding--;
     936           0 :                 TAILQ_REMOVE(&req->qpair->outstanding, req, link);
     937           0 :                 TAILQ_INSERT_TAIL(&sgroup->queued, req, link);
     938           0 :                 SPDK_DEBUGLOG(nvmf, "Subsystem '%s' is not ready for connect, retrying...\n", subsystem->subnqn);
     939           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
     940             :         }
     941             : 
     942             :         /* Ensure that hostnqn is null terminated */
     943          21 :         if (!memchr(data->hostnqn, '\0', SPDK_NVMF_NQN_MAX_LEN + 1)) {
     944           1 :                 SPDK_ERRLOG("Connect HOSTNQN is not null terminated\n");
     945           1 :                 SPDK_NVMF_INVALID_CONNECT_DATA(rsp, hostnqn);
     946           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     947             :         }
     948             : 
     949          20 :         if (!nvmf_qpair_access_allowed(req->qpair, subsystem, data->hostnqn)) {
     950           1 :                 rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
     951           1 :                 rsp->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_HOST;
     952           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
     953             :         }
     954             : 
     955          19 :         return _nvmf_ctrlr_connect(req);
     956             : }
     957             : 
     958             : static int
     959           0 : nvmf_ctrlr_association_remove(void *ctx)
     960             : {
     961           0 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
     962             :         int rc;
     963             : 
     964           0 :         nvmf_ctrlr_stop_association_timer(ctrlr);
     965             : 
     966           0 :         if (ctrlr->in_destruct) {
     967           0 :                 return SPDK_POLLER_IDLE;
     968             :         }
     969           0 :         SPDK_DEBUGLOG(nvmf, "Disconnecting host from subsystem %s due to association timeout.\n",
     970             :                       ctrlr->subsys->subnqn);
     971             : 
     972           0 :         if (ctrlr->admin_qpair) {
     973           0 :                 rc = spdk_nvmf_qpair_disconnect(ctrlr->admin_qpair, NULL, NULL);
     974           0 :                 if (rc < 0 && rc != -EINPROGRESS) {
     975           0 :                         SPDK_ERRLOG("Fail to disconnect admin ctrlr qpair\n");
     976           0 :                         assert(false);
     977             :                 }
     978             :         }
     979             : 
     980           0 :         return SPDK_POLLER_BUSY;
     981             : }
     982             : 
     983             : static int
     984           0 : _nvmf_ctrlr_cc_reset_shn_done(void *ctx)
     985             : {
     986           0 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
     987           0 :         uint64_t now = spdk_get_ticks();
     988             :         uint32_t count;
     989             : 
     990           0 :         if (ctrlr->cc_timer) {
     991           0 :                 spdk_poller_unregister(&ctrlr->cc_timer);
     992             :         }
     993             : 
     994           0 :         count = spdk_bit_array_count_set(ctrlr->qpair_mask);
     995           0 :         SPDK_DEBUGLOG(nvmf, "ctrlr %p active queue count %u\n", ctrlr, count);
     996             : 
     997           0 :         if (count > 1) {
     998           0 :                 if (now < ctrlr->cc_timeout_tsc) {
     999             :                         /* restart cc timer */
    1000           0 :                         ctrlr->cc_timer = SPDK_POLLER_REGISTER(_nvmf_ctrlr_cc_reset_shn_done, ctrlr, 100 * 1000);
    1001           0 :                         return SPDK_POLLER_IDLE;
    1002             :                 } else {
    1003             :                         /* controller fatal status */
    1004           0 :                         SPDK_WARNLOG("IO timeout, ctrlr %p is in fatal status\n", ctrlr);
    1005           0 :                         nvmf_ctrlr_set_fatal_status(ctrlr);
    1006             :                 }
    1007             :         }
    1008             : 
    1009           0 :         spdk_poller_unregister(&ctrlr->cc_timeout_timer);
    1010             : 
    1011           0 :         if (ctrlr->disconnect_is_shn) {
    1012           0 :                 ctrlr->vcprop.csts.bits.shst = SPDK_NVME_SHST_COMPLETE;
    1013           0 :                 ctrlr->disconnect_is_shn = false;
    1014             :         } else {
    1015             :                 /* Only a subset of the registers are cleared out on a reset */
    1016           0 :                 ctrlr->vcprop.cc.raw = 0;
    1017           0 :                 ctrlr->vcprop.csts.raw = 0;
    1018             :         }
    1019             : 
    1020             :         /* After CC.EN transitions to 0 (due to shutdown or reset), the association
    1021             :          * between the host and controller shall be preserved for at least 2 minutes */
    1022           0 :         if (ctrlr->association_timer) {
    1023           0 :                 SPDK_DEBUGLOG(nvmf, "Association timer already set\n");
    1024           0 :                 nvmf_ctrlr_stop_association_timer(ctrlr);
    1025             :         }
    1026           0 :         if (ctrlr->association_timeout) {
    1027           0 :                 ctrlr->association_timer = SPDK_POLLER_REGISTER(nvmf_ctrlr_association_remove, ctrlr,
    1028             :                                            ctrlr->association_timeout * 1000);
    1029             :         }
    1030           0 :         ctrlr->disconnect_in_progress = false;
    1031           0 :         return SPDK_POLLER_BUSY;
    1032             : }
    1033             : 
    1034             : static void
    1035           0 : nvmf_ctrlr_cc_reset_shn_done(struct spdk_io_channel_iter *i, int status)
    1036             : {
    1037           0 :         struct spdk_nvmf_ctrlr *ctrlr = spdk_io_channel_iter_get_ctx(i);
    1038             : 
    1039           0 :         if (status < 0) {
    1040           0 :                 SPDK_ERRLOG("Fail to disconnect io ctrlr qpairs\n");
    1041           0 :                 assert(false);
    1042             :         }
    1043             : 
    1044           0 :         _nvmf_ctrlr_cc_reset_shn_done((void *)ctrlr);
    1045           0 : }
    1046             : 
    1047             : static void
    1048           0 : nvmf_bdev_complete_reset(struct spdk_bdev_io *bdev_io, bool success, void *cb_arg)
    1049             : {
    1050           0 :         SPDK_NOTICELOG("Resetting bdev done with %s\n", success ? "success" : "failure");
    1051             : 
    1052           0 :         spdk_bdev_free_io(bdev_io);
    1053           0 : }
    1054             : 
    1055             : 
    1056             : static int
    1057           0 : nvmf_ctrlr_cc_timeout(void *ctx)
    1058             : {
    1059           0 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
    1060           0 :         struct spdk_nvmf_poll_group *group = ctrlr->admin_qpair->group;
    1061             :         struct spdk_nvmf_ns *ns;
    1062             :         struct spdk_nvmf_subsystem_pg_ns_info *ns_info;
    1063             : 
    1064           0 :         assert(group != NULL && group->sgroups != NULL);
    1065           0 :         spdk_poller_unregister(&ctrlr->cc_timeout_timer);
    1066           0 :         SPDK_DEBUGLOG(nvmf, "Ctrlr %p reset or shutdown timeout\n", ctrlr);
    1067             : 
    1068           0 :         for (ns = spdk_nvmf_subsystem_get_first_ns(ctrlr->subsys); ns != NULL;
    1069           0 :              ns = spdk_nvmf_subsystem_get_next_ns(ctrlr->subsys, ns)) {
    1070           0 :                 if (ns->bdev == NULL) {
    1071           0 :                         continue;
    1072             :                 }
    1073           0 :                 ns_info = &group->sgroups[ctrlr->subsys->id].ns_info[ns->opts.nsid - 1];
    1074           0 :                 SPDK_NOTICELOG("Ctrlr %p resetting NSID %u\n", ctrlr, ns->opts.nsid);
    1075           0 :                 spdk_bdev_reset(ns->desc, ns_info->channel, nvmf_bdev_complete_reset, NULL);
    1076             :         }
    1077             : 
    1078           0 :         return SPDK_POLLER_BUSY;
    1079             : }
    1080             : 
    1081             : const struct spdk_nvmf_registers *
    1082           0 : spdk_nvmf_ctrlr_get_regs(struct spdk_nvmf_ctrlr *ctrlr)
    1083             : {
    1084           0 :         return &ctrlr->vcprop;
    1085             : }
    1086             : 
    1087             : void
    1088           0 : nvmf_ctrlr_set_fatal_status(struct spdk_nvmf_ctrlr *ctrlr)
    1089             : {
    1090           0 :         ctrlr->vcprop.csts.bits.cfs = 1;
    1091           0 : }
    1092             : 
    1093             : static uint64_t
    1094           0 : nvmf_prop_get_cap(struct spdk_nvmf_ctrlr *ctrlr)
    1095             : {
    1096           0 :         return ctrlr->vcprop.cap.raw;
    1097             : }
    1098             : 
    1099             : static uint64_t
    1100           0 : nvmf_prop_get_vs(struct spdk_nvmf_ctrlr *ctrlr)
    1101             : {
    1102           0 :         return ctrlr->vcprop.vs.raw;
    1103             : }
    1104             : 
    1105             : static uint64_t
    1106           1 : nvmf_prop_get_cc(struct spdk_nvmf_ctrlr *ctrlr)
    1107             : {
    1108           1 :         return ctrlr->vcprop.cc.raw;
    1109             : }
    1110             : 
    1111             : static bool
    1112           1 : nvmf_prop_set_cc(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1113             : {
    1114             :         union spdk_nvme_cc_register cc, diff;
    1115             :         uint32_t cc_timeout_ms;
    1116             : 
    1117           1 :         cc.raw = value;
    1118             : 
    1119           1 :         SPDK_DEBUGLOG(nvmf, "cur CC: 0x%08x\n", ctrlr->vcprop.cc.raw);
    1120           1 :         SPDK_DEBUGLOG(nvmf, "new CC: 0x%08x\n", cc.raw);
    1121             : 
    1122             :         /*
    1123             :          * Calculate which bits changed between the current and new CC.
    1124             :          * Mark each bit as 0 once it is handled to determine if any unhandled bits were changed.
    1125             :          */
    1126           1 :         diff.raw = cc.raw ^ ctrlr->vcprop.cc.raw;
    1127             : 
    1128           1 :         if (diff.bits.en) {
    1129           0 :                 if (cc.bits.en) {
    1130           0 :                         SPDK_DEBUGLOG(nvmf, "Property Set CC Enable!\n");
    1131           0 :                         nvmf_ctrlr_stop_association_timer(ctrlr);
    1132             : 
    1133           0 :                         ctrlr->vcprop.cc.bits.en = 1;
    1134           0 :                         ctrlr->vcprop.csts.bits.rdy = 1;
    1135             :                 } else {
    1136           0 :                         SPDK_DEBUGLOG(nvmf, "Property Set CC Disable!\n");
    1137           0 :                         if (ctrlr->disconnect_in_progress) {
    1138           0 :                                 SPDK_DEBUGLOG(nvmf, "Disconnect in progress\n");
    1139           0 :                                 return true;
    1140             :                         }
    1141             : 
    1142           0 :                         ctrlr->cc_timeout_timer = SPDK_POLLER_REGISTER(nvmf_ctrlr_cc_timeout, ctrlr,
    1143             :                                                   NVMF_CC_RESET_SHN_TIMEOUT_IN_MS * 1000);
    1144             :                         /* Make sure cc_timeout_ms is between cc_timeout_timer and Host reset/shutdown timeout */
    1145           0 :                         cc_timeout_ms = (NVMF_CC_RESET_SHN_TIMEOUT_IN_MS + NVMF_CTRLR_RESET_SHN_TIMEOUT_IN_MS) / 2;
    1146           0 :                         ctrlr->cc_timeout_tsc = spdk_get_ticks() + cc_timeout_ms * spdk_get_ticks_hz() / (uint64_t)1000;
    1147             : 
    1148           0 :                         ctrlr->vcprop.cc.bits.en = 0;
    1149           0 :                         ctrlr->disconnect_in_progress = true;
    1150           0 :                         ctrlr->disconnect_is_shn = false;
    1151           0 :                         spdk_for_each_channel(ctrlr->subsys->tgt,
    1152             :                                               nvmf_ctrlr_disconnect_io_qpairs_on_pg,
    1153             :                                               ctrlr,
    1154             :                                               nvmf_ctrlr_cc_reset_shn_done);
    1155             :                 }
    1156           0 :                 diff.bits.en = 0;
    1157             :         }
    1158             : 
    1159           1 :         if (diff.bits.shn) {
    1160           0 :                 if (cc.bits.shn == SPDK_NVME_SHN_NORMAL ||
    1161           0 :                     cc.bits.shn == SPDK_NVME_SHN_ABRUPT) {
    1162           0 :                         SPDK_DEBUGLOG(nvmf, "Property Set CC Shutdown %u%ub!\n",
    1163             :                                       cc.bits.shn >> 1, cc.bits.shn & 1);
    1164           0 :                         if (ctrlr->disconnect_in_progress) {
    1165           0 :                                 SPDK_DEBUGLOG(nvmf, "Disconnect in progress\n");
    1166           0 :                                 return true;
    1167             :                         }
    1168             : 
    1169           0 :                         ctrlr->cc_timeout_timer = SPDK_POLLER_REGISTER(nvmf_ctrlr_cc_timeout, ctrlr,
    1170             :                                                   NVMF_CC_RESET_SHN_TIMEOUT_IN_MS * 1000);
    1171             :                         /* Make sure cc_timeout_ms is between cc_timeout_timer and Host reset/shutdown timeout */
    1172           0 :                         cc_timeout_ms = (NVMF_CC_RESET_SHN_TIMEOUT_IN_MS + NVMF_CTRLR_RESET_SHN_TIMEOUT_IN_MS) / 2;
    1173           0 :                         ctrlr->cc_timeout_tsc = spdk_get_ticks() + cc_timeout_ms * spdk_get_ticks_hz() / (uint64_t)1000;
    1174             : 
    1175           0 :                         ctrlr->vcprop.cc.bits.shn = cc.bits.shn;
    1176           0 :                         ctrlr->disconnect_in_progress = true;
    1177           0 :                         ctrlr->disconnect_is_shn = true;
    1178           0 :                         spdk_for_each_channel(ctrlr->subsys->tgt,
    1179             :                                               nvmf_ctrlr_disconnect_io_qpairs_on_pg,
    1180             :                                               ctrlr,
    1181             :                                               nvmf_ctrlr_cc_reset_shn_done);
    1182             : 
    1183             :                         /* From the time a shutdown is initiated the controller shall disable
    1184             :                          * Keep Alive timer */
    1185           0 :                         nvmf_ctrlr_stop_keep_alive_timer(ctrlr);
    1186           0 :                 } else if (cc.bits.shn == 0) {
    1187           0 :                         ctrlr->vcprop.cc.bits.shn = 0;
    1188             :                 } else {
    1189           0 :                         SPDK_ERRLOG("Prop Set CC: Invalid SHN value %u%ub\n",
    1190             :                                     cc.bits.shn >> 1, cc.bits.shn & 1);
    1191           0 :                         return false;
    1192             :                 }
    1193           0 :                 diff.bits.shn = 0;
    1194             :         }
    1195             : 
    1196           1 :         if (diff.bits.iosqes) {
    1197           0 :                 SPDK_DEBUGLOG(nvmf, "Prop Set IOSQES = %u (%u bytes)\n",
    1198             :                               cc.bits.iosqes, 1u << cc.bits.iosqes);
    1199           0 :                 ctrlr->vcprop.cc.bits.iosqes = cc.bits.iosqes;
    1200           0 :                 diff.bits.iosqes = 0;
    1201             :         }
    1202             : 
    1203           1 :         if (diff.bits.iocqes) {
    1204           0 :                 SPDK_DEBUGLOG(nvmf, "Prop Set IOCQES = %u (%u bytes)\n",
    1205             :                               cc.bits.iocqes, 1u << cc.bits.iocqes);
    1206           0 :                 ctrlr->vcprop.cc.bits.iocqes = cc.bits.iocqes;
    1207           0 :                 diff.bits.iocqes = 0;
    1208             :         }
    1209             : 
    1210           1 :         if (diff.bits.ams) {
    1211           0 :                 SPDK_ERRLOG("Arbitration Mechanism Selected (AMS) 0x%x not supported!\n", cc.bits.ams);
    1212           0 :                 return false;
    1213             :         }
    1214             : 
    1215           1 :         if (diff.bits.mps) {
    1216           0 :                 SPDK_ERRLOG("Memory Page Size (MPS) %u KiB not supported!\n", (1 << (2 + cc.bits.mps)));
    1217           0 :                 return false;
    1218             :         }
    1219             : 
    1220           1 :         if (diff.bits.css) {
    1221           0 :                 if (cc.bits.css > SPDK_NVME_CC_CSS_IOCS) {
    1222           0 :                         SPDK_ERRLOG("I/O Command Set Selected (CSS) 0x%x not supported!\n", cc.bits.css);
    1223           0 :                         return false;
    1224             :                 }
    1225           0 :                 diff.bits.css = 0;
    1226             :         }
    1227             : 
    1228           1 :         if (diff.raw != 0) {
    1229             :                 /* Print an error message, but don't fail the command in this case.
    1230             :                  * If we did want to fail in this case, we'd need to ensure we acted
    1231             :                  * on no other bits or the initiator gets confused. */
    1232           0 :                 SPDK_ERRLOG("Prop Set CC toggled reserved bits 0x%x!\n", diff.raw);
    1233             :         }
    1234             : 
    1235           1 :         return true;
    1236             : }
    1237             : 
    1238             : static uint64_t
    1239           0 : nvmf_prop_get_csts(struct spdk_nvmf_ctrlr *ctrlr)
    1240             : {
    1241           0 :         return ctrlr->vcprop.csts.raw;
    1242             : }
    1243             : 
    1244             : static uint64_t
    1245           0 : nvmf_prop_get_aqa(struct spdk_nvmf_ctrlr *ctrlr)
    1246             : {
    1247           0 :         return ctrlr->vcprop.aqa.raw;
    1248             : }
    1249             : 
    1250             : static bool
    1251           0 : nvmf_prop_set_aqa(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1252             : {
    1253             :         union spdk_nvme_aqa_register aqa;
    1254             : 
    1255           0 :         aqa.raw = value;
    1256             : 
    1257             :         /*
    1258             :          * We don't need to explicitly check for maximum size, as the fields are
    1259             :          * limited to 12 bits (4096).
    1260             :          */
    1261           0 :         if (aqa.bits.asqs < SPDK_NVME_ADMIN_QUEUE_MIN_ENTRIES - 1 ||
    1262           0 :             aqa.bits.acqs < SPDK_NVME_ADMIN_QUEUE_MIN_ENTRIES - 1 ||
    1263           0 :             aqa.bits.reserved1 != 0 || aqa.bits.reserved2 != 0) {
    1264           0 :                 return false;
    1265             :         }
    1266             : 
    1267           0 :         ctrlr->vcprop.aqa.raw = value;
    1268             : 
    1269           0 :         return true;
    1270             : }
    1271             : 
    1272             : static uint64_t
    1273           1 : nvmf_prop_get_asq(struct spdk_nvmf_ctrlr *ctrlr)
    1274             : {
    1275           1 :         return ctrlr->vcprop.asq;
    1276             : }
    1277             : 
    1278             : static bool
    1279           1 : nvmf_prop_set_asq_lower(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1280             : {
    1281           1 :         ctrlr->vcprop.asq = (ctrlr->vcprop.asq & (0xFFFFFFFFULL << 32ULL)) | value;
    1282             : 
    1283           1 :         return true;
    1284             : }
    1285             : 
    1286             : static bool
    1287           0 : nvmf_prop_set_asq_upper(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1288             : {
    1289           0 :         ctrlr->vcprop.asq = (ctrlr->vcprop.asq & 0xFFFFFFFFULL) | ((uint64_t)value << 32ULL);
    1290             : 
    1291           0 :         return true;
    1292             : }
    1293             : 
    1294             : static uint64_t
    1295           0 : nvmf_prop_get_acq(struct spdk_nvmf_ctrlr *ctrlr)
    1296             : {
    1297           0 :         return ctrlr->vcprop.acq;
    1298             : }
    1299             : 
    1300             : static bool
    1301           0 : nvmf_prop_set_acq_lower(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1302             : {
    1303           0 :         ctrlr->vcprop.acq = (ctrlr->vcprop.acq & (0xFFFFFFFFULL << 32ULL)) | value;
    1304             : 
    1305           0 :         return true;
    1306             : }
    1307             : 
    1308             : static bool
    1309           0 : nvmf_prop_set_acq_upper(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value)
    1310             : {
    1311           0 :         ctrlr->vcprop.acq = (ctrlr->vcprop.acq & 0xFFFFFFFFULL) | ((uint64_t)value << 32ULL);
    1312             : 
    1313           0 :         return true;
    1314             : }
    1315             : 
    1316             : struct nvmf_prop {
    1317             :         uint32_t ofst;
    1318             :         uint8_t size;
    1319             :         char name[11];
    1320             :         uint64_t (*get_cb)(struct spdk_nvmf_ctrlr *ctrlr);
    1321             :         bool (*set_cb)(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value);
    1322             :         bool (*set_upper_cb)(struct spdk_nvmf_ctrlr *ctrlr, uint32_t value);
    1323             : };
    1324             : 
    1325             : #define PROP(field, size, get_cb, set_cb, set_upper_cb) \
    1326             :         { \
    1327             :                 offsetof(struct spdk_nvme_registers, field), \
    1328             :                 size, \
    1329             :                 #field, \
    1330             :                 get_cb, set_cb, set_upper_cb \
    1331             :         }
    1332             : 
    1333             : static const struct nvmf_prop nvmf_props[] = {
    1334             :         PROP(cap,  8, nvmf_prop_get_cap,  NULL,                    NULL),
    1335             :         PROP(vs,   4, nvmf_prop_get_vs,   NULL,                    NULL),
    1336             :         PROP(cc,   4, nvmf_prop_get_cc,   nvmf_prop_set_cc,        NULL),
    1337             :         PROP(csts, 4, nvmf_prop_get_csts, NULL,                    NULL),
    1338             :         PROP(aqa,  4, nvmf_prop_get_aqa,  nvmf_prop_set_aqa,       NULL),
    1339             :         PROP(asq,  8, nvmf_prop_get_asq,  nvmf_prop_set_asq_lower, nvmf_prop_set_asq_upper),
    1340             :         PROP(acq,  8, nvmf_prop_get_acq,  nvmf_prop_set_acq_lower, nvmf_prop_set_acq_upper),
    1341             : };
    1342             : 
    1343             : static const struct nvmf_prop *
    1344           6 : find_prop(uint32_t ofst, uint8_t size)
    1345             : {
    1346             :         size_t i;
    1347             : 
    1348          28 :         for (i = 0; i < SPDK_COUNTOF(nvmf_props); i++) {
    1349          27 :                 const struct nvmf_prop *prop = &nvmf_props[i];
    1350             : 
    1351          27 :                 if ((ofst >= prop->ofst) && (ofst + size <= prop->ofst + prop->size)) {
    1352           5 :                         return prop;
    1353             :                 }
    1354             :         }
    1355             : 
    1356           1 :         return NULL;
    1357             : }
    1358             : 
    1359             : static int
    1360           3 : nvmf_property_get(struct spdk_nvmf_request *req)
    1361             : {
    1362           3 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1363           3 :         struct spdk_nvmf_fabric_prop_get_cmd *cmd = &req->cmd->prop_get_cmd;
    1364           3 :         struct spdk_nvmf_fabric_prop_get_rsp *response = &req->rsp->prop_get_rsp;
    1365             :         const struct nvmf_prop *prop;
    1366             :         uint8_t size;
    1367             : 
    1368           3 :         response->status.sc = 0;
    1369           3 :         response->value.u64 = 0;
    1370             : 
    1371           3 :         SPDK_DEBUGLOG(nvmf, "size %d, offset 0x%x\n",
    1372             :                       cmd->attrib.size, cmd->ofst);
    1373             : 
    1374           3 :         switch (cmd->attrib.size) {
    1375           3 :         case SPDK_NVMF_PROP_SIZE_4:
    1376           3 :                 size = 4;
    1377           3 :                 break;
    1378           0 :         case SPDK_NVMF_PROP_SIZE_8:
    1379           0 :                 size = 8;
    1380           0 :                 break;
    1381           0 :         default:
    1382           0 :                 SPDK_DEBUGLOG(nvmf, "Invalid size value %d\n", cmd->attrib.size);
    1383           0 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1384           0 :                 response->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
    1385           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1386             :         }
    1387             : 
    1388           3 :         prop = find_prop(cmd->ofst, size);
    1389           3 :         if (prop == NULL || prop->get_cb == NULL) {
    1390           1 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1391           1 :                 response->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
    1392           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1393             :         }
    1394             : 
    1395           2 :         SPDK_DEBUGLOG(nvmf, "name: %s\n", prop->name);
    1396             : 
    1397           2 :         response->value.u64 = prop->get_cb(ctrlr);
    1398             : 
    1399           2 :         if (size != prop->size) {
    1400             :                 /* The size must be 4 and the prop->size is 8. Figure out which part of the property to read. */
    1401           1 :                 assert(size == 4);
    1402           1 :                 assert(prop->size == 8);
    1403             : 
    1404           1 :                 if (cmd->ofst == prop->ofst) {
    1405             :                         /* Keep bottom 4 bytes only */
    1406           1 :                         response->value.u64 &= 0xFFFFFFFF;
    1407             :                 } else {
    1408             :                         /* Keep top 4 bytes only */
    1409           0 :                         response->value.u64 >>= 32;
    1410             :                 }
    1411             :         }
    1412             : 
    1413           2 :         SPDK_DEBUGLOG(nvmf, "response value: 0x%" PRIx64 "\n", response->value.u64);
    1414             : 
    1415           2 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1416             : }
    1417             : 
    1418             : static int
    1419           3 : nvmf_property_set(struct spdk_nvmf_request *req)
    1420             : {
    1421           3 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1422           3 :         struct spdk_nvmf_fabric_prop_set_cmd *cmd = &req->cmd->prop_set_cmd;
    1423           3 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1424             :         const struct nvmf_prop *prop;
    1425             :         uint64_t value;
    1426             :         uint8_t size;
    1427             :         bool ret;
    1428             : 
    1429           3 :         SPDK_DEBUGLOG(nvmf, "size %d, offset 0x%x, value 0x%" PRIx64 "\n",
    1430             :                       cmd->attrib.size, cmd->ofst, cmd->value.u64);
    1431             : 
    1432           3 :         switch (cmd->attrib.size) {
    1433           3 :         case SPDK_NVMF_PROP_SIZE_4:
    1434           3 :                 size = 4;
    1435           3 :                 break;
    1436           0 :         case SPDK_NVMF_PROP_SIZE_8:
    1437           0 :                 size = 8;
    1438           0 :                 break;
    1439           0 :         default:
    1440           0 :                 SPDK_DEBUGLOG(nvmf, "Invalid size value %d\n", cmd->attrib.size);
    1441           0 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1442           0 :                 response->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
    1443           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1444             :         }
    1445             : 
    1446           3 :         prop = find_prop(cmd->ofst, size);
    1447           3 :         if (prop == NULL || prop->set_cb == NULL) {
    1448           1 :                 SPDK_INFOLOG(nvmf, "Invalid offset 0x%x\n", cmd->ofst);
    1449           1 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1450           1 :                 response->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
    1451           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1452             :         }
    1453             : 
    1454           2 :         SPDK_DEBUGLOG(nvmf, "name: %s\n", prop->name);
    1455             : 
    1456           2 :         value = cmd->value.u64;
    1457             : 
    1458           2 :         if (prop->size == 4) {
    1459           1 :                 ret = prop->set_cb(ctrlr, (uint32_t)value);
    1460           1 :         } else if (size != prop->size) {
    1461             :                 /* The size must be 4 and the prop->size is 8. Figure out which part of the property to write. */
    1462           1 :                 assert(size == 4);
    1463           1 :                 assert(prop->size == 8);
    1464             : 
    1465           1 :                 if (cmd->ofst == prop->ofst) {
    1466           1 :                         ret = prop->set_cb(ctrlr, (uint32_t)value);
    1467             :                 } else {
    1468           0 :                         ret = prop->set_upper_cb(ctrlr, (uint32_t)value);
    1469             :                 }
    1470             :         } else {
    1471           0 :                 ret = prop->set_cb(ctrlr, (uint32_t)value);
    1472           0 :                 if (ret) {
    1473           0 :                         ret = prop->set_upper_cb(ctrlr, (uint32_t)(value >> 32));
    1474             :                 }
    1475             :         }
    1476             : 
    1477           2 :         if (!ret) {
    1478           0 :                 SPDK_ERRLOG("prop set_cb failed\n");
    1479           0 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1480           0 :                 response->status.sc = SPDK_NVMF_FABRIC_SC_INVALID_PARAM;
    1481           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1482             :         }
    1483             : 
    1484           2 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1485             : }
    1486             : 
    1487             : static int
    1488           0 : nvmf_ctrlr_set_features_arbitration(struct spdk_nvmf_request *req)
    1489             : {
    1490           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1491           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1492             : 
    1493           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Arbitration (cdw11 = 0x%0x)\n", cmd->cdw11);
    1494             : 
    1495           0 :         ctrlr->feat.arbitration.raw = cmd->cdw11;
    1496           0 :         ctrlr->feat.arbitration.bits.reserved = 0;
    1497             : 
    1498           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1499             : }
    1500             : 
    1501             : static int
    1502           0 : nvmf_ctrlr_set_features_power_management(struct spdk_nvmf_request *req)
    1503             : {
    1504           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1505           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1506           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1507             : 
    1508           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Power Management (cdw11 = 0x%0x)\n", cmd->cdw11);
    1509             : 
    1510             :         /* Only PS = 0 is allowed, since we report NPSS = 0 */
    1511           0 :         if (cmd->cdw11_bits.feat_power_management.bits.ps != 0) {
    1512           0 :                 SPDK_ERRLOG("Invalid power state %u\n", cmd->cdw11_bits.feat_power_management.bits.ps);
    1513           0 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    1514           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1515           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1516             :         }
    1517             : 
    1518           0 :         ctrlr->feat.power_management.raw = cmd->cdw11;
    1519           0 :         ctrlr->feat.power_management.bits.reserved = 0;
    1520             : 
    1521           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1522             : }
    1523             : 
    1524             : static bool
    1525           5 : temp_threshold_opts_valid(const union spdk_nvme_feat_temperature_threshold *opts)
    1526             : {
    1527             :         /*
    1528             :          * Valid TMPSEL values:
    1529             :          *  0000b - 1000b: temperature sensors
    1530             :          *  1111b: set all implemented temperature sensors
    1531             :          */
    1532           5 :         if (opts->bits.tmpsel >= 9 && opts->bits.tmpsel != 15) {
    1533             :                 /* 1001b - 1110b: reserved */
    1534           2 :                 SPDK_ERRLOG("Invalid TMPSEL %u\n", opts->bits.tmpsel);
    1535           2 :                 return false;
    1536             :         }
    1537             : 
    1538             :         /*
    1539             :          * Valid THSEL values:
    1540             :          *  00b: over temperature threshold
    1541             :          *  01b: under temperature threshold
    1542             :          */
    1543           3 :         if (opts->bits.thsel > 1) {
    1544             :                 /* 10b - 11b: reserved */
    1545           1 :                 SPDK_ERRLOG("Invalid THSEL %u\n", opts->bits.thsel);
    1546           1 :                 return false;
    1547             :         }
    1548             : 
    1549           2 :         return true;
    1550             : }
    1551             : 
    1552             : static int
    1553           3 : nvmf_ctrlr_set_features_temperature_threshold(struct spdk_nvmf_request *req)
    1554             : {
    1555           3 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1556           3 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1557             : 
    1558           3 :         SPDK_DEBUGLOG(nvmf, "Set Features - Temperature Threshold (cdw11 = 0x%0x)\n", cmd->cdw11);
    1559             : 
    1560           3 :         if (!temp_threshold_opts_valid(&cmd->cdw11_bits.feat_temp_threshold)) {
    1561           2 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    1562           2 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1563           2 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1564             :         }
    1565             : 
    1566             :         /* TODO: no sensors implemented - ignore new values */
    1567           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1568             : }
    1569             : 
    1570             : static int
    1571           2 : nvmf_ctrlr_get_features_temperature_threshold(struct spdk_nvmf_request *req)
    1572             : {
    1573           2 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1574           2 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1575             : 
    1576           2 :         SPDK_DEBUGLOG(nvmf, "Get Features - Temperature Threshold (cdw11 = 0x%0x)\n", cmd->cdw11);
    1577             : 
    1578           2 :         if (!temp_threshold_opts_valid(&cmd->cdw11_bits.feat_temp_threshold)) {
    1579           1 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    1580           1 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1581           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1582             :         }
    1583             : 
    1584             :         /* TODO: no sensors implemented - return 0 for all thresholds */
    1585           1 :         rsp->cdw0 = 0;
    1586             : 
    1587           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1588             : }
    1589             : 
    1590             : static int
    1591           0 : nvmf_ctrlr_get_features_interrupt_vector_configuration(struct spdk_nvmf_request *req)
    1592             : {
    1593           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1594           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1595           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1596           0 :         union spdk_nvme_feat_interrupt_vector_configuration iv_conf = {};
    1597             : 
    1598           0 :         SPDK_DEBUGLOG(nvmf, "Get Features - Interrupt Vector Configuration (cdw11 = 0x%0x)\n", cmd->cdw11);
    1599             : 
    1600           0 :         iv_conf.bits.iv = cmd->cdw11_bits.feat_interrupt_vector_configuration.bits.iv;
    1601           0 :         iv_conf.bits.cd = ctrlr->feat.interrupt_vector_configuration.bits.cd;
    1602           0 :         rsp->cdw0 = iv_conf.raw;
    1603             : 
    1604           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1605             : }
    1606             : 
    1607             : static int
    1608           2 : nvmf_ctrlr_set_features_error_recovery(struct spdk_nvmf_request *req)
    1609             : {
    1610           2 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1611           2 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1612           2 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1613             : 
    1614           2 :         SPDK_DEBUGLOG(nvmf, "Set Features - Error Recovery (cdw11 = 0x%0x)\n", cmd->cdw11);
    1615             : 
    1616           2 :         if (cmd->cdw11_bits.feat_error_recovery.bits.dulbe) {
    1617             :                 /*
    1618             :                  * Host is not allowed to set this bit, since we don't advertise it in
    1619             :                  * Identify Namespace.
    1620             :                  */
    1621           1 :                 SPDK_ERRLOG("Host set unsupported DULBE bit\n");
    1622           1 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    1623           1 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1624           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1625             :         }
    1626             : 
    1627           1 :         ctrlr->feat.error_recovery.raw = cmd->cdw11;
    1628           1 :         ctrlr->feat.error_recovery.bits.reserved = 0;
    1629             : 
    1630           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1631             : }
    1632             : 
    1633             : static int
    1634           0 : nvmf_ctrlr_set_features_volatile_write_cache(struct spdk_nvmf_request *req)
    1635             : {
    1636           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1637           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1638             : 
    1639           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Volatile Write Cache (cdw11 = 0x%0x)\n", cmd->cdw11);
    1640             : 
    1641           0 :         ctrlr->feat.volatile_write_cache.raw = cmd->cdw11;
    1642           0 :         ctrlr->feat.volatile_write_cache.bits.reserved = 0;
    1643             : 
    1644           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Volatile Write Cache %s\n",
    1645             :                       ctrlr->feat.volatile_write_cache.bits.wce ? "Enabled" : "Disabled");
    1646           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1647             : }
    1648             : 
    1649             : static int
    1650           0 : nvmf_ctrlr_set_features_write_atomicity(struct spdk_nvmf_request *req)
    1651             : {
    1652           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1653           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1654             : 
    1655           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Write Atomicity (cdw11 = 0x%0x)\n", cmd->cdw11);
    1656             : 
    1657           0 :         ctrlr->feat.write_atomicity.raw = cmd->cdw11;
    1658           0 :         ctrlr->feat.write_atomicity.bits.reserved = 0;
    1659             : 
    1660           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1661             : }
    1662             : 
    1663             : static int
    1664           0 : nvmf_ctrlr_set_features_host_identifier(struct spdk_nvmf_request *req)
    1665             : {
    1666           0 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1667             : 
    1668           0 :         SPDK_ERRLOG("Set Features - Host Identifier not allowed\n");
    1669           0 :         response->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    1670           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1671             : }
    1672             : 
    1673             : static int
    1674           0 : nvmf_ctrlr_get_features_host_identifier(struct spdk_nvmf_request *req)
    1675             : {
    1676           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1677           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1678           0 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1679           0 :         struct spdk_iov_xfer ix;
    1680             : 
    1681           0 :         SPDK_DEBUGLOG(nvmf, "Get Features - Host Identifier\n");
    1682             : 
    1683           0 :         if (!cmd->cdw11_bits.feat_host_identifier.bits.exhid) {
    1684             :                 /* NVMe over Fabrics requires EXHID=1 (128-bit/16-byte host ID) */
    1685           0 :                 SPDK_ERRLOG("Get Features - Host Identifier with EXHID=0 not allowed\n");
    1686           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1687           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1688             :         }
    1689             : 
    1690           0 :         if (req->iovcnt < 1 || req->length < sizeof(ctrlr->hostid)) {
    1691           0 :                 SPDK_ERRLOG("Invalid data buffer for Get Features - Host Identifier\n");
    1692           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1693           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1694             :         }
    1695             : 
    1696           0 :         spdk_iov_xfer_init(&ix, req->iov, req->iovcnt);
    1697           0 :         spdk_iov_xfer_from_buf(&ix, &ctrlr->hostid, sizeof(ctrlr->hostid));
    1698             : 
    1699           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1700             : }
    1701             : 
    1702             : static int
    1703           0 : nvmf_ctrlr_get_features_reservation_notification_mask(struct spdk_nvmf_request *req)
    1704             : {
    1705           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1706           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1707           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1708             :         struct spdk_nvmf_ns *ns;
    1709             : 
    1710           0 :         SPDK_DEBUGLOG(nvmf, "get Features - Reservation Notification Mask\n");
    1711             : 
    1712           0 :         if (cmd->nsid == SPDK_NVME_GLOBAL_NS_TAG) {
    1713           0 :                 SPDK_ERRLOG("get Features - Invalid Namespace ID\n");
    1714           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1715           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1716             :         }
    1717             : 
    1718           0 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, cmd->nsid);
    1719           0 :         if (ns == NULL) {
    1720           0 :                 SPDK_ERRLOG("Set Features - Invalid Namespace ID\n");
    1721           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1722           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1723             :         }
    1724           0 :         rsp->cdw0 = ns->mask;
    1725             : 
    1726           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1727             : }
    1728             : 
    1729             : static int
    1730           0 : nvmf_ctrlr_set_features_reservation_notification_mask(struct spdk_nvmf_request *req)
    1731             : {
    1732           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1733           0 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    1734           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1735           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1736             :         struct spdk_nvmf_ns *ns;
    1737             : 
    1738           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Reservation Notification Mask\n");
    1739             : 
    1740           0 :         if (cmd->nsid == SPDK_NVME_GLOBAL_NS_TAG) {
    1741           0 :                 for (ns = spdk_nvmf_subsystem_get_first_ns(subsystem); ns != NULL;
    1742           0 :                      ns = spdk_nvmf_subsystem_get_next_ns(subsystem, ns)) {
    1743           0 :                         ns->mask = cmd->cdw11;
    1744             :                 }
    1745           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1746             :         }
    1747             : 
    1748           0 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, cmd->nsid);
    1749           0 :         if (ns == NULL) {
    1750           0 :                 SPDK_ERRLOG("Set Features - Invalid Namespace ID\n");
    1751           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1752           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1753             :         }
    1754           0 :         ns->mask = cmd->cdw11;
    1755             : 
    1756           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1757             : }
    1758             : 
    1759             : static int
    1760           1 : nvmf_ctrlr_get_features_reservation_persistence(struct spdk_nvmf_request *req)
    1761             : {
    1762           1 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1763           1 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1764           1 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1765             :         struct spdk_nvmf_ns *ns;
    1766             : 
    1767           1 :         SPDK_DEBUGLOG(nvmf, "Get Features - Reservation Persistence\n");
    1768             : 
    1769           1 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, cmd->nsid);
    1770             :         /* NSID with SPDK_NVME_GLOBAL_NS_TAG (=0xffffffff) also included */
    1771           1 :         if (ns == NULL) {
    1772           0 :                 SPDK_ERRLOG("Get Features - Invalid Namespace ID\n");
    1773           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1774           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1775           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1776             :         }
    1777             : 
    1778           1 :         response->cdw0 = ns->ptpl_activated;
    1779             : 
    1780           1 :         response->status.sct = SPDK_NVME_SCT_GENERIC;
    1781           1 :         response->status.sc = SPDK_NVME_SC_SUCCESS;
    1782           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1783             : }
    1784             : 
    1785             : static int
    1786           1 : nvmf_ctrlr_set_features_reservation_persistence(struct spdk_nvmf_request *req)
    1787             : {
    1788           1 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1789           1 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1790           1 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1791             :         struct spdk_nvmf_ns *ns;
    1792             :         bool ptpl;
    1793             : 
    1794           1 :         SPDK_DEBUGLOG(nvmf, "Set Features - Reservation Persistence\n");
    1795             : 
    1796           1 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, cmd->nsid);
    1797           1 :         ptpl = cmd->cdw11_bits.feat_rsv_persistence.bits.ptpl;
    1798             : 
    1799           1 :         if (cmd->nsid != SPDK_NVME_GLOBAL_NS_TAG && ns && nvmf_ns_is_ptpl_capable(ns)) {
    1800           1 :                 ns->ptpl_activated = ptpl;
    1801           0 :         } else if (cmd->nsid == SPDK_NVME_GLOBAL_NS_TAG) {
    1802           0 :                 for (ns = spdk_nvmf_subsystem_get_first_ns(ctrlr->subsys); ns;
    1803           0 :                      ns = spdk_nvmf_subsystem_get_next_ns(ctrlr->subsys, ns)) {
    1804           0 :                         if (nvmf_ns_is_ptpl_capable(ns)) {
    1805           0 :                                 ns->ptpl_activated = ptpl;
    1806             :                         }
    1807             :                 }
    1808             :         } else {
    1809           0 :                 SPDK_ERRLOG("Set Features - Invalid Namespace ID or Reservation Configuration\n");
    1810           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1811           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1812           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1813             :         }
    1814             : 
    1815             :         /* TODO: Feature not changeable for now */
    1816           1 :         response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    1817           1 :         response->status.sc = SPDK_NVME_SC_FEATURE_ID_NOT_SAVEABLE;
    1818           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1819             : }
    1820             : 
    1821             : static int
    1822           3 : nvmf_ctrlr_get_features_host_behavior_support(struct spdk_nvmf_request *req)
    1823             : {
    1824           3 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1825           3 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1826           3 :         struct spdk_nvme_host_behavior host_behavior = {};
    1827           3 :         struct spdk_iov_xfer ix;
    1828             : 
    1829           3 :         SPDK_DEBUGLOG(nvmf, "Get Features - Host Behavior Support\n");
    1830             : 
    1831           3 :         if (req->iovcnt < 1 || req->length < sizeof(struct spdk_nvme_host_behavior)) {
    1832           2 :                 SPDK_ERRLOG("invalid data buffer for Host Behavior Support\n");
    1833           2 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1834           2 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1835           2 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1836             :         }
    1837             : 
    1838           1 :         host_behavior.acre = ctrlr->acre_enabled;
    1839             : 
    1840           1 :         spdk_iov_xfer_init(&ix, req->iov, req->iovcnt);
    1841           1 :         spdk_iov_xfer_from_buf(&ix, &host_behavior, sizeof(host_behavior));
    1842             : 
    1843           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1844             : }
    1845             : 
    1846             : static int
    1847           5 : nvmf_ctrlr_set_features_host_behavior_support(struct spdk_nvmf_request *req)
    1848             : {
    1849           5 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1850           5 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    1851             :         struct spdk_nvme_host_behavior *host_behavior;
    1852             : 
    1853           5 :         SPDK_DEBUGLOG(nvmf, "Set Features - Host Behavior Support\n");
    1854           5 :         if (req->iovcnt != 1) {
    1855           1 :                 SPDK_ERRLOG("Host Behavior Support invalid iovcnt: %d\n", req->iovcnt);
    1856           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1857           1 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1858           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1859             :         }
    1860           4 :         if (req->iov[0].iov_len != sizeof(struct spdk_nvme_host_behavior)) {
    1861           1 :                 SPDK_ERRLOG("Host Behavior Support invalid iov_len: %zd\n", req->iov[0].iov_len);
    1862           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1863           1 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1864           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1865             :         }
    1866             : 
    1867           3 :         host_behavior = (struct spdk_nvme_host_behavior *)req->iov[0].iov_base;
    1868           3 :         if (host_behavior->acre == 0) {
    1869           1 :                 ctrlr->acre_enabled = false;
    1870           2 :         } else if (host_behavior->acre == 1) {
    1871           1 :                 ctrlr->acre_enabled = true;
    1872             :         } else {
    1873           1 :                 SPDK_ERRLOG("Host Behavior Support invalid acre: 0x%02x\n", host_behavior->acre);
    1874           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    1875           1 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1876           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1877             :         }
    1878           2 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1879             : }
    1880             : 
    1881             : static int
    1882           0 : nvmf_ctrlr_set_features_keep_alive_timer(struct spdk_nvmf_request *req)
    1883             : {
    1884           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1885           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1886           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1887             : 
    1888           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Keep Alive Timer (%u ms)\n", cmd->cdw11);
    1889             : 
    1890             :         /*
    1891             :          * if attempts to disable keep alive by setting kato to 0h
    1892             :          * a status value of keep alive invalid shall be returned
    1893             :          */
    1894           0 :         if (cmd->cdw11_bits.feat_keep_alive_timer.bits.kato == 0) {
    1895           0 :                 rsp->status.sc = SPDK_NVME_SC_KEEP_ALIVE_INVALID;
    1896           0 :         } else if (cmd->cdw11_bits.feat_keep_alive_timer.bits.kato < MIN_KEEP_ALIVE_TIMEOUT_IN_MS) {
    1897           0 :                 ctrlr->feat.keep_alive_timer.bits.kato = MIN_KEEP_ALIVE_TIMEOUT_IN_MS;
    1898             :         } else {
    1899             :                 /* round up to milliseconds */
    1900           0 :                 ctrlr->feat.keep_alive_timer.bits.kato = spdk_divide_round_up(
    1901           0 :                                         cmd->cdw11_bits.feat_keep_alive_timer.bits.kato,
    1902             :                                         KAS_DEFAULT_VALUE * KAS_TIME_UNIT_IN_MS) *
    1903             :                                 KAS_DEFAULT_VALUE * KAS_TIME_UNIT_IN_MS;
    1904             :         }
    1905             : 
    1906             :         /*
    1907             :          * if change the keep alive timeout value successfully
    1908             :          * update the keep alive poller.
    1909             :          */
    1910           0 :         if (cmd->cdw11_bits.feat_keep_alive_timer.bits.kato != 0) {
    1911           0 :                 if (ctrlr->keep_alive_poller != NULL) {
    1912           0 :                         spdk_poller_unregister(&ctrlr->keep_alive_poller);
    1913             :                 }
    1914           0 :                 ctrlr->keep_alive_poller = SPDK_POLLER_REGISTER(nvmf_ctrlr_keep_alive_poll, ctrlr,
    1915             :                                            ctrlr->feat.keep_alive_timer.bits.kato * 1000);
    1916             :         }
    1917             : 
    1918           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Keep Alive Timer set to %u ms\n",
    1919             :                       ctrlr->feat.keep_alive_timer.bits.kato);
    1920             : 
    1921           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1922             : }
    1923             : 
    1924             : static int
    1925           0 : nvmf_ctrlr_set_features_number_of_queues(struct spdk_nvmf_request *req)
    1926             : {
    1927           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    1928           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    1929           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    1930             :         uint32_t count;
    1931             : 
    1932           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Number of Queues, cdw11 0x%x\n",
    1933             :                       req->cmd->nvme_cmd.cdw11);
    1934             : 
    1935           0 :         if (cmd->cdw11_bits.feat_num_of_queues.bits.ncqr == UINT16_MAX ||
    1936           0 :             cmd->cdw11_bits.feat_num_of_queues.bits.nsqr == UINT16_MAX) {
    1937           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    1938           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1939             :         }
    1940             : 
    1941           0 :         count = spdk_bit_array_count_set(ctrlr->qpair_mask);
    1942             :         /* verify that the controller is ready to process commands */
    1943           0 :         if (count > 1) {
    1944           0 :                 SPDK_DEBUGLOG(nvmf, "Queue pairs already active!\n");
    1945           0 :                 rsp->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    1946             :         } else {
    1947             :                 /*
    1948             :                  * Ignore the value requested by the host -
    1949             :                  * always return the pre-configured value based on max_qpairs_allowed.
    1950             :                  */
    1951           0 :                 rsp->cdw0 = ctrlr->feat.number_of_queues.raw;
    1952             :         }
    1953             : 
    1954           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    1955             : }
    1956             : 
    1957             : SPDK_STATIC_ASSERT(sizeof(struct spdk_nvmf_ctrlr) == 4920,
    1958             :                    "Please check migration fields that need to be added or not");
    1959             : 
    1960             : static void
    1961           0 : nvmf_ctrlr_migr_data_copy(struct spdk_nvmf_ctrlr_migr_data *data,
    1962             :                           const struct spdk_nvmf_ctrlr_migr_data *data_src, size_t data_size)
    1963             : {
    1964           0 :         assert(data);
    1965           0 :         assert(data_src);
    1966           0 :         assert(data_size);
    1967             : 
    1968           0 :         memcpy(&data->regs, &data_src->regs, spdk_min(data->regs_size, data_src->regs_size));
    1969           0 :         memcpy(&data->feat, &data_src->feat, spdk_min(data->feat_size, data_src->feat_size));
    1970             : 
    1971             : #define SET_FIELD(field) \
    1972             :     if (offsetof(struct spdk_nvmf_ctrlr_migr_data, field) + sizeof(data->field) <= data_size) { \
    1973             :         data->field = data_src->field; \
    1974             :     } \
    1975             : 
    1976           0 :         SET_FIELD(cntlid);
    1977           0 :         SET_FIELD(acre);
    1978           0 :         SET_FIELD(num_aer_cids);
    1979           0 :         SET_FIELD(num_async_events);
    1980           0 :         SET_FIELD(notice_aen_mask);
    1981             : #undef SET_FIELD
    1982             : 
    1983             : #define SET_ARRAY(arr) \
    1984             :     if (offsetof(struct spdk_nvmf_ctrlr_migr_data, arr) + sizeof(data->arr) <= data_size) { \
    1985             :         memcpy(&data->arr, &data_src->arr, sizeof(data->arr)); \
    1986             :     } \
    1987             : 
    1988           0 :         SET_ARRAY(async_events);
    1989           0 :         SET_ARRAY(aer_cids);
    1990             : #undef SET_ARRAY
    1991           0 : }
    1992             : 
    1993             : int
    1994           0 : spdk_nvmf_ctrlr_save_migr_data(struct spdk_nvmf_ctrlr *ctrlr,
    1995             :                                struct spdk_nvmf_ctrlr_migr_data *data)
    1996             : {
    1997             :         struct spdk_nvmf_async_event_completion *event, *event_tmp;
    1998             :         uint32_t i;
    1999           0 :         struct spdk_nvmf_ctrlr_migr_data data_local = {
    2000             :                 .data_size = offsetof(struct spdk_nvmf_ctrlr_migr_data, unused),
    2001             :                 .regs_size = sizeof(struct spdk_nvmf_registers),
    2002             :                 .feat_size = sizeof(struct spdk_nvmf_ctrlr_feat)
    2003             :         };
    2004             : 
    2005           0 :         assert(data->data_size <= sizeof(data_local));
    2006           0 :         assert(spdk_get_thread() == ctrlr->thread);
    2007             : 
    2008           0 :         memcpy(&data_local.regs, &ctrlr->vcprop, sizeof(struct spdk_nvmf_registers));
    2009           0 :         memcpy(&data_local.feat, &ctrlr->feat, sizeof(struct spdk_nvmf_ctrlr_feat));
    2010             : 
    2011           0 :         data_local.cntlid = ctrlr->cntlid;
    2012           0 :         data_local.acre = ctrlr->acre_enabled;
    2013           0 :         data_local.num_aer_cids = ctrlr->nr_aer_reqs;
    2014             : 
    2015           0 :         STAILQ_FOREACH_SAFE(event, &ctrlr->async_events, link, event_tmp) {
    2016           0 :                 if (data_local.num_async_events + 1 > SPDK_NVMF_MIGR_MAX_PENDING_AERS) {
    2017           0 :                         SPDK_ERRLOG("ctrlr %p has too many pending AERs\n", ctrlr);
    2018           0 :                         break;
    2019             :                 }
    2020             : 
    2021           0 :                 data_local.async_events[data_local.num_async_events++].raw = event->event.raw;
    2022             :         }
    2023             : 
    2024           0 :         for (i = 0; i < ctrlr->nr_aer_reqs; i++) {
    2025           0 :                 struct spdk_nvmf_request *req = ctrlr->aer_req[i];
    2026           0 :                 data_local.aer_cids[i] = req->cmd->nvme_cmd.cid;
    2027             :         }
    2028           0 :         data_local.notice_aen_mask = ctrlr->notice_aen_mask;
    2029             : 
    2030           0 :         nvmf_ctrlr_migr_data_copy(data, &data_local, spdk_min(data->data_size, data_local.data_size));
    2031           0 :         return 0;
    2032             : }
    2033             : 
    2034             : int
    2035           0 : spdk_nvmf_ctrlr_restore_migr_data(struct spdk_nvmf_ctrlr *ctrlr,
    2036             :                                   const struct spdk_nvmf_ctrlr_migr_data *data)
    2037             : {
    2038             :         uint32_t i;
    2039           0 :         struct spdk_nvmf_ctrlr_migr_data data_local = {
    2040             :                 .data_size = offsetof(struct spdk_nvmf_ctrlr_migr_data, unused),
    2041             :                 .regs_size = sizeof(struct spdk_nvmf_registers),
    2042             :                 .feat_size = sizeof(struct spdk_nvmf_ctrlr_feat)
    2043             :         };
    2044             : 
    2045           0 :         assert(data->data_size <= sizeof(data_local));
    2046           0 :         assert(spdk_get_thread() == ctrlr->thread);
    2047             : 
    2048             :         /* local version of data should have defaults set before copy */
    2049           0 :         nvmf_ctrlr_migr_data_copy(&data_local, data, spdk_min(data->data_size, data_local.data_size));
    2050           0 :         memcpy(&ctrlr->vcprop, &data_local.regs, sizeof(struct spdk_nvmf_registers));
    2051           0 :         memcpy(&ctrlr->feat, &data_local.feat, sizeof(struct spdk_nvmf_ctrlr_feat));
    2052             : 
    2053           0 :         ctrlr->cntlid = data_local.cntlid;
    2054           0 :         ctrlr->acre_enabled = data_local.acre;
    2055             : 
    2056           0 :         for (i = 0; i < data_local.num_async_events; i++) {
    2057             :                 struct spdk_nvmf_async_event_completion *event;
    2058             : 
    2059           0 :                 event = calloc(1, sizeof(*event));
    2060           0 :                 if (!event) {
    2061           0 :                         return -ENOMEM;
    2062             :                 }
    2063             : 
    2064           0 :                 event->event.raw = data_local.async_events[i].raw;
    2065           0 :                 STAILQ_INSERT_TAIL(&ctrlr->async_events, event, link);
    2066             :         }
    2067           0 :         ctrlr->notice_aen_mask = data_local.notice_aen_mask;
    2068             : 
    2069           0 :         return 0;
    2070             : }
    2071             : 
    2072             : static int
    2073           0 : nvmf_ctrlr_set_features_async_event_configuration(struct spdk_nvmf_request *req)
    2074             : {
    2075           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    2076           0 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    2077             : 
    2078           0 :         SPDK_DEBUGLOG(nvmf, "Set Features - Async Event Configuration, cdw11 0x%08x\n",
    2079             :                       cmd->cdw11);
    2080           0 :         ctrlr->feat.async_event_configuration.raw = cmd->cdw11;
    2081           0 :         ctrlr->feat.async_event_configuration.bits.reserved1 = 0;
    2082           0 :         ctrlr->feat.async_event_configuration.bits.reserved2 = 0;
    2083           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2084             : }
    2085             : 
    2086             : static int
    2087          10 : nvmf_ctrlr_async_event_request(struct spdk_nvmf_request *req)
    2088             : {
    2089          10 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    2090          10 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    2091             :         struct spdk_nvmf_async_event_completion *pending_event;
    2092             : 
    2093          10 :         SPDK_DEBUGLOG(nvmf, "Async Event Request\n");
    2094             : 
    2095             :         /* Four asynchronous events are supported for now */
    2096          10 :         if (ctrlr->nr_aer_reqs >= SPDK_NVMF_MAX_ASYNC_EVENTS) {
    2097           1 :                 SPDK_DEBUGLOG(nvmf, "AERL exceeded\n");
    2098           1 :                 rsp->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    2099           1 :                 rsp->status.sc = SPDK_NVME_SC_ASYNC_EVENT_REQUEST_LIMIT_EXCEEDED;
    2100           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2101             :         }
    2102             : 
    2103           9 :         if (!STAILQ_EMPTY(&ctrlr->async_events)) {
    2104           4 :                 pending_event = STAILQ_FIRST(&ctrlr->async_events);
    2105           4 :                 rsp->cdw0 = pending_event->event.raw;
    2106           4 :                 STAILQ_REMOVE(&ctrlr->async_events, pending_event, spdk_nvmf_async_event_completion, link);
    2107           4 :                 free(pending_event);
    2108           4 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2109             :         }
    2110             : 
    2111           5 :         ctrlr->aer_req[ctrlr->nr_aer_reqs++] = req;
    2112           5 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
    2113             : }
    2114             : 
    2115             : static void
    2116           0 : nvmf_get_firmware_slot_log_page(struct iovec *iovs, int iovcnt, uint64_t offset, uint32_t length)
    2117             : {
    2118           0 :         struct spdk_nvme_firmware_page fw_page;
    2119             :         size_t copy_len;
    2120           0 :         struct spdk_iov_xfer ix;
    2121             : 
    2122           0 :         spdk_iov_xfer_init(&ix, iovs, iovcnt);
    2123             : 
    2124           0 :         memset(&fw_page, 0, sizeof(fw_page));
    2125           0 :         fw_page.afi.active_slot = 1;
    2126           0 :         fw_page.afi.next_reset_slot = 0;
    2127           0 :         spdk_strcpy_pad(fw_page.revision[0], FW_VERSION, sizeof(fw_page.revision[0]), ' ');
    2128             : 
    2129           0 :         if (offset < sizeof(fw_page)) {
    2130           0 :                 copy_len = spdk_min(sizeof(fw_page) - offset, length);
    2131           0 :                 if (copy_len > 0) {
    2132           0 :                         spdk_iov_xfer_from_buf(&ix, (const char *)&fw_page + offset, copy_len);
    2133             :                 }
    2134             :         }
    2135           0 : }
    2136             : 
    2137             : /*
    2138             :  * Asynchronous Event Mask Bit
    2139             :  */
    2140             : enum spdk_nvme_async_event_mask_bit {
    2141             :         /* Mask Namespace Change Notification */
    2142             :         SPDK_NVME_ASYNC_EVENT_NS_ATTR_CHANGE_MASK_BIT           = 0,
    2143             :         /* Mask Asymmetric Namespace Access Change Notification */
    2144             :         SPDK_NVME_ASYNC_EVENT_ANA_CHANGE_MASK_BIT               = 1,
    2145             :         /* Mask Discovery Log Change Notification */
    2146             :         SPDK_NVME_ASYNC_EVENT_DISCOVERY_LOG_CHANGE_MASK_BIT     = 2,
    2147             :         /* Mask Reservation Log Page Available Notification */
    2148             :         SPDK_NVME_ASYNC_EVENT_RESERVATION_LOG_AVAIL_MASK_BIT    = 3,
    2149             :         /* Mask Error Event */
    2150             :         SPDK_NVME_ASYNC_EVENT_ERROR_MASK_BIT                    = 4,
    2151             :         /* 4 - 63 Reserved */
    2152             : };
    2153             : 
    2154             : static inline void
    2155          20 : nvmf_ctrlr_unmask_aen(struct spdk_nvmf_ctrlr *ctrlr,
    2156             :                       enum spdk_nvme_async_event_mask_bit mask)
    2157             : {
    2158          20 :         ctrlr->notice_aen_mask &= ~(1 << mask);
    2159          20 : }
    2160             : 
    2161             : static inline bool
    2162          11 : nvmf_ctrlr_mask_aen(struct spdk_nvmf_ctrlr *ctrlr,
    2163             :                     enum spdk_nvme_async_event_mask_bit mask)
    2164             : {
    2165          11 :         if (ctrlr->notice_aen_mask & (1 << mask)) {
    2166           5 :                 return false;
    2167             :         } else {
    2168           6 :                 ctrlr->notice_aen_mask |= (1 << mask);
    2169           6 :                 return true;
    2170             :         }
    2171             : }
    2172             : 
    2173             : /* we have to use the typedef in the function declaration to appease astyle. */
    2174             : typedef enum spdk_nvme_ana_state spdk_nvme_ana_state_t;
    2175             : 
    2176             : static inline spdk_nvme_ana_state_t
    2177          36 : nvmf_ctrlr_get_ana_state(struct spdk_nvmf_ctrlr *ctrlr, uint32_t anagrpid)
    2178             : {
    2179          36 :         if (!ctrlr->subsys->flags.ana_reporting) {
    2180          36 :                 return SPDK_NVME_ANA_OPTIMIZED_STATE;
    2181             :         }
    2182             : 
    2183           0 :         if (spdk_unlikely(ctrlr->listener == NULL)) {
    2184           0 :                 return SPDK_NVME_ANA_INACCESSIBLE_STATE;
    2185             :         }
    2186             : 
    2187           0 :         assert(anagrpid - 1 < ctrlr->subsys->max_nsid);
    2188           0 :         return ctrlr->listener->ana_state[anagrpid - 1];
    2189             : }
    2190             : 
    2191             : static spdk_nvme_ana_state_t
    2192           8 : nvmf_ctrlr_get_ana_state_from_nsid(struct spdk_nvmf_ctrlr *ctrlr, uint32_t nsid)
    2193             : {
    2194             :         struct spdk_nvmf_ns *ns;
    2195             : 
    2196             :         /* We do not have NVM subsystem specific ANA state. Hence if NSID is either
    2197             :          * SPDK_NVMF_GLOBAL_NS_TAG, invalid, or for inactive namespace, return
    2198             :          * the optimized state.
    2199             :          */
    2200           8 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, nsid);
    2201           8 :         if (ns == NULL) {
    2202           0 :                 return SPDK_NVME_ANA_OPTIMIZED_STATE;
    2203             :         }
    2204             : 
    2205           8 :         return nvmf_ctrlr_get_ana_state(ctrlr, ns->anagrpid);
    2206             : }
    2207             : 
    2208             : static void
    2209           1 : nvmf_get_error_log_page(struct spdk_nvmf_ctrlr *ctrlr, struct iovec *iovs, int iovcnt,
    2210             :                         uint64_t offset, uint32_t length, uint32_t rae)
    2211             : {
    2212           1 :         if (!rae) {
    2213           1 :                 nvmf_ctrlr_unmask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_ERROR_MASK_BIT);
    2214             :         }
    2215             : 
    2216             :         /* TODO: actually fill out log page data */
    2217           1 : }
    2218             : 
    2219             : static void
    2220          17 : nvmf_get_ana_log_page(struct spdk_nvmf_ctrlr *ctrlr, struct iovec *iovs, int iovcnt,
    2221             :                       uint64_t offset, uint32_t length, uint32_t rae)
    2222             : {
    2223          17 :         struct spdk_nvme_ana_page ana_hdr;
    2224          17 :         struct spdk_nvme_ana_group_descriptor ana_desc;
    2225             :         size_t copy_len, copied_len;
    2226          17 :         uint32_t num_anagrp = 0, anagrpid;
    2227             :         struct spdk_nvmf_ns *ns;
    2228          17 :         struct spdk_iov_xfer ix;
    2229             : 
    2230          17 :         spdk_iov_xfer_init(&ix, iovs, iovcnt);
    2231             : 
    2232          17 :         if (length == 0) {
    2233           0 :                 goto done;
    2234             :         }
    2235             : 
    2236          17 :         if (offset >= sizeof(ana_hdr)) {
    2237          13 :                 offset -= sizeof(ana_hdr);
    2238             :         } else {
    2239          20 :                 for (anagrpid = 1; anagrpid <= ctrlr->subsys->max_nsid; anagrpid++) {
    2240          16 :                         if (ctrlr->subsys->ana_group[anagrpid - 1] > 0) {
    2241          10 :                                 num_anagrp++;
    2242             :                         }
    2243             :                 }
    2244             : 
    2245           4 :                 memset(&ana_hdr, 0, sizeof(ana_hdr));
    2246             : 
    2247           4 :                 ana_hdr.num_ana_group_desc = num_anagrp;
    2248             :                 /* TODO: Support Change Count. */
    2249           4 :                 ana_hdr.change_count = 0;
    2250             : 
    2251           4 :                 copy_len = spdk_min(sizeof(ana_hdr) - offset, length);
    2252           4 :                 copied_len = spdk_iov_xfer_from_buf(&ix, (const char *)&ana_hdr + offset, copy_len);
    2253           4 :                 assert(copied_len == copy_len);
    2254           4 :                 length -= copied_len;
    2255           4 :                 offset = 0;
    2256             :         }
    2257             : 
    2258          17 :         if (length == 0) {
    2259           2 :                 goto done;
    2260             :         }
    2261             : 
    2262          37 :         for (anagrpid = 1; anagrpid <= ctrlr->subsys->max_nsid; anagrpid++) {
    2263          37 :                 if (ctrlr->subsys->ana_group[anagrpid - 1] == 0) {
    2264           7 :                         continue;
    2265             :                 }
    2266             : 
    2267          30 :                 if (offset >= sizeof(ana_desc)) {
    2268          12 :                         offset -= sizeof(ana_desc);
    2269             :                 } else {
    2270          18 :                         memset(&ana_desc, 0, sizeof(ana_desc));
    2271             : 
    2272          18 :                         ana_desc.ana_group_id = anagrpid;
    2273          18 :                         ana_desc.num_of_nsid = ctrlr->subsys->ana_group[anagrpid - 1];
    2274          18 :                         ana_desc.ana_state = nvmf_ctrlr_get_ana_state(ctrlr, anagrpid);
    2275             : 
    2276          18 :                         copy_len = spdk_min(sizeof(ana_desc) - offset, length);
    2277          18 :                         copied_len = spdk_iov_xfer_from_buf(&ix, (const char *)&ana_desc + offset,
    2278             :                                                             copy_len);
    2279          18 :                         assert(copied_len == copy_len);
    2280          18 :                         length -= copied_len;
    2281          18 :                         offset = 0;
    2282             : 
    2283          18 :                         if (length == 0) {
    2284          10 :                                 goto done;
    2285             :                         }
    2286             :                 }
    2287             : 
    2288             :                 /* TODO: Revisit here about O(n^2) cost if we have subsystem with
    2289             :                  * many namespaces in the future.
    2290             :                  */
    2291          86 :                 for (ns = spdk_nvmf_subsystem_get_first_ns(ctrlr->subsys); ns != NULL;
    2292          66 :                      ns = spdk_nvmf_subsystem_get_next_ns(ctrlr->subsys, ns)) {
    2293          71 :                         if (ns->anagrpid != anagrpid) {
    2294          39 :                                 continue;
    2295             :                         }
    2296             : 
    2297          32 :                         if (offset >= sizeof(uint32_t)) {
    2298          16 :                                 offset -= sizeof(uint32_t);
    2299          16 :                                 continue;
    2300             :                         }
    2301             : 
    2302          16 :                         copy_len = spdk_min(sizeof(uint32_t) - offset, length);
    2303          16 :                         copied_len = spdk_iov_xfer_from_buf(&ix, (const char *)&ns->nsid + offset,
    2304             :                                                             copy_len);
    2305          16 :                         assert(copied_len == copy_len);
    2306          16 :                         length -= copied_len;
    2307          16 :                         offset = 0;
    2308             : 
    2309          16 :                         if (length == 0) {
    2310           5 :                                 goto done;
    2311             :                         }
    2312             :                 }
    2313             :         }
    2314             : 
    2315          17 : done:
    2316          17 :         if (!rae) {
    2317          17 :                 nvmf_ctrlr_unmask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_ANA_CHANGE_MASK_BIT);
    2318             :         }
    2319          17 : }
    2320             : 
    2321             : void
    2322           0 : nvmf_ctrlr_ns_changed(struct spdk_nvmf_ctrlr *ctrlr, uint32_t nsid)
    2323             : {
    2324           0 :         uint16_t max_changes = SPDK_COUNTOF(ctrlr->changed_ns_list.ns_list);
    2325             :         uint16_t i;
    2326           0 :         bool found = false;
    2327             : 
    2328           0 :         for (i = 0; i < ctrlr->changed_ns_list_count; i++) {
    2329           0 :                 if (ctrlr->changed_ns_list.ns_list[i] == nsid) {
    2330             :                         /* nsid is already in the list */
    2331           0 :                         found = true;
    2332           0 :                         break;
    2333             :                 }
    2334             :         }
    2335             : 
    2336           0 :         if (!found) {
    2337           0 :                 if (ctrlr->changed_ns_list_count == max_changes) {
    2338             :                         /* Out of space - set first entry to FFFFFFFFh and zero-fill the rest. */
    2339           0 :                         ctrlr->changed_ns_list.ns_list[0] = 0xFFFFFFFFu;
    2340           0 :                         for (i = 1; i < max_changes; i++) {
    2341           0 :                                 ctrlr->changed_ns_list.ns_list[i] = 0;
    2342             :                         }
    2343             :                 } else {
    2344           0 :                         ctrlr->changed_ns_list.ns_list[ctrlr->changed_ns_list_count++] = nsid;
    2345             :                 }
    2346             :         }
    2347           0 : }
    2348             : 
    2349             : static void
    2350           2 : nvmf_get_changed_ns_list_log_page(struct spdk_nvmf_ctrlr *ctrlr,
    2351             :                                   struct iovec *iovs, int iovcnt, uint64_t offset, uint32_t length, uint32_t rae)
    2352             : {
    2353             :         size_t copy_length;
    2354           2 :         struct spdk_iov_xfer ix;
    2355             : 
    2356           2 :         spdk_iov_xfer_init(&ix, iovs, iovcnt);
    2357             : 
    2358           2 :         if (offset < sizeof(ctrlr->changed_ns_list)) {
    2359           2 :                 copy_length = spdk_min(length, sizeof(ctrlr->changed_ns_list) - offset);
    2360           2 :                 if (copy_length) {
    2361           2 :                         spdk_iov_xfer_from_buf(&ix, (char *)&ctrlr->changed_ns_list + offset, copy_length);
    2362             :                 }
    2363             :         }
    2364             : 
    2365             :         /* Clear log page each time it is read */
    2366           2 :         ctrlr->changed_ns_list_count = 0;
    2367           2 :         memset(&ctrlr->changed_ns_list, 0, sizeof(ctrlr->changed_ns_list));
    2368             : 
    2369           2 :         if (!rae) {
    2370           1 :                 nvmf_ctrlr_unmask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_NS_ATTR_CHANGE_MASK_BIT);
    2371             :         }
    2372           2 : }
    2373             : 
    2374             : /* The structure can be modified if we provide support for other commands in future */
    2375             : static const struct spdk_nvme_cmds_and_effect_log_page g_cmds_and_effect_log_page = {
    2376             :         .admin_cmds_supported = {
    2377             :                 /* CSUPP, LBCC, NCC, NIC, CCC, CSE */
    2378             :                 /* Get Log Page */
    2379             :                 [SPDK_NVME_OPC_GET_LOG_PAGE]            = {1, 0, 0, 0, 0, 0, 0, 0},
    2380             :                 /* Identify */
    2381             :                 [SPDK_NVME_OPC_IDENTIFY]                = {1, 0, 0, 0, 0, 0, 0, 0},
    2382             :                 /* Abort */
    2383             :                 [SPDK_NVME_OPC_ABORT]                   = {1, 0, 0, 0, 0, 0, 0, 0},
    2384             :                 /* Set Features */
    2385             :                 [SPDK_NVME_OPC_SET_FEATURES]            = {1, 0, 0, 0, 0, 0, 0, 0},
    2386             :                 /* Get Features */
    2387             :                 [SPDK_NVME_OPC_GET_FEATURES]            = {1, 0, 0, 0, 0, 0, 0, 0},
    2388             :                 /* Async Event Request */
    2389             :                 [SPDK_NVME_OPC_ASYNC_EVENT_REQUEST]     = {1, 0, 0, 0, 0, 0, 0, 0},
    2390             :                 /* Keep Alive */
    2391             :                 [SPDK_NVME_OPC_KEEP_ALIVE]              = {1, 0, 0, 0, 0, 0, 0, 0},
    2392             :         },
    2393             :         .io_cmds_supported = {
    2394             :                 /* FLUSH */
    2395             :                 [SPDK_NVME_OPC_FLUSH]                   = {1, 1, 0, 0, 0, 0, 0, 0},
    2396             :                 /* WRITE */
    2397             :                 [SPDK_NVME_OPC_WRITE]                   = {1, 1, 0, 0, 0, 0, 0, 0},
    2398             :                 /* READ */
    2399             :                 [SPDK_NVME_OPC_READ]                    = {1, 0, 0, 0, 0, 0, 0, 0},
    2400             :                 /* WRITE ZEROES */
    2401             :                 [SPDK_NVME_OPC_WRITE_ZEROES]            = {1, 1, 0, 0, 0, 0, 0, 0},
    2402             :                 /* DATASET MANAGEMENT */
    2403             :                 [SPDK_NVME_OPC_DATASET_MANAGEMENT]      = {1, 1, 0, 0, 0, 0, 0, 0},
    2404             :                 /* COMPARE */
    2405             :                 [SPDK_NVME_OPC_COMPARE]                 = {1, 0, 0, 0, 0, 0, 0, 0},
    2406             :                 /* ZONE MANAGEMENT SEND */
    2407             :                 [SPDK_NVME_OPC_ZONE_MGMT_SEND]          = {1, 1, 0, 0, 0, 0, 0, 0},
    2408             :                 /* ZONE MANAGEMENT RECEIVE */
    2409             :                 [SPDK_NVME_OPC_ZONE_MGMT_RECV]          = {1, 0, 0, 0, 0, 0, 0, 0},
    2410             :                 /* COPY */
    2411             :                 [SPDK_NVME_OPC_COPY]                    = {1, 1, 0, 0, 0, 0, 0, 0},
    2412             :         },
    2413             : };
    2414             : 
    2415             : static void
    2416           0 : nvmf_get_cmds_and_effects_log_page(struct spdk_nvmf_ctrlr *ctrlr, struct iovec *iovs, int iovcnt,
    2417             :                                    uint64_t offset, uint32_t length)
    2418             : {
    2419           0 :         uint32_t page_size = sizeof(struct spdk_nvme_cmds_and_effect_log_page);
    2420           0 :         size_t copy_len = 0;
    2421           0 :         struct spdk_nvme_cmds_and_effect_log_page cmds_and_effect_log_page = g_cmds_and_effect_log_page;
    2422           0 :         struct spdk_nvme_cmds_and_effect_entry csupp_and_lbcc_effect_entry = {1, 1, 0, 0, 0, 0, 0, 0};
    2423           0 :         struct spdk_iov_xfer ix;
    2424             : 
    2425           0 :         spdk_iov_xfer_init(&ix, iovs, iovcnt);
    2426             : 
    2427           0 :         if (offset < page_size) {
    2428           0 :                 if (ctrlr->subsys->zone_append_supported) {
    2429           0 :                         cmds_and_effect_log_page.io_cmds_supported[SPDK_NVME_OPC_ZONE_APPEND] =
    2430             :                                 csupp_and_lbcc_effect_entry;
    2431             :                 }
    2432           0 :                 copy_len = spdk_min(page_size - offset, length);
    2433           0 :                 spdk_iov_xfer_from_buf(&ix, (char *)(&cmds_and_effect_log_page) + offset, copy_len);
    2434             :         }
    2435           0 : }
    2436             : 
    2437             : static void
    2438           1 : nvmf_get_reservation_notification_log_page(struct spdk_nvmf_ctrlr *ctrlr,
    2439             :                 struct iovec *iovs, int iovcnt, uint64_t offset, uint32_t length, uint32_t rae)
    2440             : {
    2441             :         uint32_t unit_log_len, avail_log_len, next_pos, copy_len;
    2442             :         struct spdk_nvmf_reservation_log *log, *log_tmp;
    2443           1 :         struct spdk_iov_xfer ix;
    2444             : 
    2445           1 :         spdk_iov_xfer_init(&ix, iovs, iovcnt);
    2446             : 
    2447           1 :         unit_log_len = sizeof(struct spdk_nvme_reservation_notification_log);
    2448             :         /* No available log, return zeroed log pages */
    2449           1 :         if (!ctrlr->num_avail_log_pages) {
    2450           0 :                 return;
    2451             :         }
    2452             : 
    2453           1 :         avail_log_len = ctrlr->num_avail_log_pages * unit_log_len;
    2454           1 :         if (offset >= avail_log_len) {
    2455           0 :                 return;
    2456             :         }
    2457             : 
    2458           1 :         next_pos = 0;
    2459           3 :         TAILQ_FOREACH_SAFE(log, &ctrlr->log_head, link, log_tmp) {
    2460           3 :                 TAILQ_REMOVE(&ctrlr->log_head, log, link);
    2461           3 :                 ctrlr->num_avail_log_pages--;
    2462             : 
    2463           3 :                 next_pos += unit_log_len;
    2464           3 :                 if (next_pos > offset) {
    2465           3 :                         copy_len = spdk_min(next_pos - offset, length);
    2466           3 :                         spdk_iov_xfer_from_buf(&ix, &log->log, copy_len);
    2467           3 :                         length -= copy_len;
    2468           3 :                         offset += copy_len;
    2469             :                 }
    2470           3 :                 free(log);
    2471             : 
    2472           3 :                 if (length == 0) {
    2473           1 :                         break;
    2474             :                 }
    2475             :         }
    2476             : 
    2477           1 :         if (!rae) {
    2478           1 :                 nvmf_ctrlr_unmask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_RESERVATION_LOG_AVAIL_MASK_BIT);
    2479             :         }
    2480           1 :         return;
    2481             : }
    2482             : 
    2483             : static int
    2484           6 : nvmf_ctrlr_get_log_page(struct spdk_nvmf_request *req)
    2485             : {
    2486           6 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    2487           6 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    2488           6 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    2489           6 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    2490           6 :         struct spdk_nvme_transport_id cmd_source_trid;
    2491             :         uint64_t offset, len;
    2492             :         uint32_t rae, numdl, numdu;
    2493             :         uint8_t lid;
    2494             : 
    2495           6 :         if (req->iovcnt < 1) {
    2496           1 :                 SPDK_DEBUGLOG(nvmf, "get log command with no buffer\n");
    2497           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    2498           1 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    2499           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2500             :         }
    2501             : 
    2502           5 :         offset = (uint64_t)cmd->cdw12 | ((uint64_t)cmd->cdw13 << 32);
    2503           5 :         if (offset & 3) {
    2504           1 :                 SPDK_ERRLOG("Invalid log page offset 0x%" PRIx64 "\n", offset);
    2505           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    2506           1 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    2507           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2508             :         }
    2509             : 
    2510           4 :         rae = cmd->cdw10_bits.get_log_page.rae;
    2511           4 :         numdl = cmd->cdw10_bits.get_log_page.numdl;
    2512           4 :         numdu = cmd->cdw11_bits.get_log_page.numdu;
    2513           4 :         len = ((numdu << 16) + numdl + (uint64_t)1) * 4;
    2514           4 :         if (len > req->length) {
    2515           0 :                 SPDK_ERRLOG("Get log page: len (%" PRIu64 ") > buf size (%u)\n",
    2516             :                             len, req->length);
    2517           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    2518           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    2519           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2520             :         }
    2521             : 
    2522           4 :         lid = cmd->cdw10_bits.get_log_page.lid;
    2523           4 :         SPDK_DEBUGLOG(nvmf, "Get log page: LID=0x%02X offset=0x%" PRIx64 " len=0x%" PRIx64 " rae=%u\n",
    2524             :                       lid, offset, len, rae);
    2525             : 
    2526           4 :         if (spdk_nvmf_subsystem_is_discovery(subsystem)) {
    2527           0 :                 switch (lid) {
    2528           0 :                 case SPDK_NVME_LOG_DISCOVERY:
    2529           0 :                         if (spdk_nvmf_qpair_get_listen_trid(req->qpair, &cmd_source_trid)) {
    2530           0 :                                 SPDK_ERRLOG("Failed to get LOG_DISCOVERY source trid\n");
    2531           0 :                                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    2532           0 :                                 response->status.sc = SPDK_NVME_SC_INTERNAL_DEVICE_ERROR;
    2533           0 :                                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2534             :                         }
    2535           0 :                         nvmf_get_discovery_log_page(subsystem->tgt, ctrlr->hostnqn, req->iov, req->iovcnt,
    2536             :                                                     offset, len, &cmd_source_trid);
    2537           0 :                         if (!rae) {
    2538           0 :                                 nvmf_ctrlr_unmask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_DISCOVERY_LOG_CHANGE_MASK_BIT);
    2539             :                         }
    2540           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2541           0 :                 default:
    2542           0 :                         goto invalid_log_page;
    2543             :                 }
    2544             :         } else {
    2545           4 :                 if (offset > len) {
    2546           0 :                         SPDK_ERRLOG("Get log page: offset (%" PRIu64 ") > len (%" PRIu64 ")\n",
    2547             :                                     offset, len);
    2548           0 :                         response->status.sct = SPDK_NVME_SCT_GENERIC;
    2549           0 :                         response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    2550           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2551             :                 }
    2552             : 
    2553           4 :                 switch (lid) {
    2554           1 :                 case SPDK_NVME_LOG_ERROR:
    2555           1 :                         nvmf_get_error_log_page(ctrlr, req->iov, req->iovcnt, offset, len, rae);
    2556           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2557           0 :                 case SPDK_NVME_LOG_HEALTH_INFORMATION:
    2558             :                         /* TODO: actually fill out log page data */
    2559           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2560           0 :                 case SPDK_NVME_LOG_FIRMWARE_SLOT:
    2561           0 :                         nvmf_get_firmware_slot_log_page(req->iov, req->iovcnt, offset, len);
    2562           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2563           0 :                 case SPDK_NVME_LOG_ASYMMETRIC_NAMESPACE_ACCESS:
    2564           0 :                         if (subsystem->flags.ana_reporting) {
    2565           0 :                                 nvmf_get_ana_log_page(ctrlr, req->iov, req->iovcnt, offset, len, rae);
    2566           0 :                                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2567             :                         } else {
    2568           0 :                                 goto invalid_log_page;
    2569             :                         }
    2570           0 :                 case SPDK_NVME_LOG_COMMAND_EFFECTS_LOG:
    2571           0 :                         nvmf_get_cmds_and_effects_log_page(ctrlr, req->iov, req->iovcnt, offset, len);
    2572           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2573           2 :                 case SPDK_NVME_LOG_CHANGED_NS_LIST:
    2574           2 :                         nvmf_get_changed_ns_list_log_page(ctrlr, req->iov, req->iovcnt, offset, len, rae);
    2575           2 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2576           0 :                 case SPDK_NVME_LOG_RESERVATION_NOTIFICATION:
    2577           0 :                         nvmf_get_reservation_notification_log_page(ctrlr, req->iov, req->iovcnt, offset, len, rae);
    2578           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2579           1 :                 default:
    2580           1 :                         goto invalid_log_page;
    2581             :                 }
    2582             :         }
    2583             : 
    2584           1 : invalid_log_page:
    2585           1 :         SPDK_INFOLOG(nvmf, "Unsupported Get Log Page 0x%02X\n", lid);
    2586           1 :         response->status.sct = SPDK_NVME_SCT_GENERIC;
    2587           1 :         response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    2588           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2589             : }
    2590             : 
    2591             : static struct spdk_nvmf_ns *
    2592          10 : _nvmf_subsystem_get_ns_safe(struct spdk_nvmf_subsystem *subsystem,
    2593             :                             uint32_t nsid,
    2594             :                             struct spdk_nvme_cpl *rsp)
    2595             : {
    2596             :         struct spdk_nvmf_ns *ns;
    2597          10 :         if (nsid == 0 || nsid > subsystem->max_nsid) {
    2598           5 :                 SPDK_ERRLOG("Identify Namespace for invalid NSID %u\n", nsid);
    2599           5 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2600           5 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    2601           5 :                 return NULL;
    2602             :         }
    2603             : 
    2604           5 :         ns = _nvmf_subsystem_get_ns(subsystem, nsid);
    2605           5 :         if (ns == NULL || ns->bdev == NULL) {
    2606             :                 /*
    2607             :                  * Inactive namespaces should return a zero filled data structure.
    2608             :                  * The data buffer is already zeroed by nvmf_ctrlr_process_admin_cmd(),
    2609             :                  * so we can just return early here.
    2610             :                  */
    2611           1 :                 SPDK_DEBUGLOG(nvmf, "Identify Namespace for inactive NSID %u\n", nsid);
    2612           1 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2613           1 :                 rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2614           1 :                 return NULL;
    2615             :         }
    2616           4 :         return ns;
    2617             : }
    2618             : 
    2619             : int
    2620           6 : spdk_nvmf_ctrlr_identify_ns(struct spdk_nvmf_ctrlr *ctrlr,
    2621             :                             struct spdk_nvme_cmd *cmd,
    2622             :                             struct spdk_nvme_cpl *rsp,
    2623             :                             struct spdk_nvme_ns_data *nsdata)
    2624             : {
    2625           6 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    2626             :         struct spdk_nvmf_ns *ns;
    2627             :         uint32_t max_num_blocks, format_index;
    2628             :         enum spdk_nvme_ana_state ana_state;
    2629             : 
    2630           6 :         ns = _nvmf_subsystem_get_ns_safe(subsystem, cmd->nsid, rsp);
    2631           6 :         if (ns == NULL) {
    2632           4 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2633             :         }
    2634             : 
    2635           2 :         nvmf_bdev_ctrlr_identify_ns(ns, nsdata, ctrlr->dif_insert_or_strip);
    2636             : 
    2637           2 :         assert(ctrlr->admin_qpair);
    2638             : 
    2639           2 :         format_index = spdk_nvme_ns_get_format_index(nsdata);
    2640             : 
    2641             :         /* Due to bug in the Linux kernel NVMe driver we have to set noiob no larger than mdts */
    2642           2 :         max_num_blocks = ctrlr->admin_qpair->transport->opts.max_io_size /
    2643           2 :                          (1U << nsdata->lbaf[format_index].lbads);
    2644           2 :         if (nsdata->noiob > max_num_blocks) {
    2645           0 :                 nsdata->noiob = max_num_blocks;
    2646             :         }
    2647             : 
    2648             :         /* Set NOWS equal to Controller MDTS */
    2649           2 :         if (nsdata->nsfeat.optperf) {
    2650           0 :                 nsdata->nows = max_num_blocks - 1;
    2651             :         }
    2652             : 
    2653           2 :         if (subsystem->flags.ana_reporting) {
    2654           0 :                 assert(ns->anagrpid - 1 < subsystem->max_nsid);
    2655           0 :                 nsdata->anagrpid = ns->anagrpid;
    2656             : 
    2657           0 :                 ana_state = nvmf_ctrlr_get_ana_state(ctrlr, ns->anagrpid);
    2658           0 :                 if (ana_state == SPDK_NVME_ANA_INACCESSIBLE_STATE ||
    2659             :                     ana_state == SPDK_NVME_ANA_PERSISTENT_LOSS_STATE) {
    2660           0 :                         nsdata->nuse = 0;
    2661             :                 }
    2662             :         }
    2663             : 
    2664           2 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2665             : }
    2666             : 
    2667             : static void
    2668           3 : nvmf_ctrlr_populate_oacs(struct spdk_nvmf_ctrlr *ctrlr,
    2669             :                          struct spdk_nvme_ctrlr_data *cdata)
    2670             : {
    2671           3 :         cdata->oacs = ctrlr->cdata.oacs;
    2672             : 
    2673           3 :         cdata->oacs.virtualization_management =
    2674           3 :                 g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_VIRTUALIZATION_MANAGEMENT].hdlr != NULL;
    2675           6 :         cdata->oacs.nvme_mi = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_NVME_MI_SEND].hdlr != NULL
    2676           3 :                               && g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_NVME_MI_RECEIVE].hdlr != NULL;
    2677           6 :         cdata->oacs.directives = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_DIRECTIVE_SEND].hdlr != NULL
    2678           3 :                                  && g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_DIRECTIVE_RECEIVE].hdlr != NULL;
    2679           3 :         cdata->oacs.device_self_test =
    2680           3 :                 g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_DEVICE_SELF_TEST].hdlr != NULL;
    2681           6 :         cdata->oacs.ns_manage = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_NS_MANAGEMENT].hdlr != NULL
    2682           3 :                                 && g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_NS_ATTACHMENT].hdlr != NULL;
    2683           6 :         cdata->oacs.firmware = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_FIRMWARE_IMAGE_DOWNLOAD].hdlr !=
    2684             :                                NULL
    2685           3 :                                && g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_FIRMWARE_COMMIT].hdlr != NULL;
    2686           3 :         cdata->oacs.format =
    2687           3 :                 g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_FORMAT_NVM].hdlr != NULL;
    2688           6 :         cdata->oacs.security = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_SECURITY_SEND].hdlr != NULL
    2689           3 :                                && g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_SECURITY_RECEIVE].hdlr != NULL;
    2690           3 :         cdata->oacs.get_lba_status = g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_GET_LBA_STATUS].hdlr !=
    2691             :                                      NULL;
    2692           3 : }
    2693             : 
    2694             : int
    2695           3 : spdk_nvmf_ctrlr_identify_ctrlr(struct spdk_nvmf_ctrlr *ctrlr, struct spdk_nvme_ctrlr_data *cdata)
    2696             : {
    2697           3 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    2698             :         struct spdk_nvmf_transport *transport;
    2699             : 
    2700             :         /*
    2701             :          * Common fields for discovery and NVM subsystems
    2702             :          */
    2703           3 :         assert(ctrlr->admin_qpair);
    2704           3 :         transport = ctrlr->admin_qpair->transport;
    2705           3 :         spdk_strcpy_pad(cdata->fr, FW_VERSION, sizeof(cdata->fr), ' ');
    2706           3 :         assert((transport->opts.max_io_size % 4096) == 0);
    2707           3 :         cdata->mdts = spdk_u32log2(transport->opts.max_io_size / 4096);
    2708           3 :         cdata->cntlid = ctrlr->cntlid;
    2709           3 :         cdata->ver = ctrlr->vcprop.vs;
    2710           3 :         cdata->aerl = ctrlr->cdata.aerl;
    2711           3 :         cdata->lpa.edlp = 1;
    2712           3 :         cdata->elpe = 127;
    2713           3 :         cdata->maxcmd = transport->opts.max_queue_depth;
    2714           3 :         cdata->sgls = ctrlr->cdata.sgls;
    2715           3 :         cdata->fuses = ctrlr->cdata.fuses;
    2716           3 :         cdata->acwu = 0; /* ACWU is 0-based. */
    2717           3 :         if (subsystem->flags.ana_reporting) {
    2718           0 :                 cdata->mnan = subsystem->max_nsid;
    2719             :         }
    2720           3 :         spdk_strcpy_pad(cdata->subnqn, subsystem->subnqn, sizeof(cdata->subnqn), '\0');
    2721             : 
    2722           3 :         SPDK_DEBUGLOG(nvmf, "ctrlr data: maxcmd 0x%x\n", cdata->maxcmd);
    2723           3 :         SPDK_DEBUGLOG(nvmf, "sgls data: 0x%x\n", from_le32(&cdata->sgls));
    2724             : 
    2725             : 
    2726           3 :         if (spdk_nvmf_subsystem_is_discovery(subsystem)) {
    2727             :                 /*
    2728             :                  * NVM Discovery subsystem fields
    2729             :                  */
    2730           0 :                 cdata->oaes.discovery_log_change_notices = 1;
    2731             :         } else {
    2732           3 :                 cdata->vid = ctrlr->cdata.vid;
    2733           3 :                 cdata->ssvid = ctrlr->cdata.ssvid;
    2734           3 :                 cdata->ieee[0] = ctrlr->cdata.ieee[0];
    2735           3 :                 cdata->ieee[1] = ctrlr->cdata.ieee[1];
    2736           3 :                 cdata->ieee[2] = ctrlr->cdata.ieee[2];
    2737             : 
    2738             :                 /*
    2739             :                  * NVM subsystem fields (reserved for discovery subsystems)
    2740             :                  */
    2741           3 :                 spdk_strcpy_pad(cdata->mn, spdk_nvmf_subsystem_get_mn(subsystem), sizeof(cdata->mn), ' ');
    2742           3 :                 spdk_strcpy_pad(cdata->sn, spdk_nvmf_subsystem_get_sn(subsystem), sizeof(cdata->sn), ' ');
    2743           3 :                 cdata->kas = ctrlr->cdata.kas;
    2744             : 
    2745           3 :                 cdata->rab = 6;
    2746           3 :                 cdata->cmic.multi_port = 1;
    2747           3 :                 cdata->cmic.multi_ctrlr = 1;
    2748           3 :                 cdata->oaes.ns_attribute_notices = 1;
    2749           3 :                 cdata->ctratt.host_id_exhid_supported = 1;
    2750             :                 /* We do not have any actual limitation to the number of abort commands.
    2751             :                  * We follow the recommendation by the NVMe specification.
    2752             :                  */
    2753           3 :                 cdata->acl = NVMF_ABORT_COMMAND_LIMIT;
    2754           3 :                 cdata->frmw.slot1_ro = 1;
    2755           3 :                 cdata->frmw.num_slots = 1;
    2756             : 
    2757           3 :                 cdata->lpa.celp = 1; /* Command Effects log page supported */
    2758             : 
    2759           3 :                 cdata->sqes.min = 6;
    2760           3 :                 cdata->sqes.max = 6;
    2761           3 :                 cdata->cqes.min = 4;
    2762           3 :                 cdata->cqes.max = 4;
    2763           3 :                 cdata->nn = subsystem->max_nsid;
    2764           3 :                 cdata->vwc.present = 1;
    2765           3 :                 cdata->vwc.flush_broadcast = SPDK_NVME_FLUSH_BROADCAST_NOT_SUPPORTED;
    2766             : 
    2767           3 :                 cdata->nvmf_specific = ctrlr->cdata.nvmf_specific;
    2768             : 
    2769           3 :                 cdata->oncs.compare = ctrlr->cdata.oncs.compare;
    2770           3 :                 cdata->oncs.dsm = nvmf_ctrlr_dsm_supported(ctrlr);
    2771           3 :                 cdata->oncs.write_zeroes = nvmf_ctrlr_write_zeroes_supported(ctrlr);
    2772           3 :                 cdata->oncs.reservations = ctrlr->cdata.oncs.reservations;
    2773           3 :                 cdata->oncs.copy = ctrlr->cdata.oncs.copy;
    2774           3 :                 cdata->ocfs.copy_format0 = cdata->oncs.copy;
    2775           3 :                 if (subsystem->flags.ana_reporting) {
    2776             :                         /* Asymmetric Namespace Access Reporting is supported. */
    2777           0 :                         cdata->cmic.ana_reporting = 1;
    2778           0 :                         cdata->oaes.ana_change_notices = 1;
    2779             : 
    2780           0 :                         cdata->anatt = ANA_TRANSITION_TIME_IN_SEC;
    2781             :                         /* ANA Change state is not used, and ANA Persistent Loss state
    2782             :                          * is not supported for now.
    2783             :                          */
    2784           0 :                         cdata->anacap.ana_optimized_state = 1;
    2785           0 :                         cdata->anacap.ana_non_optimized_state = 1;
    2786           0 :                         cdata->anacap.ana_inaccessible_state = 1;
    2787             :                         /* ANAGRPID does not change while namespace is attached to controller */
    2788           0 :                         cdata->anacap.no_change_anagrpid = 1;
    2789           0 :                         cdata->anagrpmax = subsystem->max_nsid;
    2790           0 :                         cdata->nanagrpid = subsystem->max_nsid;
    2791             :                 }
    2792             : 
    2793           3 :                 nvmf_ctrlr_populate_oacs(ctrlr, cdata);
    2794             : 
    2795           3 :                 assert(subsystem->tgt != NULL);
    2796           3 :                 cdata->crdt[0] = subsystem->tgt->crdt[0];
    2797           3 :                 cdata->crdt[1] = subsystem->tgt->crdt[1];
    2798           3 :                 cdata->crdt[2] = subsystem->tgt->crdt[2];
    2799             : 
    2800           3 :                 SPDK_DEBUGLOG(nvmf, "ext ctrlr data: ioccsz 0x%x\n",
    2801             :                               cdata->nvmf_specific.ioccsz);
    2802           3 :                 SPDK_DEBUGLOG(nvmf, "ext ctrlr data: iorcsz 0x%x\n",
    2803             :                               cdata->nvmf_specific.iorcsz);
    2804           3 :                 SPDK_DEBUGLOG(nvmf, "ext ctrlr data: icdoff 0x%x\n",
    2805             :                               cdata->nvmf_specific.icdoff);
    2806           3 :                 SPDK_DEBUGLOG(nvmf, "ext ctrlr data: ctrattr 0x%x\n",
    2807             :                               *(uint8_t *)&cdata->nvmf_specific.ctrattr);
    2808           3 :                 SPDK_DEBUGLOG(nvmf, "ext ctrlr data: msdbd 0x%x\n",
    2809             :                               cdata->nvmf_specific.msdbd);
    2810             :         }
    2811             : 
    2812           3 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2813             : }
    2814             : 
    2815             : static int
    2816           1 : nvmf_ns_identify_iocs_zns(struct spdk_nvmf_ns *ns,
    2817             :                           struct spdk_nvme_cmd *cmd,
    2818             :                           struct spdk_nvme_cpl *rsp,
    2819             :                           struct spdk_nvme_zns_ns_data *nsdata_zns)
    2820             : {
    2821           1 :         nsdata_zns->zoc.variable_zone_capacity = 0;
    2822           1 :         nsdata_zns->zoc.zone_active_excursions = 0;
    2823           1 :         nsdata_zns->ozcs.read_across_zone_boundaries = 1;
    2824             :         /* Underflowing the zero based mar and mor bdev helper results in the correct
    2825             :            value of FFFFFFFFh. */
    2826           1 :         nsdata_zns->mar = spdk_bdev_get_max_active_zones(ns->bdev) - 1;
    2827           1 :         nsdata_zns->mor = spdk_bdev_get_max_open_zones(ns->bdev) - 1;
    2828           1 :         nsdata_zns->rrl = 0;
    2829           1 :         nsdata_zns->frl = 0;
    2830           1 :         nsdata_zns->lbafe[0].zsze = spdk_bdev_get_zone_size(ns->bdev);
    2831             : 
    2832           1 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2833           1 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2834           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2835             : }
    2836             : 
    2837             : int
    2838           4 : spdk_nvmf_ns_identify_iocs_specific(struct spdk_nvmf_ctrlr *ctrlr,
    2839             :                                     struct spdk_nvme_cmd *cmd,
    2840             :                                     struct spdk_nvme_cpl *rsp,
    2841             :                                     void *nsdata,
    2842             :                                     size_t nsdata_size)
    2843             : {
    2844           4 :         uint8_t csi = cmd->cdw11_bits.identify.csi;
    2845           4 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    2846           4 :         struct spdk_nvmf_ns *ns = _nvmf_subsystem_get_ns_safe(subsystem, cmd->nsid, rsp);
    2847             : 
    2848           4 :         memset(nsdata, 0, nsdata_size);
    2849             : 
    2850           4 :         if (ns == NULL) {
    2851           2 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2852           2 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    2853           2 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2854             :         }
    2855             : 
    2856           2 :         switch (csi) {
    2857           1 :         case SPDK_NVME_CSI_ZNS:
    2858           1 :                 return nvmf_ns_identify_iocs_zns(ns, cmd, rsp, nsdata);
    2859           1 :         default:
    2860           1 :                 break;
    2861             :         }
    2862             : 
    2863           1 :         SPDK_DEBUGLOG(nvmf,
    2864             :                       "Returning zero filled struct for the iocs specific ns "
    2865             :                       "identify command and CSI 0x%02x\n",
    2866             :                       csi);
    2867           1 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2868           1 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2869           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2870             : }
    2871             : 
    2872             : static int
    2873           1 : nvmf_ctrlr_identify_iocs_nvm(struct spdk_nvmf_ctrlr *ctrlr,
    2874             :                              struct spdk_nvme_cmd *cmd,
    2875             :                              struct spdk_nvme_cpl *rsp,
    2876             :                              struct spdk_nvme_nvm_ctrlr_data *cdata_nvm)
    2877             : {
    2878             :         /* The unit of max_write_zeroes_size_kib is KiB.
    2879             :          * The unit of wzsl is the minimum memory page size(2 ^ (12 + CAP.MPSMIN) bytes)
    2880             :          * and is reported as a power of two (2^n).
    2881             :          */
    2882           2 :         cdata_nvm->wzsl = spdk_u64log2(ctrlr->subsys->max_write_zeroes_size_kib >>
    2883           1 :                                        (2 + ctrlr->vcprop.cap.bits.mpsmin));
    2884             : 
    2885             :         /* The unit of max_discard_size_kib is KiB.
    2886             :          * The dmrsl indicates the maximum number of logical blocks for
    2887             :          * dataset management command.
    2888             :          */
    2889           1 :         cdata_nvm->dmrsl = ctrlr->subsys->max_discard_size_kib << 1;
    2890           1 :         cdata_nvm->dmrl = 1;
    2891             : 
    2892           1 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2893           1 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2894           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2895             : }
    2896             : 
    2897             : static int
    2898           4 : nvmf_ctrlr_identify_iocs_zns(struct spdk_nvmf_ctrlr *ctrlr,
    2899             :                              struct spdk_nvme_cmd *cmd,
    2900             :                              struct spdk_nvme_cpl *rsp,
    2901             :                              struct spdk_nvme_zns_ctrlr_data *cdata_zns)
    2902             : {
    2903             :         /* The unit of max_zone_append_size_kib is KiB.
    2904             :         The unit of zasl is the minimum memory page size
    2905             :         (2 ^ (12 + CAP.MPSMIN) KiB)
    2906             :         and is reported as a power of two (2^n). */
    2907           8 :         cdata_zns->zasl = spdk_u64log2(ctrlr->subsys->max_zone_append_size_kib >>
    2908           4 :                                        (12 + ctrlr->vcprop.cap.bits.mpsmin));
    2909             : 
    2910           4 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2911           4 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2912           4 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2913             : }
    2914             : 
    2915             : int
    2916           5 : spdk_nvmf_ctrlr_identify_iocs_specific(struct spdk_nvmf_ctrlr *ctrlr,
    2917             :                                        struct spdk_nvme_cmd *cmd,
    2918             :                                        struct spdk_nvme_cpl *rsp,
    2919             :                                        void *cdata,
    2920             :                                        size_t cdata_size)
    2921             : {
    2922           5 :         uint8_t csi = cmd->cdw11_bits.identify.csi;
    2923             : 
    2924           5 :         memset(cdata, 0, cdata_size);
    2925             : 
    2926           5 :         switch (csi) {
    2927           1 :         case SPDK_NVME_CSI_NVM:
    2928           1 :                 return nvmf_ctrlr_identify_iocs_nvm(ctrlr, cmd, rsp, cdata);
    2929           4 :         case SPDK_NVME_CSI_ZNS:
    2930           4 :                 return nvmf_ctrlr_identify_iocs_zns(ctrlr, cmd, rsp, cdata);
    2931           0 :         default:
    2932           0 :                 break;
    2933             :         }
    2934             : 
    2935           0 :         SPDK_DEBUGLOG(nvmf,
    2936             :                       "Returning zero filled struct for the iocs specific ctrlr "
    2937             :                       "identify command and CSI 0x%02x\n",
    2938             :                       csi);
    2939           0 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    2940           0 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    2941           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2942             : }
    2943             : 
    2944             : static int
    2945           0 : nvmf_ctrlr_identify_active_ns_list(struct spdk_nvmf_subsystem *subsystem,
    2946             :                                    struct spdk_nvme_cmd *cmd,
    2947             :                                    struct spdk_nvme_cpl *rsp,
    2948             :                                    struct spdk_nvme_ns_list *ns_list)
    2949             : {
    2950             :         struct spdk_nvmf_ns *ns;
    2951           0 :         uint32_t count = 0;
    2952             : 
    2953           0 :         if (cmd->nsid >= 0xfffffffeUL) {
    2954           0 :                 SPDK_ERRLOG("Identify Active Namespace List with invalid NSID %u\n", cmd->nsid);
    2955           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    2956           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2957             :         }
    2958             : 
    2959           0 :         memset(ns_list, 0, sizeof(*ns_list));
    2960             : 
    2961           0 :         for (ns = spdk_nvmf_subsystem_get_first_ns(subsystem); ns != NULL;
    2962           0 :              ns = spdk_nvmf_subsystem_get_next_ns(subsystem, ns)) {
    2963           0 :                 if (ns->opts.nsid <= cmd->nsid) {
    2964           0 :                         continue;
    2965             :                 }
    2966             : 
    2967           0 :                 ns_list->ns_list[count++] = ns->opts.nsid;
    2968           0 :                 if (count == SPDK_COUNTOF(ns_list->ns_list)) {
    2969           0 :                         break;
    2970             :                 }
    2971             :         }
    2972             : 
    2973           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    2974             : }
    2975             : 
    2976             : static void
    2977           7 : _add_ns_id_desc(void **buf_ptr, size_t *buf_remain,
    2978             :                 enum spdk_nvme_nidt type,
    2979             :                 const void *data, size_t data_size)
    2980             : {
    2981             :         struct spdk_nvme_ns_id_desc *desc;
    2982           7 :         size_t desc_size = sizeof(*desc) + data_size;
    2983             : 
    2984             :         /*
    2985             :          * These should never fail in practice, since all valid NS ID descriptors
    2986             :          * should be defined so that they fit in the available 4096-byte buffer.
    2987             :          */
    2988           7 :         assert(data_size > 0);
    2989           7 :         assert(data_size <= UINT8_MAX);
    2990           7 :         assert(desc_size < *buf_remain);
    2991           7 :         if (data_size == 0 || data_size > UINT8_MAX || desc_size > *buf_remain) {
    2992           0 :                 return;
    2993             :         }
    2994             : 
    2995           7 :         desc = *buf_ptr;
    2996           7 :         desc->nidt = type;
    2997           7 :         desc->nidl = data_size;
    2998           7 :         memcpy(desc->nid, data, data_size);
    2999             : 
    3000           7 :         *buf_ptr += desc_size;
    3001           7 :         *buf_remain -= desc_size;
    3002             : }
    3003             : 
    3004             : static int
    3005           6 : nvmf_ctrlr_identify_ns_id_descriptor_list(
    3006             :         struct spdk_nvmf_subsystem *subsystem,
    3007             :         struct spdk_nvme_cmd *cmd,
    3008             :         struct spdk_nvme_cpl *rsp,
    3009             :         void *id_desc_list, size_t id_desc_list_size)
    3010             : {
    3011             :         struct spdk_nvmf_ns *ns;
    3012           6 :         size_t buf_remain = id_desc_list_size;
    3013           6 :         void *buf_ptr = id_desc_list;
    3014             : 
    3015           6 :         ns = _nvmf_subsystem_get_ns(subsystem, cmd->nsid);
    3016           6 :         if (ns == NULL || ns->bdev == NULL) {
    3017           1 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3018           1 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    3019           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3020             :         }
    3021             : 
    3022             : #define ADD_ID_DESC(type, data, size) \
    3023             :         do { \
    3024             :                 if (!spdk_mem_all_zero(data, size)) { \
    3025             :                         _add_ns_id_desc(&buf_ptr, &buf_remain, type, data, size); \
    3026             :                 } \
    3027             :         } while (0)
    3028             : 
    3029           5 :         ADD_ID_DESC(SPDK_NVME_NIDT_EUI64, ns->opts.eui64, sizeof(ns->opts.eui64));
    3030           5 :         ADD_ID_DESC(SPDK_NVME_NIDT_NGUID, ns->opts.nguid, sizeof(ns->opts.nguid));
    3031           5 :         ADD_ID_DESC(SPDK_NVME_NIDT_UUID, &ns->opts.uuid, sizeof(ns->opts.uuid));
    3032           5 :         ADD_ID_DESC(SPDK_NVME_NIDT_CSI, &ns->csi, sizeof(uint8_t));
    3033             : 
    3034             :         /*
    3035             :          * The list is automatically 0-terminated, both in the temporary buffer
    3036             :          * used by nvmf_ctrlr_identify(), and the eventual iov destination -
    3037             :          * controller to host buffers in admin commands always get zeroed in
    3038             :          * nvmf_ctrlr_process_admin_cmd().
    3039             :          */
    3040             : 
    3041             : #undef ADD_ID_DESC
    3042             : 
    3043           5 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3044             : }
    3045             : 
    3046             : static int
    3047           0 : nvmf_ctrlr_identify_iocs(struct spdk_nvmf_ctrlr *ctrlr,
    3048             :                          struct spdk_nvme_cmd *cmd,
    3049             :                          struct spdk_nvme_cpl *rsp,
    3050             :                          void *cdata, size_t cdata_size)
    3051             : {
    3052             :         struct spdk_nvme_iocs_vector *vector;
    3053             :         struct spdk_nvmf_ns *ns;
    3054             : 
    3055           0 :         if (cdata_size < sizeof(struct spdk_nvme_iocs_vector)) {
    3056           0 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3057           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3058           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3059             :         }
    3060             : 
    3061             :         /* For now we only support this command sent to the current
    3062             :          * controller.
    3063             :          */
    3064           0 :         if (cmd->cdw10_bits.identify.cntid != 0xFFFF &&
    3065           0 :             cmd->cdw10_bits.identify.cntid != ctrlr->cntlid) {
    3066           0 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3067           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3068           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3069             :         }
    3070           0 :         memset(cdata, 0, cdata_size);
    3071             : 
    3072           0 :         vector = cdata;
    3073           0 :         vector->nvm = 1;
    3074           0 :         for (ns = spdk_nvmf_subsystem_get_first_ns(ctrlr->subsys); ns != NULL;
    3075           0 :              ns = spdk_nvmf_subsystem_get_next_ns(ctrlr->subsys, ns)) {
    3076           0 :                 if (ns->bdev == NULL) {
    3077           0 :                         continue;
    3078             :                 }
    3079           0 :                 if (spdk_bdev_is_zoned(ns->bdev)) {
    3080           0 :                         vector->zns = 1;
    3081             :                 }
    3082             :         }
    3083             : 
    3084           0 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3085           0 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    3086           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3087             : }
    3088             : 
    3089             : static int
    3090           6 : nvmf_ctrlr_identify(struct spdk_nvmf_request *req)
    3091             : {
    3092             :         uint8_t cns;
    3093           6 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    3094           6 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    3095           6 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    3096           6 :         struct spdk_nvmf_subsystem *subsystem = ctrlr->subsys;
    3097           6 :         int ret = SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3098           6 :         char tmpbuf[SPDK_NVME_IDENTIFY_BUFLEN] = "";
    3099           6 :         struct spdk_iov_xfer ix;
    3100             : 
    3101           6 :         if (req->iovcnt < 1 || req->length < SPDK_NVME_IDENTIFY_BUFLEN) {
    3102           0 :                 SPDK_DEBUGLOG(nvmf, "identify command with invalid buffer\n");
    3103           0 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3104           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3105           0 :                 return ret;
    3106             :         }
    3107             : 
    3108           6 :         cns = cmd->cdw10_bits.identify.cns;
    3109             : 
    3110           6 :         if (spdk_nvmf_subsystem_is_discovery(subsystem) &&
    3111             :             cns != SPDK_NVME_IDENTIFY_CTRLR) {
    3112             :                 /* Discovery controllers only support Identify Controller */
    3113           0 :                 goto invalid_cns;
    3114             :         }
    3115             : 
    3116             :         /*
    3117             :          * We must use a temporary buffer: it's entirely possible the out buffer
    3118             :          * is split across more than one IOV.
    3119             :          */
    3120           6 :         spdk_iov_xfer_init(&ix, req->iov, req->iovcnt);
    3121             : 
    3122           6 :         SPDK_DEBUGLOG(nvmf, "Received identify command with CNS 0x%02x\n", cns);
    3123             : 
    3124           6 :         switch (cns) {
    3125           0 :         case SPDK_NVME_IDENTIFY_NS:
    3126           0 :                 ret = spdk_nvmf_ctrlr_identify_ns(ctrlr, cmd, rsp, (void *)&tmpbuf);
    3127           0 :                 break;
    3128           0 :         case SPDK_NVME_IDENTIFY_CTRLR:
    3129           0 :                 ret = spdk_nvmf_ctrlr_identify_ctrlr(ctrlr, (void *)&tmpbuf);
    3130           0 :                 break;
    3131           0 :         case SPDK_NVME_IDENTIFY_ACTIVE_NS_LIST:
    3132           0 :                 ret = nvmf_ctrlr_identify_active_ns_list(subsystem, cmd, rsp, (void *)&tmpbuf);
    3133           0 :                 break;
    3134           6 :         case SPDK_NVME_IDENTIFY_NS_ID_DESCRIPTOR_LIST:
    3135           6 :                 ret = nvmf_ctrlr_identify_ns_id_descriptor_list(subsystem, cmd, rsp,
    3136           6 :                                 tmpbuf, req->length);
    3137           6 :                 break;
    3138           0 :         case SPDK_NVME_IDENTIFY_NS_IOCS:
    3139           0 :                 ret = spdk_nvmf_ns_identify_iocs_specific(ctrlr, cmd, rsp, (void *)&tmpbuf, req->length);
    3140           0 :                 break;
    3141           0 :         case SPDK_NVME_IDENTIFY_CTRLR_IOCS:
    3142           0 :                 ret = spdk_nvmf_ctrlr_identify_iocs_specific(ctrlr, cmd, rsp, (void *)&tmpbuf, req->length);
    3143           0 :                 break;
    3144           0 :         case SPDK_NVME_IDENTIFY_IOCS:
    3145           0 :                 ret = nvmf_ctrlr_identify_iocs(ctrlr, cmd, rsp, (void *)&tmpbuf, req->length);
    3146           0 :                 break;
    3147           0 :         default:
    3148           0 :                 goto invalid_cns;
    3149             :         }
    3150             : 
    3151           6 :         if (ret == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
    3152           6 :                 spdk_iov_xfer_from_buf(&ix, tmpbuf, sizeof(tmpbuf));
    3153             :         }
    3154             : 
    3155           6 :         return ret;
    3156             : 
    3157           0 : invalid_cns:
    3158           0 :         SPDK_DEBUGLOG(nvmf, "Identify command with unsupported CNS 0x%02x\n", cns);
    3159           0 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3160           0 :         rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3161           0 :         return ret;
    3162             : }
    3163             : 
    3164             : static bool
    3165           2 : nvmf_qpair_abort_aer(struct spdk_nvmf_qpair *qpair, uint16_t cid)
    3166             : {
    3167           2 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
    3168             :         struct spdk_nvmf_request *req;
    3169             :         int i;
    3170             : 
    3171           2 :         if (!nvmf_qpair_is_admin_queue(qpair)) {
    3172           0 :                 return false;
    3173             :         }
    3174             : 
    3175           2 :         assert(spdk_get_thread() == ctrlr->thread);
    3176             : 
    3177           6 :         for (i = 0; i < ctrlr->nr_aer_reqs; i++) {
    3178           6 :                 if (ctrlr->aer_req[i]->cmd->nvme_cmd.cid == cid) {
    3179           2 :                         SPDK_DEBUGLOG(nvmf, "Aborting AER request\n");
    3180           2 :                         req = ctrlr->aer_req[i];
    3181           2 :                         ctrlr->aer_req[i] = NULL;
    3182           2 :                         ctrlr->nr_aer_reqs--;
    3183             : 
    3184             :                         /* Move the last req to the aborting position for making aer_reqs
    3185             :                          * in continuous
    3186             :                          */
    3187           2 :                         if (i < ctrlr->nr_aer_reqs) {
    3188           1 :                                 ctrlr->aer_req[i] = ctrlr->aer_req[ctrlr->nr_aer_reqs];
    3189           1 :                                 ctrlr->aer_req[ctrlr->nr_aer_reqs] = NULL;
    3190             :                         }
    3191             : 
    3192           2 :                         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3193           2 :                         req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_ABORTED_BY_REQUEST;
    3194           2 :                         _nvmf_request_complete(req);
    3195           2 :                         return true;
    3196             :                 }
    3197             :         }
    3198             : 
    3199           0 :         return false;
    3200             : }
    3201             : 
    3202             : void
    3203           0 : nvmf_qpair_abort_pending_zcopy_reqs(struct spdk_nvmf_qpair *qpair)
    3204             : {
    3205             :         struct spdk_nvmf_request *req, *tmp;
    3206             : 
    3207           0 :         TAILQ_FOREACH_SAFE(req, &qpair->outstanding, link, tmp) {
    3208           0 :                 if (req->zcopy_phase == NVMF_ZCOPY_PHASE_EXECUTE) {
    3209             :                         /* Zero-copy requests are kept on the outstanding queue from the moment
    3210             :                          * zcopy_start is sent until a zcopy_end callback is received.  Therefore,
    3211             :                          * we can't remove them from the outstanding queue here, but need to rely on
    3212             :                          * the transport to do a zcopy_end to release their buffers and, in turn,
    3213             :                          * remove them from the queue.
    3214             :                          */
    3215           0 :                         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3216           0 :                         req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_ABORTED_BY_REQUEST;
    3217           0 :                         nvmf_transport_req_free(req);
    3218             :                 }
    3219             :         }
    3220           0 : }
    3221             : 
    3222             : static void
    3223           0 : nvmf_qpair_abort_request(struct spdk_nvmf_qpair *qpair, struct spdk_nvmf_request *req)
    3224             : {
    3225           0 :         uint16_t cid = req->cmd->nvme_cmd.cdw10_bits.abort.cid;
    3226             : 
    3227           0 :         if (nvmf_qpair_abort_aer(qpair, cid)) {
    3228           0 :                 SPDK_DEBUGLOG(nvmf, "abort ctrlr=%p sqid=%u cid=%u successful\n",
    3229             :                               qpair->ctrlr, qpair->qid, cid);
    3230           0 :                 req->rsp->nvme_cpl.cdw0 &= ~1U; /* Command successfully aborted */
    3231             : 
    3232           0 :                 spdk_nvmf_request_complete(req);
    3233           0 :                 return;
    3234             :         }
    3235             : 
    3236           0 :         nvmf_transport_qpair_abort_request(qpair, req);
    3237             : }
    3238             : 
    3239             : static void
    3240           0 : nvmf_ctrlr_abort_done(struct spdk_io_channel_iter *i, int status)
    3241             : {
    3242           0 :         struct spdk_nvmf_request *req = spdk_io_channel_iter_get_ctx(i);
    3243             : 
    3244           0 :         if (status == 0) {
    3245             :                 /* There was no qpair whose ID matches SQID of the abort command.
    3246             :                  * Hence call _nvmf_request_complete() here.
    3247             :                  */
    3248           0 :                 _nvmf_request_complete(req);
    3249             :         }
    3250           0 : }
    3251             : 
    3252             : static void
    3253           0 : nvmf_ctrlr_abort_on_pg(struct spdk_io_channel_iter *i)
    3254             : {
    3255           0 :         struct spdk_nvmf_request *req = spdk_io_channel_iter_get_ctx(i);
    3256           0 :         struct spdk_io_channel *ch = spdk_io_channel_iter_get_channel(i);
    3257           0 :         struct spdk_nvmf_poll_group *group = spdk_io_channel_get_ctx(ch);
    3258           0 :         uint16_t sqid = req->cmd->nvme_cmd.cdw10_bits.abort.sqid;
    3259             :         struct spdk_nvmf_qpair *qpair;
    3260             : 
    3261           0 :         TAILQ_FOREACH(qpair, &group->qpairs, link) {
    3262           0 :                 if (qpair->ctrlr == req->qpair->ctrlr && qpair->qid == sqid) {
    3263             :                         /* Found the qpair */
    3264             : 
    3265           0 :                         nvmf_qpair_abort_request(qpair, req);
    3266             : 
    3267             :                         /* Return -1 for the status so the iteration across threads stops. */
    3268           0 :                         spdk_for_each_channel_continue(i, -1);
    3269           0 :                         return;
    3270             :                 }
    3271             :         }
    3272             : 
    3273           0 :         spdk_for_each_channel_continue(i, 0);
    3274             : }
    3275             : 
    3276             : static int
    3277           0 : nvmf_ctrlr_abort(struct spdk_nvmf_request *req)
    3278             : {
    3279           0 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    3280             : 
    3281           0 :         rsp->cdw0 = 1U; /* Command not aborted */
    3282           0 :         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    3283           0 :         rsp->status.sc = SPDK_NVME_SC_SUCCESS;
    3284             : 
    3285             :         /* Send a message to each poll group, searching for this ctrlr, sqid, and command. */
    3286           0 :         spdk_for_each_channel(req->qpair->ctrlr->subsys->tgt,
    3287             :                               nvmf_ctrlr_abort_on_pg,
    3288             :                               req,
    3289             :                               nvmf_ctrlr_abort_done
    3290             :                              );
    3291             : 
    3292           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
    3293             : }
    3294             : 
    3295             : int
    3296           0 : nvmf_ctrlr_abort_request(struct spdk_nvmf_request *req)
    3297             : {
    3298           0 :         struct spdk_nvmf_request *req_to_abort = req->req_to_abort;
    3299           0 :         struct spdk_bdev *bdev;
    3300           0 :         struct spdk_bdev_desc *desc;
    3301           0 :         struct spdk_io_channel *ch;
    3302             :         int rc;
    3303             : 
    3304           0 :         assert(req_to_abort != NULL);
    3305             : 
    3306           0 :         if (g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_ABORT].hdlr &&
    3307           0 :             nvmf_qpair_is_admin_queue(req_to_abort->qpair)) {
    3308           0 :                 return g_nvmf_custom_admin_cmd_hdlrs[SPDK_NVME_OPC_ABORT].hdlr(req);
    3309             :         }
    3310             : 
    3311           0 :         rc = spdk_nvmf_request_get_bdev(req_to_abort->cmd->nvme_cmd.nsid, req_to_abort,
    3312             :                                         &bdev, &desc, &ch);
    3313           0 :         if (rc != 0) {
    3314           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3315             :         }
    3316             : 
    3317           0 :         return spdk_nvmf_bdev_ctrlr_abort_cmd(bdev, desc, ch, req, req_to_abort);
    3318             : }
    3319             : 
    3320             : static int
    3321           1 : get_features_generic(struct spdk_nvmf_request *req, uint32_t cdw0)
    3322             : {
    3323           1 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    3324             : 
    3325           1 :         rsp->cdw0 = cdw0;
    3326           1 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3327             : }
    3328             : 
    3329             : /* we have to use the typedef in the function declaration to appease astyle. */
    3330             : typedef enum spdk_nvme_path_status_code spdk_nvme_path_status_code_t;
    3331             : 
    3332             : static spdk_nvme_path_status_code_t
    3333           0 : _nvme_ana_state_to_path_status(enum spdk_nvme_ana_state ana_state)
    3334             : {
    3335           0 :         switch (ana_state) {
    3336           0 :         case SPDK_NVME_ANA_INACCESSIBLE_STATE:
    3337           0 :                 return SPDK_NVME_SC_ASYMMETRIC_ACCESS_INACCESSIBLE;
    3338           0 :         case SPDK_NVME_ANA_PERSISTENT_LOSS_STATE:
    3339           0 :                 return SPDK_NVME_SC_ASYMMETRIC_ACCESS_PERSISTENT_LOSS;
    3340           0 :         case SPDK_NVME_ANA_CHANGE_STATE:
    3341           0 :                 return SPDK_NVME_SC_ASYMMETRIC_ACCESS_TRANSITION;
    3342           0 :         default:
    3343           0 :                 return SPDK_NVME_SC_INTERNAL_PATH_ERROR;
    3344             :         }
    3345             : }
    3346             : 
    3347             : static int
    3348           3 : nvmf_ctrlr_get_features(struct spdk_nvmf_request *req)
    3349             : {
    3350             :         uint8_t feature;
    3351           3 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    3352           3 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    3353           3 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    3354             :         enum spdk_nvme_ana_state ana_state;
    3355             : 
    3356           3 :         feature = cmd->cdw10_bits.get_features.fid;
    3357             : 
    3358           3 :         if (spdk_nvmf_subsystem_is_discovery(ctrlr->subsys)) {
    3359             :                 /*
    3360             :                  * Features supported by Discovery controller
    3361             :                  */
    3362           0 :                 switch (feature) {
    3363           0 :                 case SPDK_NVME_FEAT_KEEP_ALIVE_TIMER:
    3364           0 :                         return get_features_generic(req, ctrlr->feat.keep_alive_timer.raw);
    3365           0 :                 case SPDK_NVME_FEAT_ASYNC_EVENT_CONFIGURATION:
    3366           0 :                         return get_features_generic(req, ctrlr->feat.async_event_configuration.raw);
    3367           0 :                 default:
    3368           0 :                         SPDK_INFOLOG(nvmf, "Get Features command with unsupported feature ID 0x%02x\n", feature);
    3369           0 :                         response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3370           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3371             :                 }
    3372             :         }
    3373             :         /*
    3374             :          * Process Get Features command for non-discovery controller
    3375             :          */
    3376           3 :         ana_state = nvmf_ctrlr_get_ana_state_from_nsid(ctrlr, cmd->nsid);
    3377           3 :         switch (ana_state) {
    3378           0 :         case SPDK_NVME_ANA_INACCESSIBLE_STATE:
    3379             :         case SPDK_NVME_ANA_PERSISTENT_LOSS_STATE:
    3380             :         case SPDK_NVME_ANA_CHANGE_STATE:
    3381           0 :                 switch (feature) {
    3382           0 :                 case SPDK_NVME_FEAT_ERROR_RECOVERY:
    3383             :                 case SPDK_NVME_FEAT_WRITE_ATOMICITY:
    3384             :                 case SPDK_NVME_FEAT_HOST_RESERVE_MASK:
    3385             :                 case SPDK_NVME_FEAT_HOST_RESERVE_PERSIST:
    3386           0 :                         response->status.sct = SPDK_NVME_SCT_PATH;
    3387           0 :                         response->status.sc = _nvme_ana_state_to_path_status(ana_state);
    3388           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3389           0 :                 default:
    3390           0 :                         break;
    3391             :                 }
    3392           0 :                 break;
    3393           3 :         default:
    3394           3 :                 break;
    3395             :         }
    3396             : 
    3397           3 :         switch (feature) {
    3398           0 :         case SPDK_NVME_FEAT_ARBITRATION:
    3399           0 :                 return get_features_generic(req, ctrlr->feat.arbitration.raw);
    3400           0 :         case SPDK_NVME_FEAT_POWER_MANAGEMENT:
    3401           0 :                 return get_features_generic(req, ctrlr->feat.power_management.raw);
    3402           2 :         case SPDK_NVME_FEAT_TEMPERATURE_THRESHOLD:
    3403           2 :                 return nvmf_ctrlr_get_features_temperature_threshold(req);
    3404           1 :         case SPDK_NVME_FEAT_ERROR_RECOVERY:
    3405           1 :                 return get_features_generic(req, ctrlr->feat.error_recovery.raw);
    3406           0 :         case SPDK_NVME_FEAT_VOLATILE_WRITE_CACHE:
    3407           0 :                 return get_features_generic(req, ctrlr->feat.volatile_write_cache.raw);
    3408           0 :         case SPDK_NVME_FEAT_NUMBER_OF_QUEUES:
    3409           0 :                 return get_features_generic(req, ctrlr->feat.number_of_queues.raw);
    3410           0 :         case SPDK_NVME_FEAT_INTERRUPT_COALESCING:
    3411           0 :                 return get_features_generic(req, ctrlr->feat.interrupt_coalescing.raw);
    3412           0 :         case SPDK_NVME_FEAT_INTERRUPT_VECTOR_CONFIGURATION:
    3413           0 :                 return nvmf_ctrlr_get_features_interrupt_vector_configuration(req);
    3414           0 :         case SPDK_NVME_FEAT_WRITE_ATOMICITY:
    3415           0 :                 return get_features_generic(req, ctrlr->feat.write_atomicity.raw);
    3416           0 :         case SPDK_NVME_FEAT_ASYNC_EVENT_CONFIGURATION:
    3417           0 :                 return get_features_generic(req, ctrlr->feat.async_event_configuration.raw);
    3418           0 :         case SPDK_NVME_FEAT_KEEP_ALIVE_TIMER:
    3419           0 :                 return get_features_generic(req, ctrlr->feat.keep_alive_timer.raw);
    3420           0 :         case SPDK_NVME_FEAT_HOST_IDENTIFIER:
    3421           0 :                 return nvmf_ctrlr_get_features_host_identifier(req);
    3422           0 :         case SPDK_NVME_FEAT_HOST_RESERVE_MASK:
    3423           0 :                 return nvmf_ctrlr_get_features_reservation_notification_mask(req);
    3424           0 :         case SPDK_NVME_FEAT_HOST_RESERVE_PERSIST:
    3425           0 :                 return nvmf_ctrlr_get_features_reservation_persistence(req);
    3426           0 :         case SPDK_NVME_FEAT_HOST_BEHAVIOR_SUPPORT:
    3427           0 :                 return nvmf_ctrlr_get_features_host_behavior_support(req);
    3428           0 :         default:
    3429           0 :                 SPDK_INFOLOG(nvmf, "Get Features command with unsupported feature ID 0x%02x\n", feature);
    3430           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3431           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3432             :         }
    3433             : }
    3434             : 
    3435             : static int
    3436           5 : nvmf_ctrlr_set_features(struct spdk_nvmf_request *req)
    3437             : {
    3438             :         uint8_t feature, save;
    3439           5 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    3440           5 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    3441           5 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    3442             :         enum spdk_nvme_ana_state ana_state;
    3443             :         /*
    3444             :          * Features are not saveable by the controller as indicated by
    3445             :          * ONCS field of the Identify Controller data.
    3446             :          * */
    3447           5 :         save = cmd->cdw10_bits.set_features.sv;
    3448           5 :         if (save) {
    3449           0 :                 response->status.sc = SPDK_NVME_SC_FEATURE_ID_NOT_SAVEABLE;
    3450           0 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    3451           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3452             :         }
    3453             : 
    3454           5 :         feature = cmd->cdw10_bits.set_features.fid;
    3455             : 
    3456           5 :         if (spdk_nvmf_subsystem_is_discovery(ctrlr->subsys)) {
    3457             :                 /*
    3458             :                  * Features supported by Discovery controller
    3459             :                  */
    3460           0 :                 switch (feature) {
    3461           0 :                 case SPDK_NVME_FEAT_KEEP_ALIVE_TIMER:
    3462           0 :                         return nvmf_ctrlr_set_features_keep_alive_timer(req);
    3463           0 :                 case SPDK_NVME_FEAT_ASYNC_EVENT_CONFIGURATION:
    3464           0 :                         return nvmf_ctrlr_set_features_async_event_configuration(req);
    3465           0 :                 default:
    3466           0 :                         SPDK_INFOLOG(nvmf, "Set Features command with unsupported feature ID 0x%02x\n", feature);
    3467           0 :                         response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3468           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3469             :                 }
    3470             :         }
    3471             :         /*
    3472             :          * Process Set Features command for non-discovery controller
    3473             :          */
    3474           5 :         ana_state = nvmf_ctrlr_get_ana_state_from_nsid(ctrlr, cmd->nsid);
    3475           5 :         switch (ana_state) {
    3476           0 :         case SPDK_NVME_ANA_INACCESSIBLE_STATE:
    3477             :         case SPDK_NVME_ANA_CHANGE_STATE:
    3478           0 :                 if (cmd->nsid == SPDK_NVME_GLOBAL_NS_TAG) {
    3479           0 :                         response->status.sct = SPDK_NVME_SCT_PATH;
    3480           0 :                         response->status.sc = _nvme_ana_state_to_path_status(ana_state);
    3481           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3482             :                 } else {
    3483           0 :                         switch (feature) {
    3484           0 :                         case SPDK_NVME_FEAT_ERROR_RECOVERY:
    3485             :                         case SPDK_NVME_FEAT_WRITE_ATOMICITY:
    3486             :                         case SPDK_NVME_FEAT_HOST_RESERVE_MASK:
    3487             :                         case SPDK_NVME_FEAT_HOST_RESERVE_PERSIST:
    3488           0 :                                 response->status.sct = SPDK_NVME_SCT_PATH;
    3489           0 :                                 response->status.sc = _nvme_ana_state_to_path_status(ana_state);
    3490           0 :                                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3491           0 :                         default:
    3492           0 :                                 break;
    3493             :                         }
    3494             :                 }
    3495           0 :                 break;
    3496           0 :         case SPDK_NVME_ANA_PERSISTENT_LOSS_STATE:
    3497           0 :                 response->status.sct = SPDK_NVME_SCT_PATH;
    3498           0 :                 response->status.sc = SPDK_NVME_SC_ASYMMETRIC_ACCESS_PERSISTENT_LOSS;
    3499           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3500           5 :         default:
    3501           5 :                 break;
    3502             :         }
    3503             : 
    3504           5 :         switch (feature) {
    3505           0 :         case SPDK_NVME_FEAT_ARBITRATION:
    3506           0 :                 return nvmf_ctrlr_set_features_arbitration(req);
    3507           0 :         case SPDK_NVME_FEAT_POWER_MANAGEMENT:
    3508           0 :                 return nvmf_ctrlr_set_features_power_management(req);
    3509           3 :         case SPDK_NVME_FEAT_TEMPERATURE_THRESHOLD:
    3510           3 :                 return nvmf_ctrlr_set_features_temperature_threshold(req);
    3511           2 :         case SPDK_NVME_FEAT_ERROR_RECOVERY:
    3512           2 :                 return nvmf_ctrlr_set_features_error_recovery(req);
    3513           0 :         case SPDK_NVME_FEAT_VOLATILE_WRITE_CACHE:
    3514           0 :                 return nvmf_ctrlr_set_features_volatile_write_cache(req);
    3515           0 :         case SPDK_NVME_FEAT_NUMBER_OF_QUEUES:
    3516           0 :                 return nvmf_ctrlr_set_features_number_of_queues(req);
    3517           0 :         case SPDK_NVME_FEAT_INTERRUPT_COALESCING:
    3518           0 :                 response->status.sct = SPDK_NVME_SCT_COMMAND_SPECIFIC;
    3519           0 :                 response->status.sc = SPDK_NVME_SC_FEATURE_NOT_CHANGEABLE;
    3520           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3521           0 :         case SPDK_NVME_FEAT_WRITE_ATOMICITY:
    3522           0 :                 return nvmf_ctrlr_set_features_write_atomicity(req);
    3523           0 :         case SPDK_NVME_FEAT_ASYNC_EVENT_CONFIGURATION:
    3524           0 :                 return nvmf_ctrlr_set_features_async_event_configuration(req);
    3525           0 :         case SPDK_NVME_FEAT_KEEP_ALIVE_TIMER:
    3526           0 :                 return nvmf_ctrlr_set_features_keep_alive_timer(req);
    3527           0 :         case SPDK_NVME_FEAT_HOST_IDENTIFIER:
    3528           0 :                 return nvmf_ctrlr_set_features_host_identifier(req);
    3529           0 :         case SPDK_NVME_FEAT_HOST_RESERVE_MASK:
    3530           0 :                 return nvmf_ctrlr_set_features_reservation_notification_mask(req);
    3531           0 :         case SPDK_NVME_FEAT_HOST_RESERVE_PERSIST:
    3532           0 :                 return nvmf_ctrlr_set_features_reservation_persistence(req);
    3533           0 :         case SPDK_NVME_FEAT_HOST_BEHAVIOR_SUPPORT:
    3534           0 :                 return nvmf_ctrlr_set_features_host_behavior_support(req);
    3535           0 :         default:
    3536           0 :                 SPDK_INFOLOG(nvmf, "Set Features command with unsupported feature ID 0x%02x\n", feature);
    3537           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3538           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3539             :         }
    3540             : }
    3541             : 
    3542             : static int
    3543           0 : nvmf_ctrlr_keep_alive(struct spdk_nvmf_request *req)
    3544             : {
    3545           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    3546             : 
    3547           0 :         SPDK_DEBUGLOG(nvmf, "Keep Alive\n");
    3548             :         /*
    3549             :          * To handle keep alive just clear or reset the
    3550             :          * ctrlr based keep alive duration counter.
    3551             :          * When added, a separate timer based process
    3552             :          * will monitor if the time since last recorded
    3553             :          * keep alive has exceeded the max duration and
    3554             :          * take appropriate action.
    3555             :          */
    3556           0 :         ctrlr->last_keep_alive_tick = spdk_get_ticks();
    3557             : 
    3558           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3559             : }
    3560             : 
    3561             : int
    3562          17 : nvmf_ctrlr_process_admin_cmd(struct spdk_nvmf_request *req)
    3563             : {
    3564          17 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    3565          17 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    3566          17 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    3567             :         struct spdk_nvmf_subsystem_poll_group *sgroup;
    3568             :         int rc;
    3569             : 
    3570          17 :         if (ctrlr == NULL) {
    3571           0 :                 SPDK_ERRLOG("Admin command sent before CONNECT\n");
    3572           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    3573           0 :                 response->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    3574           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3575             :         }
    3576             : 
    3577          17 :         if (cmd->opc == SPDK_NVME_OPC_ASYNC_EVENT_REQUEST) {
    3578             :                 /* We do not want to treat AERs as outstanding commands,
    3579             :                  * so decrement mgmt_io_outstanding here to offset
    3580             :                  * the increment that happened prior to this call.
    3581             :                  */
    3582          10 :                 sgroup = &req->qpair->group->sgroups[ctrlr->subsys->id];
    3583          10 :                 assert(sgroup != NULL);
    3584          10 :                 sgroup->mgmt_io_outstanding--;
    3585             :         }
    3586             : 
    3587          17 :         assert(spdk_get_thread() == ctrlr->thread);
    3588             : 
    3589          17 :         if (cmd->fuse != 0) {
    3590             :                 /* Fused admin commands are not supported. */
    3591           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    3592           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    3593           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3594             :         }
    3595             : 
    3596          17 :         if (ctrlr->vcprop.cc.bits.en != 1) {
    3597           0 :                 SPDK_ERRLOG("Admin command sent to disabled controller\n");
    3598           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    3599           0 :                 response->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    3600           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3601             :         }
    3602             : 
    3603          17 :         if (req->iovcnt && spdk_nvme_opc_get_data_transfer(cmd->opc) == SPDK_NVME_DATA_CONTROLLER_TO_HOST) {
    3604           6 :                 spdk_iov_memset(req->iov, req->iovcnt, 0);
    3605             :         }
    3606             : 
    3607          17 :         if (spdk_nvmf_subsystem_is_discovery(ctrlr->subsys)) {
    3608             :                 /* Discovery controllers only support these admin OPS. */
    3609           0 :                 switch (cmd->opc) {
    3610           0 :                 case SPDK_NVME_OPC_IDENTIFY:
    3611             :                 case SPDK_NVME_OPC_GET_LOG_PAGE:
    3612             :                 case SPDK_NVME_OPC_KEEP_ALIVE:
    3613             :                 case SPDK_NVME_OPC_SET_FEATURES:
    3614             :                 case SPDK_NVME_OPC_GET_FEATURES:
    3615             :                 case SPDK_NVME_OPC_ASYNC_EVENT_REQUEST:
    3616           0 :                         break;
    3617           0 :                 default:
    3618           0 :                         goto invalid_opcode;
    3619             :                 }
    3620          17 :         }
    3621             : 
    3622             :         /* Call a custom adm cmd handler if set. Aborts are handled in a different path (see nvmf_passthru_admin_cmd) */
    3623          17 :         if (g_nvmf_custom_admin_cmd_hdlrs[cmd->opc].hdlr && cmd->opc != SPDK_NVME_OPC_ABORT) {
    3624           1 :                 rc = g_nvmf_custom_admin_cmd_hdlrs[cmd->opc].hdlr(req);
    3625           1 :                 if (rc >= SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
    3626             :                         /* The handler took care of this command */
    3627           1 :                         return rc;
    3628             :                 }
    3629             :         }
    3630             : 
    3631          16 :         switch (cmd->opc) {
    3632           0 :         case SPDK_NVME_OPC_GET_LOG_PAGE:
    3633           0 :                 return nvmf_ctrlr_get_log_page(req);
    3634           6 :         case SPDK_NVME_OPC_IDENTIFY:
    3635           6 :                 return nvmf_ctrlr_identify(req);
    3636           0 :         case SPDK_NVME_OPC_ABORT:
    3637           0 :                 return nvmf_ctrlr_abort(req);
    3638           0 :         case SPDK_NVME_OPC_GET_FEATURES:
    3639           0 :                 return nvmf_ctrlr_get_features(req);
    3640           0 :         case SPDK_NVME_OPC_SET_FEATURES:
    3641           0 :                 return nvmf_ctrlr_set_features(req);
    3642          10 :         case SPDK_NVME_OPC_ASYNC_EVENT_REQUEST:
    3643          10 :                 return nvmf_ctrlr_async_event_request(req);
    3644           0 :         case SPDK_NVME_OPC_KEEP_ALIVE:
    3645           0 :                 return nvmf_ctrlr_keep_alive(req);
    3646             : 
    3647           0 :         case SPDK_NVME_OPC_CREATE_IO_SQ:
    3648             :         case SPDK_NVME_OPC_CREATE_IO_CQ:
    3649             :         case SPDK_NVME_OPC_DELETE_IO_SQ:
    3650             :         case SPDK_NVME_OPC_DELETE_IO_CQ:
    3651             :                 /* Create and Delete I/O CQ/SQ not allowed in NVMe-oF */
    3652           0 :                 goto invalid_opcode;
    3653             : 
    3654           0 :         default:
    3655           0 :                 goto invalid_opcode;
    3656             :         }
    3657             : 
    3658           0 : invalid_opcode:
    3659           0 :         SPDK_INFOLOG(nvmf, "Unsupported admin opcode 0x%x\n", cmd->opc);
    3660           0 :         response->status.sct = SPDK_NVME_SCT_GENERIC;
    3661           0 :         response->status.sc = SPDK_NVME_SC_INVALID_OPCODE;
    3662           0 :         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3663             : }
    3664             : 
    3665             : static int
    3666           1 : nvmf_ctrlr_process_fabrics_cmd(struct spdk_nvmf_request *req)
    3667             : {
    3668           1 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    3669             :         struct spdk_nvmf_capsule_cmd *cap_hdr;
    3670             : 
    3671           1 :         cap_hdr = &req->cmd->nvmf_cmd;
    3672             : 
    3673           1 :         if (qpair->ctrlr == NULL) {
    3674             :                 /* No ctrlr established yet; the only valid command is Connect */
    3675           1 :                 if (cap_hdr->fctype == SPDK_NVMF_FABRIC_COMMAND_CONNECT) {
    3676           0 :                         return nvmf_ctrlr_cmd_connect(req);
    3677             :                 } else {
    3678           1 :                         SPDK_DEBUGLOG(nvmf, "Got fctype 0x%x, expected Connect\n",
    3679             :                                       cap_hdr->fctype);
    3680           1 :                         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3681           1 :                         req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    3682           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3683             :                 }
    3684           0 :         } else if (nvmf_qpair_is_admin_queue(qpair)) {
    3685             :                 /*
    3686             :                  * Controller session is established, and this is an admin queue.
    3687             :                  * Disallow Connect and allow other fabrics commands.
    3688             :                  */
    3689           0 :                 switch (cap_hdr->fctype) {
    3690           0 :                 case SPDK_NVMF_FABRIC_COMMAND_PROPERTY_SET:
    3691           0 :                         return nvmf_property_set(req);
    3692           0 :                 case SPDK_NVMF_FABRIC_COMMAND_PROPERTY_GET:
    3693           0 :                         return nvmf_property_get(req);
    3694           0 :                 default:
    3695           0 :                         SPDK_DEBUGLOG(nvmf, "unknown fctype 0x%02x\n",
    3696             :                                       cap_hdr->fctype);
    3697           0 :                         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3698           0 :                         req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_INVALID_OPCODE;
    3699           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3700             :                 }
    3701             :         } else {
    3702             :                 /* Controller session is established, and this is an I/O queue */
    3703             :                 /* For now, no I/O-specific Fabrics commands are implemented (other than Connect) */
    3704           0 :                 SPDK_DEBUGLOG(nvmf, "Unexpected I/O fctype 0x%x\n", cap_hdr->fctype);
    3705           0 :                 req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3706           0 :                 req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_INVALID_OPCODE;
    3707           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    3708             :         }
    3709             : }
    3710             : 
    3711             : static inline void
    3712           5 : nvmf_ctrlr_queue_pending_async_event(struct spdk_nvmf_ctrlr *ctrlr,
    3713             :                                      union spdk_nvme_async_event_completion *event)
    3714             : {
    3715             :         struct spdk_nvmf_async_event_completion *nvmf_event;
    3716             : 
    3717           5 :         nvmf_event = calloc(1, sizeof(*nvmf_event));
    3718           5 :         if (!nvmf_event) {
    3719           0 :                 SPDK_ERRLOG("Alloc nvmf event failed, ignore the event\n");
    3720           0 :                 return;
    3721             :         }
    3722           5 :         nvmf_event->event.raw = event->raw;
    3723           5 :         STAILQ_INSERT_TAIL(&ctrlr->async_events, nvmf_event, link);
    3724             : }
    3725             : 
    3726             : static inline int
    3727           6 : nvmf_ctrlr_async_event_notification(struct spdk_nvmf_ctrlr *ctrlr,
    3728             :                                     union spdk_nvme_async_event_completion *event)
    3729             : {
    3730             :         struct spdk_nvmf_request *req;
    3731             :         struct spdk_nvme_cpl *rsp;
    3732             : 
    3733           6 :         assert(spdk_get_thread() == ctrlr->thread);
    3734             : 
    3735             :         /* If there is no outstanding AER request, queue the event.  Then
    3736             :          * if an AER is later submitted, this event can be sent as a
    3737             :          * response.
    3738             :          */
    3739           6 :         if (ctrlr->nr_aer_reqs == 0) {
    3740           5 :                 nvmf_ctrlr_queue_pending_async_event(ctrlr, event);
    3741           5 :                 return 0;
    3742             :         }
    3743             : 
    3744           1 :         req = ctrlr->aer_req[--ctrlr->nr_aer_reqs];
    3745           1 :         rsp = &req->rsp->nvme_cpl;
    3746             : 
    3747           1 :         rsp->cdw0 = event->raw;
    3748             : 
    3749           1 :         _nvmf_request_complete(req);
    3750           1 :         ctrlr->aer_req[ctrlr->nr_aer_reqs] = NULL;
    3751             : 
    3752           1 :         return 0;
    3753             : }
    3754             : 
    3755             : int
    3756           6 : nvmf_ctrlr_async_event_ns_notice(struct spdk_nvmf_ctrlr *ctrlr)
    3757             : {
    3758           6 :         union spdk_nvme_async_event_completion event = {0};
    3759             : 
    3760             :         /* Users may disable the event notification */
    3761           6 :         if (!ctrlr->feat.async_event_configuration.bits.ns_attr_notice) {
    3762           0 :                 return 0;
    3763             :         }
    3764             : 
    3765           6 :         if (!nvmf_ctrlr_mask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_NS_ATTR_CHANGE_MASK_BIT)) {
    3766           3 :                 return 0;
    3767             :         }
    3768             : 
    3769           3 :         event.bits.async_event_type = SPDK_NVME_ASYNC_EVENT_TYPE_NOTICE;
    3770           3 :         event.bits.async_event_info = SPDK_NVME_ASYNC_EVENT_NS_ATTR_CHANGED;
    3771           3 :         event.bits.log_page_identifier = SPDK_NVME_LOG_CHANGED_NS_LIST;
    3772             : 
    3773           3 :         return nvmf_ctrlr_async_event_notification(ctrlr, &event);
    3774             : }
    3775             : 
    3776             : int
    3777           1 : nvmf_ctrlr_async_event_ana_change_notice(struct spdk_nvmf_ctrlr *ctrlr)
    3778             : {
    3779           1 :         union spdk_nvme_async_event_completion event = {0};
    3780             : 
    3781             :         /* Users may disable the event notification */
    3782           1 :         if (!ctrlr->feat.async_event_configuration.bits.ana_change_notice) {
    3783           0 :                 return 0;
    3784             :         }
    3785             : 
    3786           1 :         if (!nvmf_ctrlr_mask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_ANA_CHANGE_MASK_BIT)) {
    3787           0 :                 return 0;
    3788             :         }
    3789             : 
    3790           1 :         event.bits.async_event_type = SPDK_NVME_ASYNC_EVENT_TYPE_NOTICE;
    3791           1 :         event.bits.async_event_info = SPDK_NVME_ASYNC_EVENT_ANA_CHANGE;
    3792           1 :         event.bits.log_page_identifier = SPDK_NVME_LOG_ASYMMETRIC_NAMESPACE_ACCESS;
    3793             : 
    3794           1 :         return nvmf_ctrlr_async_event_notification(ctrlr, &event);
    3795             : }
    3796             : 
    3797             : void
    3798           3 : nvmf_ctrlr_async_event_reservation_notification(struct spdk_nvmf_ctrlr *ctrlr)
    3799             : {
    3800           3 :         union spdk_nvme_async_event_completion event = {0};
    3801             : 
    3802           3 :         if (!ctrlr->num_avail_log_pages) {
    3803           0 :                 return;
    3804             :         }
    3805             : 
    3806           3 :         if (!nvmf_ctrlr_mask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_RESERVATION_LOG_AVAIL_MASK_BIT)) {
    3807           2 :                 return;
    3808             :         }
    3809             : 
    3810           1 :         event.bits.async_event_type = SPDK_NVME_ASYNC_EVENT_TYPE_IO;
    3811           1 :         event.bits.async_event_info = SPDK_NVME_ASYNC_EVENT_RESERVATION_LOG_AVAIL;
    3812           1 :         event.bits.log_page_identifier = SPDK_NVME_LOG_RESERVATION_NOTIFICATION;
    3813             : 
    3814           1 :         nvmf_ctrlr_async_event_notification(ctrlr, &event);
    3815             : }
    3816             : 
    3817             : void
    3818           1 : nvmf_ctrlr_async_event_discovery_log_change_notice(void *ctx)
    3819             : {
    3820           1 :         union spdk_nvme_async_event_completion event = {0};
    3821           1 :         struct spdk_nvmf_ctrlr *ctrlr = ctx;
    3822             : 
    3823             :         /* Users may disable the event notification manually or
    3824             :          * it may not be enabled due to keep alive timeout
    3825             :          * not being set in connect command to discovery controller.
    3826             :          */
    3827           1 :         if (!ctrlr->feat.async_event_configuration.bits.discovery_log_change_notice) {
    3828           0 :                 return;
    3829             :         }
    3830             : 
    3831           1 :         if (!nvmf_ctrlr_mask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_DISCOVERY_LOG_CHANGE_MASK_BIT)) {
    3832           0 :                 return;
    3833             :         }
    3834             : 
    3835           1 :         event.bits.async_event_type = SPDK_NVME_ASYNC_EVENT_TYPE_NOTICE;
    3836           1 :         event.bits.async_event_info = SPDK_NVME_ASYNC_EVENT_DISCOVERY_LOG_CHANGE;
    3837           1 :         event.bits.log_page_identifier = SPDK_NVME_LOG_DISCOVERY;
    3838             : 
    3839           1 :         nvmf_ctrlr_async_event_notification(ctrlr, &event);
    3840             : }
    3841             : 
    3842             : int
    3843           0 : spdk_nvmf_ctrlr_async_event_error_event(struct spdk_nvmf_ctrlr *ctrlr,
    3844             :                                         enum spdk_nvme_async_event_info_error info)
    3845             : {
    3846           0 :         union spdk_nvme_async_event_completion event;
    3847             : 
    3848           0 :         if (!nvmf_ctrlr_mask_aen(ctrlr, SPDK_NVME_ASYNC_EVENT_ERROR_MASK_BIT)) {
    3849           0 :                 return 0;
    3850             :         }
    3851             : 
    3852           0 :         if (info > SPDK_NVME_ASYNC_EVENT_FW_IMAGE_LOAD) {
    3853           0 :                 return 0;
    3854             :         }
    3855             : 
    3856           0 :         event.bits.async_event_type = SPDK_NVME_ASYNC_EVENT_TYPE_ERROR;
    3857           0 :         event.bits.log_page_identifier = SPDK_NVME_LOG_ERROR;
    3858           0 :         event.bits.async_event_info = info;
    3859             : 
    3860           0 :         return nvmf_ctrlr_async_event_notification(ctrlr, &event);
    3861             : }
    3862             : 
    3863             : void
    3864           0 : nvmf_qpair_free_aer(struct spdk_nvmf_qpair *qpair)
    3865             : {
    3866           0 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
    3867             :         int i;
    3868             : 
    3869           0 :         if (!nvmf_qpair_is_admin_queue(qpair)) {
    3870           0 :                 return;
    3871             :         }
    3872             : 
    3873           0 :         assert(spdk_get_thread() == ctrlr->thread);
    3874             : 
    3875           0 :         for (i = 0; i < ctrlr->nr_aer_reqs; i++) {
    3876           0 :                 spdk_nvmf_request_free(ctrlr->aer_req[i]);
    3877           0 :                 ctrlr->aer_req[i] = NULL;
    3878             :         }
    3879             : 
    3880           0 :         ctrlr->nr_aer_reqs = 0;
    3881             : }
    3882             : 
    3883             : void
    3884           0 : spdk_nvmf_ctrlr_abort_aer(struct spdk_nvmf_ctrlr *ctrlr)
    3885             : {
    3886             :         struct spdk_nvmf_request *req;
    3887             :         int i;
    3888             : 
    3889           0 :         assert(spdk_get_thread() == ctrlr->thread);
    3890             : 
    3891           0 :         if (!ctrlr->nr_aer_reqs) {
    3892           0 :                 return;
    3893             :         }
    3894             : 
    3895           0 :         for (i = 0; i < ctrlr->nr_aer_reqs; i++) {
    3896           0 :                 req = ctrlr->aer_req[i];
    3897             : 
    3898           0 :                 req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    3899           0 :                 req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_ABORTED_BY_REQUEST;
    3900           0 :                 _nvmf_request_complete(req);
    3901             : 
    3902           0 :                 ctrlr->aer_req[i] = NULL;
    3903             :         }
    3904             : 
    3905           0 :         ctrlr->nr_aer_reqs = 0;
    3906             : }
    3907             : 
    3908             : static void
    3909           3 : _nvmf_ctrlr_add_reservation_log(void *ctx)
    3910             : {
    3911           3 :         struct spdk_nvmf_reservation_log *log = (struct spdk_nvmf_reservation_log *)ctx;
    3912           3 :         struct spdk_nvmf_ctrlr *ctrlr = log->ctrlr;
    3913             : 
    3914           3 :         ctrlr->log_page_count++;
    3915             : 
    3916             :         /* Maximum number of queued log pages is 255 */
    3917           3 :         if (ctrlr->num_avail_log_pages == 0xff) {
    3918             :                 struct spdk_nvmf_reservation_log *entry;
    3919           0 :                 entry = TAILQ_LAST(&ctrlr->log_head, log_page_head);
    3920           0 :                 entry->log.log_page_count = ctrlr->log_page_count;
    3921           0 :                 free(log);
    3922           0 :                 return;
    3923             :         }
    3924             : 
    3925           3 :         log->log.log_page_count = ctrlr->log_page_count;
    3926           3 :         log->log.num_avail_log_pages = ctrlr->num_avail_log_pages++;
    3927           3 :         TAILQ_INSERT_TAIL(&ctrlr->log_head, log, link);
    3928             : 
    3929           3 :         nvmf_ctrlr_async_event_reservation_notification(ctrlr);
    3930             : }
    3931             : 
    3932             : void
    3933           6 : nvmf_ctrlr_reservation_notice_log(struct spdk_nvmf_ctrlr *ctrlr,
    3934             :                                   struct spdk_nvmf_ns *ns,
    3935             :                                   enum spdk_nvme_reservation_notification_log_page_type type)
    3936             : {
    3937             :         struct spdk_nvmf_reservation_log *log;
    3938             : 
    3939           6 :         switch (type) {
    3940           0 :         case SPDK_NVME_RESERVATION_LOG_PAGE_EMPTY:
    3941           0 :                 return;
    3942           2 :         case SPDK_NVME_REGISTRATION_PREEMPTED:
    3943           2 :                 if (ns->mask & SPDK_NVME_REGISTRATION_PREEMPTED_MASK) {
    3944           1 :                         return;
    3945             :                 }
    3946           1 :                 break;
    3947           2 :         case SPDK_NVME_RESERVATION_RELEASED:
    3948           2 :                 if (ns->mask & SPDK_NVME_RESERVATION_RELEASED_MASK) {
    3949           1 :                         return;
    3950             :                 }
    3951           1 :                 break;
    3952           2 :         case SPDK_NVME_RESERVATION_PREEMPTED:
    3953           2 :                 if (ns->mask & SPDK_NVME_RESERVATION_PREEMPTED_MASK) {
    3954           1 :                         return;
    3955             :                 }
    3956           1 :                 break;
    3957           0 :         default:
    3958           0 :                 return;
    3959             :         }
    3960             : 
    3961           3 :         log = calloc(1, sizeof(*log));
    3962           3 :         if (!log) {
    3963           0 :                 SPDK_ERRLOG("Alloc log page failed, ignore the log\n");
    3964           0 :                 return;
    3965             :         }
    3966           3 :         log->ctrlr = ctrlr;
    3967           3 :         log->log.type = type;
    3968           3 :         log->log.nsid = ns->nsid;
    3969             : 
    3970           3 :         spdk_thread_send_msg(ctrlr->thread, _nvmf_ctrlr_add_reservation_log, log);
    3971             : }
    3972             : 
    3973             : /* Check from subsystem poll group's namespace information data structure */
    3974             : static bool
    3975          28 : nvmf_ns_info_ctrlr_is_registrant(struct spdk_nvmf_subsystem_pg_ns_info *ns_info,
    3976             :                                  struct spdk_nvmf_ctrlr *ctrlr)
    3977             : {
    3978             :         uint32_t i;
    3979             : 
    3980         205 :         for (i = 0; i < SPDK_NVMF_MAX_NUM_REGISTRANTS; i++) {
    3981         195 :                 if (!spdk_uuid_compare(&ns_info->reg_hostid[i], &ctrlr->hostid)) {
    3982          18 :                         return true;
    3983             :                 }
    3984             :         }
    3985             : 
    3986          10 :         return false;
    3987             : }
    3988             : 
    3989             : /*
    3990             :  * Check the NVMe command is permitted or not for current controller(Host).
    3991             :  */
    3992             : static int
    3993          38 : nvmf_ns_reservation_request_check(struct spdk_nvmf_subsystem_pg_ns_info *ns_info,
    3994             :                                   struct spdk_nvmf_ctrlr *ctrlr,
    3995             :                                   struct spdk_nvmf_request *req)
    3996             : {
    3997          38 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    3998          38 :         enum spdk_nvme_reservation_type rtype = ns_info->rtype;
    3999          38 :         uint8_t status = SPDK_NVME_SC_SUCCESS;
    4000             :         uint8_t racqa;
    4001             :         bool is_registrant;
    4002             : 
    4003             :         /* No valid reservation */
    4004          38 :         if (!rtype) {
    4005          10 :                 return 0;
    4006             :         }
    4007             : 
    4008          28 :         is_registrant = nvmf_ns_info_ctrlr_is_registrant(ns_info, ctrlr);
    4009             :         /* All registrants type and current ctrlr is a valid registrant */
    4010          28 :         if ((rtype == SPDK_NVME_RESERVE_WRITE_EXCLUSIVE_ALL_REGS ||
    4011           9 :              rtype == SPDK_NVME_RESERVE_EXCLUSIVE_ACCESS_ALL_REGS) && is_registrant) {
    4012           5 :                 return 0;
    4013          23 :         } else if (!spdk_uuid_compare(&ns_info->holder_id, &ctrlr->hostid)) {
    4014           4 :                 return 0;
    4015             :         }
    4016             : 
    4017             :         /* Non-holder for current controller */
    4018          19 :         switch (cmd->opc) {
    4019           9 :         case SPDK_NVME_OPC_READ:
    4020             :         case SPDK_NVME_OPC_COMPARE:
    4021           9 :                 if (rtype == SPDK_NVME_RESERVE_EXCLUSIVE_ACCESS) {
    4022           1 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4023           1 :                         goto exit;
    4024             :                 }
    4025           8 :                 if ((rtype == SPDK_NVME_RESERVE_EXCLUSIVE_ACCESS_REG_ONLY ||
    4026           2 :                      rtype == SPDK_NVME_RESERVE_EXCLUSIVE_ACCESS_ALL_REGS) && !is_registrant) {
    4027           2 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4028             :                 }
    4029           8 :                 break;
    4030           9 :         case SPDK_NVME_OPC_FLUSH:
    4031             :         case SPDK_NVME_OPC_WRITE:
    4032             :         case SPDK_NVME_OPC_WRITE_UNCORRECTABLE:
    4033             :         case SPDK_NVME_OPC_WRITE_ZEROES:
    4034             :         case SPDK_NVME_OPC_DATASET_MANAGEMENT:
    4035           9 :                 if (rtype == SPDK_NVME_RESERVE_WRITE_EXCLUSIVE ||
    4036             :                     rtype == SPDK_NVME_RESERVE_EXCLUSIVE_ACCESS) {
    4037           3 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4038           3 :                         goto exit;
    4039             :                 }
    4040           6 :                 if (!is_registrant) {
    4041           4 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4042             :                 }
    4043           6 :                 break;
    4044           0 :         case SPDK_NVME_OPC_RESERVATION_ACQUIRE:
    4045           0 :                 racqa = cmd->cdw10_bits.resv_acquire.racqa;
    4046           0 :                 if (racqa == SPDK_NVME_RESERVE_ACQUIRE) {
    4047           0 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4048           0 :                         goto exit;
    4049             :                 }
    4050           0 :                 if (!is_registrant) {
    4051           0 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4052             :                 }
    4053           0 :                 break;
    4054           1 :         case SPDK_NVME_OPC_RESERVATION_RELEASE:
    4055           1 :                 if (!is_registrant) {
    4056           0 :                         status = SPDK_NVME_SC_RESERVATION_CONFLICT;
    4057             :                 }
    4058           1 :                 break;
    4059           0 :         default:
    4060           0 :                 break;
    4061             :         }
    4062             : 
    4063          19 : exit:
    4064          19 :         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    4065          19 :         req->rsp->nvme_cpl.status.sc = status;
    4066          19 :         if (status == SPDK_NVME_SC_RESERVATION_CONFLICT) {
    4067          10 :                 return -EPERM;
    4068             :         }
    4069             : 
    4070           9 :         return 0;
    4071             : }
    4072             : 
    4073             : static int
    4074           6 : nvmf_ctrlr_process_io_fused_cmd(struct spdk_nvmf_request *req, struct spdk_bdev *bdev,
    4075             :                                 struct spdk_bdev_desc *desc, struct spdk_io_channel *ch)
    4076             : {
    4077           6 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    4078           6 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    4079           6 :         struct spdk_nvmf_request *first_fused_req = req->qpair->first_fused_req;
    4080             :         int rc;
    4081             : 
    4082           6 :         if (cmd->fuse == SPDK_NVME_CMD_FUSE_FIRST) {
    4083             :                 /* first fused operation (should be compare) */
    4084           3 :                 if (first_fused_req != NULL) {
    4085           0 :                         struct spdk_nvme_cpl *fused_response = &first_fused_req->rsp->nvme_cpl;
    4086             : 
    4087           0 :                         SPDK_ERRLOG("Wrong sequence of fused operations\n");
    4088             : 
    4089             :                         /* abort req->qpair->first_fused_request and continue with new fused command */
    4090           0 :                         fused_response->status.sc = SPDK_NVME_SC_ABORTED_MISSING_FUSED;
    4091           0 :                         fused_response->status.sct = SPDK_NVME_SCT_GENERIC;
    4092           0 :                         _nvmf_request_complete(first_fused_req);
    4093           3 :                 } else if (cmd->opc != SPDK_NVME_OPC_COMPARE) {
    4094           1 :                         SPDK_ERRLOG("Wrong op code of fused operations\n");
    4095           1 :                         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    4096           1 :                         rsp->status.sc = SPDK_NVME_SC_INVALID_OPCODE;
    4097           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4098             :                 }
    4099             : 
    4100           2 :                 req->qpair->first_fused_req = req;
    4101           2 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
    4102           3 :         } else if (cmd->fuse == SPDK_NVME_CMD_FUSE_SECOND) {
    4103             :                 /* second fused operation (should be write) */
    4104           3 :                 if (first_fused_req == NULL) {
    4105           1 :                         SPDK_ERRLOG("Wrong sequence of fused operations\n");
    4106           1 :                         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    4107           1 :                         rsp->status.sc = SPDK_NVME_SC_ABORTED_MISSING_FUSED;
    4108           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4109           2 :                 } else if (cmd->opc != SPDK_NVME_OPC_WRITE) {
    4110           1 :                         struct spdk_nvme_cpl *fused_response = &first_fused_req->rsp->nvme_cpl;
    4111             : 
    4112           1 :                         SPDK_ERRLOG("Wrong op code of fused operations\n");
    4113             : 
    4114             :                         /* abort req->qpair->first_fused_request and fail current command */
    4115           1 :                         fused_response->status.sc = SPDK_NVME_SC_ABORTED_MISSING_FUSED;
    4116           1 :                         fused_response->status.sct = SPDK_NVME_SCT_GENERIC;
    4117           1 :                         _nvmf_request_complete(first_fused_req);
    4118             : 
    4119           1 :                         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    4120           1 :                         rsp->status.sc = SPDK_NVME_SC_INVALID_OPCODE;
    4121           1 :                         req->qpair->first_fused_req = NULL;
    4122           1 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4123             :                 }
    4124             : 
    4125             :                 /* save request of first command to generate response later */
    4126           1 :                 req->first_fused_req = first_fused_req;
    4127           1 :                 req->first_fused = true;
    4128           1 :                 req->qpair->first_fused_req = NULL;
    4129             :         } else {
    4130           0 :                 SPDK_ERRLOG("Invalid fused command fuse field.\n");
    4131           0 :                 rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    4132           0 :                 rsp->status.sc = SPDK_NVME_SC_INVALID_FIELD;
    4133           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4134             :         }
    4135             : 
    4136           1 :         rc = nvmf_bdev_ctrlr_compare_and_write_cmd(bdev, desc, ch, req->first_fused_req, req);
    4137             : 
    4138           1 :         if (rc == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
    4139           1 :                 if (spdk_nvme_cpl_is_error(rsp)) {
    4140           0 :                         struct spdk_nvme_cpl *fused_response = &first_fused_req->rsp->nvme_cpl;
    4141             : 
    4142           0 :                         fused_response->status = rsp->status;
    4143           0 :                         rsp->status.sct = SPDK_NVME_SCT_GENERIC;
    4144           0 :                         rsp->status.sc = SPDK_NVME_SC_ABORTED_FAILED_FUSED;
    4145             :                         /* Complete first of fused commands. Second will be completed by upper layer */
    4146           0 :                         _nvmf_request_complete(first_fused_req);
    4147           0 :                         req->first_fused_req = NULL;
    4148           0 :                         req->first_fused = false;
    4149             :                 }
    4150             :         }
    4151             : 
    4152           1 :         return rc;
    4153             : }
    4154             : 
    4155             : bool
    4156         271 : nvmf_ctrlr_use_zcopy(struct spdk_nvmf_request *req)
    4157             : {
    4158         271 :         struct spdk_nvmf_transport *transport = req->qpair->transport;
    4159             :         struct spdk_nvmf_ns *ns;
    4160             : 
    4161         271 :         assert(req->zcopy_phase == NVMF_ZCOPY_PHASE_NONE);
    4162             : 
    4163         271 :         if (!transport->opts.zcopy) {
    4164           3 :                 return false;
    4165             :         }
    4166             : 
    4167         268 :         if (nvmf_qpair_is_admin_queue(req->qpair)) {
    4168             :                 /* Admin queue */
    4169           1 :                 return false;
    4170             :         }
    4171             : 
    4172         267 :         if ((req->cmd->nvme_cmd.opc != SPDK_NVME_OPC_WRITE) &&
    4173         262 :             (req->cmd->nvme_cmd.opc != SPDK_NVME_OPC_READ)) {
    4174             :                 /* Not a READ or WRITE command */
    4175         254 :                 return false;
    4176             :         }
    4177             : 
    4178          13 :         if (req->cmd->nvme_cmd.fuse != SPDK_NVME_CMD_FUSE_NONE) {
    4179             :                 /* Fused commands dont use zcopy buffers */
    4180           1 :                 return false;
    4181             :         }
    4182             : 
    4183          12 :         ns = _nvmf_subsystem_get_ns(req->qpair->ctrlr->subsys, req->cmd->nvme_cmd.nsid);
    4184          12 :         if (ns == NULL || ns->bdev == NULL || !ns->zcopy) {
    4185           2 :                 return false;
    4186             :         }
    4187             : 
    4188          10 :         req->zcopy_phase = NVMF_ZCOPY_PHASE_INIT;
    4189          10 :         return true;
    4190             : }
    4191             : 
    4192             : void
    4193           9 : spdk_nvmf_request_zcopy_start(struct spdk_nvmf_request *req)
    4194             : {
    4195           9 :         assert(req->zcopy_phase == NVMF_ZCOPY_PHASE_INIT);
    4196             : 
    4197             :         /* Set iovcnt to be the maximum number of iovs that the ZCOPY can use */
    4198           9 :         req->iovcnt = NVMF_REQ_MAX_BUFFERS;
    4199             : 
    4200           9 :         spdk_nvmf_request_exec(req);
    4201           9 : }
    4202             : 
    4203             : void
    4204           2 : spdk_nvmf_request_zcopy_end(struct spdk_nvmf_request *req, bool commit)
    4205             : {
    4206           2 :         assert(req->zcopy_phase == NVMF_ZCOPY_PHASE_EXECUTE);
    4207           2 :         req->zcopy_phase = NVMF_ZCOPY_PHASE_END_PENDING;
    4208             : 
    4209           2 :         nvmf_bdev_ctrlr_zcopy_end(req, commit);
    4210           2 : }
    4211             : 
    4212             : int
    4213          11 : nvmf_ctrlr_process_io_cmd(struct spdk_nvmf_request *req)
    4214             : {
    4215             :         uint32_t nsid;
    4216             :         struct spdk_nvmf_ns *ns;
    4217             :         struct spdk_bdev *bdev;
    4218             :         struct spdk_bdev_desc *desc;
    4219             :         struct spdk_io_channel *ch;
    4220          11 :         struct spdk_nvmf_poll_group *group = req->qpair->group;
    4221          11 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    4222          11 :         struct spdk_nvme_cmd *cmd = &req->cmd->nvme_cmd;
    4223          11 :         struct spdk_nvme_cpl *response = &req->rsp->nvme_cpl;
    4224             :         struct spdk_nvmf_subsystem_pg_ns_info *ns_info;
    4225             :         enum spdk_nvme_ana_state ana_state;
    4226             : 
    4227             :         /* pre-set response details for this command */
    4228          11 :         response->status.sc = SPDK_NVME_SC_SUCCESS;
    4229          11 :         nsid = cmd->nsid;
    4230             : 
    4231          11 :         if (spdk_unlikely(ctrlr == NULL)) {
    4232           1 :                 SPDK_ERRLOG("I/O command sent before CONNECT\n");
    4233           1 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    4234           1 :                 response->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    4235           1 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4236             :         }
    4237             : 
    4238          10 :         if (spdk_unlikely(ctrlr->vcprop.cc.bits.en != 1)) {
    4239           0 :                 SPDK_ERRLOG("I/O command sent to disabled controller\n");
    4240           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    4241           0 :                 response->status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    4242           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4243             :         }
    4244             : 
    4245          10 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, nsid);
    4246          10 :         if (spdk_unlikely(ns == NULL || ns->bdev == NULL)) {
    4247           0 :                 SPDK_DEBUGLOG(nvmf, "Unsuccessful query for nsid %u\n", cmd->nsid);
    4248           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    4249           0 :                 response->status.dnr = 1;
    4250           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4251             :         }
    4252             : 
    4253          10 :         ana_state = nvmf_ctrlr_get_ana_state(ctrlr, ns->anagrpid);
    4254          10 :         if (spdk_unlikely(ana_state != SPDK_NVME_ANA_OPTIMIZED_STATE &&
    4255             :                           ana_state != SPDK_NVME_ANA_NON_OPTIMIZED_STATE)) {
    4256           0 :                 SPDK_DEBUGLOG(nvmf, "Fail I/O command due to ANA state %d\n",
    4257             :                               ana_state);
    4258           0 :                 response->status.sct = SPDK_NVME_SCT_PATH;
    4259           0 :                 response->status.sc = _nvme_ana_state_to_path_status(ana_state);
    4260           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4261             :         }
    4262             : 
    4263          10 :         if (spdk_likely(ctrlr->listener != NULL)) {
    4264             :                 SPDK_DTRACE_PROBE3_TICKS(nvmf_request_io_exec_path, req,
    4265             :                                          ctrlr->listener->trid->traddr,
    4266             :                                          ctrlr->listener->trid->trsvcid);
    4267             :         }
    4268             : 
    4269             :         /* scan-build falsely reporting dereference of null pointer */
    4270          10 :         assert(group != NULL && group->sgroups != NULL);
    4271          10 :         ns_info = &group->sgroups[ctrlr->subsys->id].ns_info[nsid - 1];
    4272          10 :         if (nvmf_ns_reservation_request_check(ns_info, ctrlr, req)) {
    4273           0 :                 SPDK_DEBUGLOG(nvmf, "Reservation Conflict for nsid %u, opcode %u\n",
    4274             :                               cmd->nsid, cmd->opc);
    4275           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4276             :         }
    4277             : 
    4278          10 :         bdev = ns->bdev;
    4279          10 :         desc = ns->desc;
    4280          10 :         ch = ns_info->channel;
    4281             : 
    4282          10 :         if (spdk_unlikely(cmd->fuse & SPDK_NVME_CMD_FUSE_MASK)) {
    4283           6 :                 return nvmf_ctrlr_process_io_fused_cmd(req, bdev, desc, ch);
    4284           4 :         } else if (spdk_unlikely(req->qpair->first_fused_req != NULL)) {
    4285           0 :                 struct spdk_nvme_cpl *fused_response = &req->qpair->first_fused_req->rsp->nvme_cpl;
    4286             : 
    4287           0 :                 SPDK_ERRLOG("Expected second of fused commands - failing first of fused commands\n");
    4288             : 
    4289             :                 /* abort req->qpair->first_fused_request and continue with new command */
    4290           0 :                 fused_response->status.sc = SPDK_NVME_SC_ABORTED_MISSING_FUSED;
    4291           0 :                 fused_response->status.sct = SPDK_NVME_SCT_GENERIC;
    4292           0 :                 _nvmf_request_complete(req->qpair->first_fused_req);
    4293           0 :                 req->qpair->first_fused_req = NULL;
    4294             :         }
    4295             : 
    4296           4 :         if (spdk_nvmf_request_using_zcopy(req)) {
    4297           4 :                 assert(req->zcopy_phase == NVMF_ZCOPY_PHASE_INIT);
    4298           4 :                 return nvmf_bdev_ctrlr_zcopy_start(bdev, desc, ch, req);
    4299             :         } else {
    4300           0 :                 switch (cmd->opc) {
    4301           0 :                 case SPDK_NVME_OPC_READ:
    4302           0 :                         return nvmf_bdev_ctrlr_read_cmd(bdev, desc, ch, req);
    4303           0 :                 case SPDK_NVME_OPC_WRITE:
    4304           0 :                         return nvmf_bdev_ctrlr_write_cmd(bdev, desc, ch, req);
    4305           0 :                 case SPDK_NVME_OPC_COMPARE:
    4306           0 :                         return nvmf_bdev_ctrlr_compare_cmd(bdev, desc, ch, req);
    4307           0 :                 case SPDK_NVME_OPC_WRITE_ZEROES:
    4308           0 :                         return nvmf_bdev_ctrlr_write_zeroes_cmd(bdev, desc, ch, req);
    4309           0 :                 case SPDK_NVME_OPC_FLUSH:
    4310           0 :                         return nvmf_bdev_ctrlr_flush_cmd(bdev, desc, ch, req);
    4311           0 :                 case SPDK_NVME_OPC_DATASET_MANAGEMENT:
    4312           0 :                         return nvmf_bdev_ctrlr_dsm_cmd(bdev, desc, ch, req);
    4313           0 :                 case SPDK_NVME_OPC_RESERVATION_REGISTER:
    4314             :                 case SPDK_NVME_OPC_RESERVATION_ACQUIRE:
    4315             :                 case SPDK_NVME_OPC_RESERVATION_RELEASE:
    4316             :                 case SPDK_NVME_OPC_RESERVATION_REPORT:
    4317           0 :                         spdk_thread_send_msg(ctrlr->subsys->thread, nvmf_ns_reservation_request, req);
    4318           0 :                         return SPDK_NVMF_REQUEST_EXEC_STATUS_ASYNCHRONOUS;
    4319           0 :                 case SPDK_NVME_OPC_COPY:
    4320           0 :                         return nvmf_bdev_ctrlr_copy_cmd(bdev, desc, ch, req);
    4321           0 :                 default:
    4322           0 :                         return nvmf_bdev_ctrlr_nvme_passthru_io(bdev, desc, ch, req);
    4323             :                 }
    4324             :         }
    4325             : }
    4326             : 
    4327             : static void
    4328          33 : nvmf_qpair_request_cleanup(struct spdk_nvmf_qpair *qpair)
    4329             : {
    4330          33 :         if (spdk_unlikely(qpair->state == SPDK_NVMF_QPAIR_DEACTIVATING)) {
    4331           1 :                 assert(qpair->state_cb != NULL);
    4332             : 
    4333           1 :                 if (TAILQ_EMPTY(&qpair->outstanding)) {
    4334           1 :                         qpair->state_cb(qpair->state_cb_arg, 0);
    4335             :                 }
    4336             :         }
    4337          33 : }
    4338             : 
    4339             : int
    4340           0 : spdk_nvmf_request_free(struct spdk_nvmf_request *req)
    4341             : {
    4342           0 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4343             : 
    4344           0 :         TAILQ_REMOVE(&qpair->outstanding, req, link);
    4345           0 :         if (spdk_unlikely(nvmf_transport_req_free(req))) {
    4346           0 :                 SPDK_ERRLOG("Unable to free transport level request resources.\n");
    4347             :         }
    4348             : 
    4349           0 :         nvmf_qpair_request_cleanup(qpair);
    4350             : 
    4351           0 :         return 0;
    4352             : }
    4353             : 
    4354             : static void
    4355          33 : _nvmf_request_complete(void *ctx)
    4356             : {
    4357          33 :         struct spdk_nvmf_request *req = ctx;
    4358          33 :         struct spdk_nvme_cpl *rsp = &req->rsp->nvme_cpl;
    4359             :         struct spdk_nvmf_qpair *qpair;
    4360          33 :         struct spdk_nvmf_subsystem_poll_group *sgroup = NULL;
    4361             :         struct spdk_nvmf_subsystem_pg_ns_info *ns_info;
    4362          33 :         bool is_aer = false;
    4363             :         uint32_t nsid;
    4364             :         bool paused;
    4365             :         uint8_t opcode;
    4366             : 
    4367          33 :         rsp->sqid = 0;
    4368          33 :         rsp->status.p = 0;
    4369          33 :         rsp->cid = req->cmd->nvme_cmd.cid;
    4370          33 :         nsid = req->cmd->nvme_cmd.nsid;
    4371          33 :         opcode = req->cmd->nvmf_cmd.opcode;
    4372             : 
    4373          33 :         qpair = req->qpair;
    4374          33 :         if (spdk_likely(qpair->ctrlr)) {
    4375          23 :                 sgroup = &qpair->group->sgroups[qpair->ctrlr->subsys->id];
    4376          23 :                 assert(sgroup != NULL);
    4377          23 :                 is_aer = req->cmd->nvme_cmd.opc == SPDK_NVME_OPC_ASYNC_EVENT_REQUEST;
    4378          23 :                 if (spdk_likely(qpair->qid != 0)) {
    4379          16 :                         qpair->group->stat.completed_nvme_io++;
    4380             :                 }
    4381             : 
    4382             :                 /*
    4383             :                  * Set the crd value.
    4384             :                  * If the the IO has any error, and dnr (DoNotRetry) is not 1,
    4385             :                  * and ACRE is enabled, we will set the crd to 1 to select the first CRDT.
    4386             :                  */
    4387          23 :                 if (spdk_unlikely(spdk_nvme_cpl_is_error(rsp) &&
    4388             :                                   rsp->status.dnr == 0 &&
    4389             :                                   qpair->ctrlr->acre_enabled)) {
    4390           0 :                         rsp->status.crd = 1;
    4391             :                 }
    4392          10 :         } else if (spdk_unlikely(nvmf_request_is_fabric_connect(req))) {
    4393           8 :                 sgroup = nvmf_subsystem_pg_from_connect_cmd(req);
    4394             :         }
    4395             : 
    4396          33 :         if (SPDK_DEBUGLOG_FLAG_ENABLED("nvmf")) {
    4397           0 :                 spdk_nvme_print_completion(qpair->qid, rsp);
    4398             :         }
    4399             : 
    4400          33 :         switch (req->zcopy_phase) {
    4401          23 :         case NVMF_ZCOPY_PHASE_NONE:
    4402          23 :                 TAILQ_REMOVE(&qpair->outstanding, req, link);
    4403          23 :                 break;
    4404           8 :         case NVMF_ZCOPY_PHASE_INIT:
    4405           8 :                 if (spdk_unlikely(spdk_nvme_cpl_is_error(rsp))) {
    4406           5 :                         req->zcopy_phase = NVMF_ZCOPY_PHASE_INIT_FAILED;
    4407           5 :                         TAILQ_REMOVE(&qpair->outstanding, req, link);
    4408             :                 } else {
    4409           3 :                         req->zcopy_phase = NVMF_ZCOPY_PHASE_EXECUTE;
    4410             :                 }
    4411           8 :                 break;
    4412           0 :         case NVMF_ZCOPY_PHASE_EXECUTE:
    4413           0 :                 break;
    4414           2 :         case NVMF_ZCOPY_PHASE_END_PENDING:
    4415           2 :                 TAILQ_REMOVE(&qpair->outstanding, req, link);
    4416           2 :                 req->zcopy_phase = NVMF_ZCOPY_PHASE_COMPLETE;
    4417           2 :                 break;
    4418           0 :         default:
    4419           0 :                 SPDK_ERRLOG("Invalid ZCOPY phase %u\n", req->zcopy_phase);
    4420           0 :                 break;
    4421             :         }
    4422             : 
    4423          33 :         if (spdk_unlikely(nvmf_transport_req_complete(req))) {
    4424           0 :                 SPDK_ERRLOG("Transport request completion error!\n");
    4425             :         }
    4426             : 
    4427             :         /* AER cmd is an exception */
    4428          33 :         if (spdk_likely(sgroup && !is_aer)) {
    4429          29 :                 if (spdk_unlikely(opcode == SPDK_NVME_OPC_FABRIC ||
    4430             :                                   nvmf_qpair_is_admin_queue(qpair))) {
    4431          15 :                         assert(sgroup->mgmt_io_outstanding > 0);
    4432          15 :                         sgroup->mgmt_io_outstanding--;
    4433             :                 } else {
    4434          14 :                         if (req->zcopy_phase == NVMF_ZCOPY_PHASE_NONE ||
    4435           9 :                             req->zcopy_phase == NVMF_ZCOPY_PHASE_COMPLETE ||
    4436           7 :                             req->zcopy_phase == NVMF_ZCOPY_PHASE_INIT_FAILED) {
    4437             :                                 /* End of request */
    4438             : 
    4439             :                                 /* NOTE: This implicitly also checks for 0, since 0 - 1 wraps around to UINT32_MAX. */
    4440          11 :                                 if (spdk_likely(nsid - 1 < sgroup->num_ns)) {
    4441          10 :                                         sgroup->ns_info[nsid - 1].io_outstanding--;
    4442             :                                 }
    4443             :                         }
    4444             :                 }
    4445             : 
    4446          29 :                 if (spdk_unlikely(sgroup->state == SPDK_NVMF_SUBSYSTEM_PAUSING &&
    4447             :                                   sgroup->mgmt_io_outstanding == 0)) {
    4448           0 :                         paused = true;
    4449           0 :                         for (nsid = 0; nsid < sgroup->num_ns; nsid++) {
    4450           0 :                                 ns_info = &sgroup->ns_info[nsid];
    4451             : 
    4452           0 :                                 if (ns_info->state == SPDK_NVMF_SUBSYSTEM_PAUSING &&
    4453           0 :                                     ns_info->io_outstanding > 0) {
    4454           0 :                                         paused = false;
    4455           0 :                                         break;
    4456             :                                 }
    4457             :                         }
    4458             : 
    4459           0 :                         if (paused) {
    4460           0 :                                 sgroup->state = SPDK_NVMF_SUBSYSTEM_PAUSED;
    4461           0 :                                 sgroup->cb_fn(sgroup->cb_arg, 0);
    4462           0 :                                 sgroup->cb_fn = NULL;
    4463           0 :                                 sgroup->cb_arg = NULL;
    4464             :                         }
    4465             :                 }
    4466             : 
    4467             :         }
    4468             : 
    4469          33 :         nvmf_qpair_request_cleanup(qpair);
    4470          33 : }
    4471             : 
    4472             : int
    4473          17 : spdk_nvmf_request_complete(struct spdk_nvmf_request *req)
    4474             : {
    4475          17 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4476             : 
    4477          17 :         spdk_thread_exec_msg(qpair->group->thread, _nvmf_request_complete, req);
    4478             : 
    4479          17 :         return 0;
    4480             : }
    4481             : 
    4482             : void
    4483           0 : spdk_nvmf_request_exec_fabrics(struct spdk_nvmf_request *req)
    4484             : {
    4485           0 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4486           0 :         struct spdk_nvmf_subsystem_poll_group *sgroup = NULL;
    4487             :         enum spdk_nvmf_request_exec_status status;
    4488             : 
    4489           0 :         if (qpair->ctrlr) {
    4490           0 :                 sgroup = &qpair->group->sgroups[qpair->ctrlr->subsys->id];
    4491           0 :         } else if (spdk_unlikely(nvmf_request_is_fabric_connect(req))) {
    4492           0 :                 sgroup = nvmf_subsystem_pg_from_connect_cmd(req);
    4493             :         }
    4494             : 
    4495           0 :         assert(sgroup != NULL);
    4496           0 :         sgroup->mgmt_io_outstanding++;
    4497             : 
    4498             :         /* Place the request on the outstanding list so we can keep track of it */
    4499           0 :         TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
    4500             : 
    4501           0 :         assert(req->cmd->nvmf_cmd.opcode == SPDK_NVME_OPC_FABRIC);
    4502           0 :         status = nvmf_ctrlr_process_fabrics_cmd(req);
    4503             : 
    4504           0 :         if (status == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
    4505           0 :                 _nvmf_request_complete(req);
    4506             :         }
    4507           0 : }
    4508             : 
    4509             : static bool
    4510          15 : nvmf_check_subsystem_active(struct spdk_nvmf_request *req)
    4511             : {
    4512          15 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4513          15 :         struct spdk_nvmf_subsystem_poll_group *sgroup = NULL;
    4514             :         struct spdk_nvmf_subsystem_pg_ns_info *ns_info;
    4515             :         uint32_t nsid;
    4516             : 
    4517          15 :         if (spdk_likely(qpair->ctrlr)) {
    4518          14 :                 sgroup = &qpair->group->sgroups[qpair->ctrlr->subsys->id];
    4519          14 :                 assert(sgroup != NULL);
    4520           1 :         } else if (spdk_unlikely(nvmf_request_is_fabric_connect(req))) {
    4521           0 :                 sgroup = nvmf_subsystem_pg_from_connect_cmd(req);
    4522             :         }
    4523             : 
    4524             :         /* Check if the subsystem is paused (if there is a subsystem) */
    4525          15 :         if (spdk_likely(sgroup != NULL)) {
    4526          14 :                 if (spdk_unlikely(req->cmd->nvmf_cmd.opcode == SPDK_NVME_OPC_FABRIC ||
    4527             :                                   nvmf_qpair_is_admin_queue(qpair))) {
    4528           0 :                         if (sgroup->state != SPDK_NVMF_SUBSYSTEM_ACTIVE) {
    4529             :                                 /* The subsystem is not currently active. Queue this request. */
    4530           0 :                                 TAILQ_INSERT_TAIL(&sgroup->queued, req, link);
    4531           0 :                                 return false;
    4532             :                         }
    4533           0 :                         sgroup->mgmt_io_outstanding++;
    4534             :                 } else {
    4535          14 :                         nsid = req->cmd->nvme_cmd.nsid;
    4536             : 
    4537             :                         /* NOTE: This implicitly also checks for 0, since 0 - 1 wraps around to UINT32_MAX. */
    4538          14 :                         if (spdk_unlikely(nsid - 1 >= sgroup->num_ns)) {
    4539           1 :                                 req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    4540           1 :                                 req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    4541           1 :                                 req->rsp->nvme_cpl.status.dnr = 1;
    4542           1 :                                 TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
    4543           1 :                                 _nvmf_request_complete(req);
    4544           1 :                                 return false;
    4545             :                         }
    4546             : 
    4547          13 :                         ns_info = &sgroup->ns_info[nsid - 1];
    4548          13 :                         if (spdk_unlikely(ns_info->channel == NULL)) {
    4549             :                                 /* This can can happen if host sends I/O to a namespace that is
    4550             :                                  * in the process of being added, but before the full addition
    4551             :                                  * process is complete.  Report invalid namespace in that case.
    4552             :                                  */
    4553           1 :                                 req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    4554           1 :                                 req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    4555           1 :                                 req->rsp->nvme_cpl.status.dnr = 1;
    4556           1 :                                 TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
    4557           1 :                                 ns_info->io_outstanding++;
    4558           1 :                                 _nvmf_request_complete(req);
    4559           1 :                                 return false;
    4560             :                         }
    4561             : 
    4562          12 :                         if (spdk_unlikely(ns_info->state != SPDK_NVMF_SUBSYSTEM_ACTIVE)) {
    4563             :                                 /* The namespace is not currently active. Queue this request. */
    4564           1 :                                 TAILQ_INSERT_TAIL(&sgroup->queued, req, link);
    4565           1 :                                 return false;
    4566             :                         }
    4567             : 
    4568          11 :                         ns_info->io_outstanding++;
    4569             :                 }
    4570             : 
    4571          11 :                 if (spdk_unlikely(qpair->state != SPDK_NVMF_QPAIR_ACTIVE)) {
    4572           1 :                         req->rsp->nvme_cpl.status.sct = SPDK_NVME_SCT_GENERIC;
    4573           1 :                         req->rsp->nvme_cpl.status.sc = SPDK_NVME_SC_COMMAND_SEQUENCE_ERROR;
    4574           1 :                         TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
    4575           1 :                         _nvmf_request_complete(req);
    4576           1 :                         return false;
    4577             :                 }
    4578             :         }
    4579             : 
    4580          11 :         return true;
    4581             : }
    4582             : 
    4583             : void
    4584          15 : spdk_nvmf_request_exec(struct spdk_nvmf_request *req)
    4585             : {
    4586          15 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4587          15 :         struct spdk_nvmf_transport *transport = qpair->transport;
    4588             :         enum spdk_nvmf_request_exec_status status;
    4589             : 
    4590          15 :         if (spdk_unlikely(!nvmf_check_subsystem_active(req))) {
    4591           4 :                 return;
    4592             :         }
    4593             : 
    4594          11 :         if (SPDK_DEBUGLOG_FLAG_ENABLED("nvmf")) {
    4595           0 :                 spdk_nvme_print_command(qpair->qid, &req->cmd->nvme_cmd);
    4596             :         }
    4597             : 
    4598             :         /* Place the request on the outstanding list so we can keep track of it */
    4599          11 :         TAILQ_INSERT_TAIL(&qpair->outstanding, req, link);
    4600             : 
    4601          11 :         if (spdk_unlikely((req->cmd->nvmf_cmd.opcode == SPDK_NVME_OPC_FABRIC) &&
    4602             :                           spdk_nvme_trtype_is_fabrics(transport->ops->type))) {
    4603           0 :                 status = nvmf_ctrlr_process_fabrics_cmd(req);
    4604          11 :         } else if (spdk_unlikely(nvmf_qpair_is_admin_queue(qpair))) {
    4605           0 :                 status = nvmf_ctrlr_process_admin_cmd(req);
    4606             :         } else {
    4607          11 :                 status = nvmf_ctrlr_process_io_cmd(req);
    4608             :         }
    4609             : 
    4610          11 :         if (status == SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE) {
    4611           9 :                 _nvmf_request_complete(req);
    4612             :         }
    4613             : }
    4614             : 
    4615             : static bool
    4616           4 : nvmf_ctrlr_get_dif_ctx(struct spdk_nvmf_ctrlr *ctrlr, struct spdk_nvme_cmd *cmd,
    4617             :                        struct spdk_dif_ctx *dif_ctx)
    4618             : {
    4619             :         struct spdk_nvmf_ns *ns;
    4620             :         struct spdk_bdev *bdev;
    4621             : 
    4622           4 :         if (ctrlr == NULL || cmd == NULL) {
    4623           0 :                 return false;
    4624             :         }
    4625             : 
    4626           4 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, cmd->nsid);
    4627           4 :         if (ns == NULL || ns->bdev == NULL) {
    4628           2 :                 return false;
    4629             :         }
    4630             : 
    4631           2 :         bdev = ns->bdev;
    4632             : 
    4633           2 :         switch (cmd->opc) {
    4634           1 :         case SPDK_NVME_OPC_READ:
    4635             :         case SPDK_NVME_OPC_WRITE:
    4636             :         case SPDK_NVME_OPC_COMPARE:
    4637           1 :                 return nvmf_bdev_ctrlr_get_dif_ctx(bdev, cmd, dif_ctx);
    4638           1 :         default:
    4639           1 :                 break;
    4640             :         }
    4641             : 
    4642           1 :         return false;
    4643             : }
    4644             : 
    4645             : bool
    4646          12 : spdk_nvmf_request_get_dif_ctx(struct spdk_nvmf_request *req, struct spdk_dif_ctx *dif_ctx)
    4647             : {
    4648          12 :         struct spdk_nvmf_qpair *qpair = req->qpair;
    4649          12 :         struct spdk_nvmf_ctrlr *ctrlr = qpair->ctrlr;
    4650             : 
    4651          12 :         if (spdk_likely(ctrlr == NULL || !ctrlr->dif_insert_or_strip)) {
    4652           5 :                 return false;
    4653             :         }
    4654             : 
    4655           7 :         if (spdk_unlikely(qpair->state != SPDK_NVMF_QPAIR_ACTIVE)) {
    4656           1 :                 return false;
    4657             :         }
    4658             : 
    4659           6 :         if (spdk_unlikely(req->cmd->nvmf_cmd.opcode == SPDK_NVME_OPC_FABRIC)) {
    4660           1 :                 return false;
    4661             :         }
    4662             : 
    4663           5 :         if (spdk_unlikely(nvmf_qpair_is_admin_queue(qpair))) {
    4664           1 :                 return false;
    4665             :         }
    4666             : 
    4667           4 :         return nvmf_ctrlr_get_dif_ctx(ctrlr, &req->cmd->nvme_cmd, dif_ctx);
    4668             : }
    4669             : 
    4670             : void
    4671           1 : spdk_nvmf_set_custom_admin_cmd_hdlr(uint8_t opc, spdk_nvmf_custom_cmd_hdlr hdlr)
    4672             : {
    4673           1 :         g_nvmf_custom_admin_cmd_hdlrs[opc].hdlr = hdlr;
    4674           1 : }
    4675             : 
    4676             : static int
    4677           0 : nvmf_passthru_admin_cmd_for_bdev_nsid(struct spdk_nvmf_request *req, uint32_t bdev_nsid)
    4678             : {
    4679           0 :         struct spdk_bdev *bdev;
    4680           0 :         struct spdk_bdev_desc *desc;
    4681           0 :         struct spdk_io_channel *ch;
    4682           0 :         struct spdk_nvme_cpl *response = spdk_nvmf_request_get_response(req);
    4683             :         int rc;
    4684             : 
    4685           0 :         rc = spdk_nvmf_request_get_bdev(bdev_nsid, req, &bdev, &desc, &ch);
    4686           0 :         if (rc) {
    4687           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    4688           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    4689           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4690             :         }
    4691           0 :         return spdk_nvmf_bdev_ctrlr_nvme_passthru_admin(bdev, desc, ch, req, NULL);
    4692             : }
    4693             : 
    4694             : static int
    4695           0 : nvmf_passthru_admin_cmd(struct spdk_nvmf_request *req)
    4696             : {
    4697           0 :         struct spdk_nvme_cmd *cmd = spdk_nvmf_request_get_cmd(req);
    4698             :         uint32_t bdev_nsid;
    4699             : 
    4700           0 :         if (g_nvmf_custom_admin_cmd_hdlrs[cmd->opc].nsid != 0) {
    4701           0 :                 bdev_nsid = g_nvmf_custom_admin_cmd_hdlrs[cmd->opc].nsid;
    4702             :         } else {
    4703           0 :                 bdev_nsid = cmd->nsid;
    4704             :         }
    4705             : 
    4706           0 :         return nvmf_passthru_admin_cmd_for_bdev_nsid(req, bdev_nsid);
    4707             : }
    4708             : 
    4709             : int
    4710           0 : nvmf_passthru_admin_cmd_for_ctrlr(struct spdk_nvmf_request *req, struct spdk_nvmf_ctrlr *ctrlr)
    4711             : {
    4712           0 :         struct spdk_nvme_cpl *response = spdk_nvmf_request_get_response(req);
    4713             :         struct spdk_nvmf_ns *ns;
    4714             : 
    4715           0 :         ns = spdk_nvmf_subsystem_get_first_ns(ctrlr->subsys);
    4716           0 :         if (ns == NULL) {
    4717             :                 /* Is there a better sc to use here? */
    4718           0 :                 response->status.sct = SPDK_NVME_SCT_GENERIC;
    4719           0 :                 response->status.sc = SPDK_NVME_SC_INVALID_NAMESPACE_OR_FORMAT;
    4720           0 :                 return SPDK_NVMF_REQUEST_EXEC_STATUS_COMPLETE;
    4721             :         }
    4722             : 
    4723           0 :         return nvmf_passthru_admin_cmd_for_bdev_nsid(req, ns->nsid);
    4724             : }
    4725             : 
    4726             : void
    4727           0 : spdk_nvmf_set_passthru_admin_cmd(uint8_t opc, uint32_t forward_nsid)
    4728             : {
    4729           0 :         g_nvmf_custom_admin_cmd_hdlrs[opc].hdlr = nvmf_passthru_admin_cmd;
    4730           0 :         g_nvmf_custom_admin_cmd_hdlrs[opc].nsid = forward_nsid;
    4731           0 : }
    4732             : 
    4733             : int
    4734           0 : spdk_nvmf_request_get_bdev(uint32_t nsid, struct spdk_nvmf_request *req,
    4735             :                            struct spdk_bdev **bdev, struct spdk_bdev_desc **desc, struct spdk_io_channel **ch)
    4736             : {
    4737           0 :         struct spdk_nvmf_ctrlr *ctrlr = req->qpair->ctrlr;
    4738             :         struct spdk_nvmf_ns *ns;
    4739           0 :         struct spdk_nvmf_poll_group *group = req->qpair->group;
    4740             :         struct spdk_nvmf_subsystem_pg_ns_info *ns_info;
    4741             : 
    4742           0 :         *bdev = NULL;
    4743           0 :         *desc = NULL;
    4744           0 :         *ch = NULL;
    4745             : 
    4746           0 :         ns = _nvmf_subsystem_get_ns(ctrlr->subsys, nsid);
    4747           0 :         if (ns == NULL || ns->bdev == NULL) {
    4748           0 :                 return -EINVAL;
    4749             :         }
    4750             : 
    4751           0 :         assert(group != NULL && group->sgroups != NULL);
    4752           0 :         ns_info = &group->sgroups[ctrlr->subsys->id].ns_info[nsid - 1];
    4753           0 :         *bdev = ns->bdev;
    4754           0 :         *desc = ns->desc;
    4755           0 :         *ch = ns_info->channel;
    4756             : 
    4757           0 :         return 0;
    4758             : }
    4759             : 
    4760           0 : struct spdk_nvmf_ctrlr *spdk_nvmf_request_get_ctrlr(struct spdk_nvmf_request *req)
    4761             : {
    4762           0 :         return req->qpair->ctrlr;
    4763             : }
    4764             : 
    4765           0 : struct spdk_nvme_cmd *spdk_nvmf_request_get_cmd(struct spdk_nvmf_request *req)
    4766             : {
    4767           0 :         return &req->cmd->nvme_cmd;
    4768             : }
    4769             : 
    4770           0 : struct spdk_nvme_cpl *spdk_nvmf_request_get_response(struct spdk_nvmf_request *req)
    4771             : {
    4772           0 :         return &req->rsp->nvme_cpl;
    4773             : }
    4774             : 
    4775           0 : struct spdk_nvmf_subsystem *spdk_nvmf_request_get_subsystem(struct spdk_nvmf_request *req)
    4776             : {
    4777           0 :         return req->qpair->ctrlr->subsys;
    4778             : }
    4779             : 
    4780             : size_t
    4781           0 : spdk_nvmf_request_copy_from_buf(struct spdk_nvmf_request *req,
    4782             :                                 void *buf, size_t buflen)
    4783             : {
    4784           0 :         struct spdk_iov_xfer ix;
    4785             : 
    4786           0 :         spdk_iov_xfer_init(&ix, req->iov, req->iovcnt);
    4787           0 :         return spdk_iov_xfer_from_buf(&ix, buf, buflen);
    4788             : }
    4789             : 
    4790             : size_t
    4791           0 : spdk_nvmf_request_copy_to_buf(struct spdk_nvmf_request *req,
    4792             :                               void *buf, size_t buflen)
    4793             : {
    4794           0 :         struct spdk_iov_xfer ix;
    4795             : 
    4796           0 :         spdk_iov_xfer_init(&ix, req->iov, req->iovcnt);
    4797           0 :         return spdk_iov_xfer_to_buf(&ix, buf, buflen);
    4798             : }
    4799             : 
    4800           0 : struct spdk_nvmf_subsystem *spdk_nvmf_ctrlr_get_subsystem(struct spdk_nvmf_ctrlr *ctrlr)
    4801             : {
    4802           0 :         return ctrlr->subsys;
    4803             : }
    4804             : 
    4805             : uint16_t
    4806           0 : spdk_nvmf_ctrlr_get_id(struct spdk_nvmf_ctrlr *ctrlr)
    4807             : {
    4808           0 :         return ctrlr->cntlid;
    4809             : }
    4810             : 
    4811           0 : struct spdk_nvmf_request *spdk_nvmf_request_get_req_to_abort(struct spdk_nvmf_request *req)
    4812             : {
    4813           0 :         return req->req_to_abort;
    4814             : }

Generated by: LCOV version 1.15