| /spdk/python/spdk/rpc/ |
| H A D | ublk.py | 19 def ublk_start_disk(client, bdev_name, ublk_id=1, num_queues=1, queue_depth=128): argument 26 if queue_depth: 27 params['queue_depth'] = queue_depth
|
| H A D | iscsi.py | 202 queue_depth, argument 232 'queue_depth': queue_depth,
|
| H A D | bdev.py | 984 selector: Multipath selector (round_robin, queue_depth) 1608 queue_depth=None, corrupt_offset=None, corrupt_value=None): 1615 queue_depth: the queue depth at which to trigger the error 1625 if queue_depth is not None: 1626 params['queue_depth'] = queue_depth 1560 bdev_error_inject_error(client, name, io_type, error_type, num=None, queue_depth=None, corrupt_offset=None, corrupt_value=None) global() argument
|
| /spdk/lib/ftl/ |
| H A D | ftl_band_ops.c | 30 assert(rq->io.band->queue_depth > 0); in write_rq_end() 31 rq->io.band->queue_depth--; in write_rq_end() 76 band->queue_depth++; in ftl_band_rq_write() 104 assert(band->queue_depth > 0); in read_rq_end() 105 band->queue_depth--; in read_rq_end() 153 band->queue_depth++; in ftl_band_rq_read() 166 assert(band->queue_depth > 0); in write_brq_end() 167 band->queue_depth--; in write_brq_end() 209 band->queue_depth++; in ftl_band_basic_rq_write() 227 assert(band->queue_depth > 0); in read_brq_end() [all …]
|
| H A D | ftl_writer.c | 64 if (band->queue_depth) { in close_full_bands() 179 assert(0 == writer->band->queue_depth); in ftl_writer_pad_band() 214 if (writer->band->queue_depth) { in ftl_writer_is_halted()
|
| H A D | ftl_band.h | 105 uint64_t queue_depth; member 234 return band->queue_depth; in ftl_band_qd()
|
| H A D | ftl_p2l.c | 155 assert(rq->io.band->queue_depth > 0); in ftl_p2l_ckpt_issue_end() 156 rq->io.band->queue_depth--; in ftl_p2l_ckpt_issue_end()
|
| /spdk/lib/ublk/ |
| H A D | ublk_rpc.c | 83 uint32_t queue_depth; member 91 {"queue_depth", offsetof(struct rpc_ublk_start_disk, queue_depth), spdk_json_decode_uint32, true}, 132 req->queue_depth = UBLK_DEV_QUEUE_DEPTH; in rpc_ublk_start_disk() 144 rc = ublk_start_disk(req->bdev_name, req->ublk_id, req->num_queues, req->queue_depth, in rpc_ublk_start_disk()
|
| H A D | ublk.c | 109 /* ring depth == dev_info->queue_depth. */ 128 uint32_t queue_depth; 917 return ublk->queue_depth; in spdk_ublk_write_config_json() 959 spdk_json_write_named_uint32(w, "queue_depth", ublk->queue_depth); 1693 .queue_depth = ublk->queue_depth, in ublk_dev_info_init() 1838 q->q_depth = ublk->queue_depth; in ublk_ios_init() 1889 uint32_t num_queues, uint32_t queue_depth, in ublk_start_disk() 1924 UBLK_DEBUGLOG(ublk, "bdev %s num_queues %d queue_depth in ublk_start_disk() 127 uint32_t queue_depth; global() member 1873 ublk_start_disk(const char * bdev_name,uint32_t ublk_id,uint32_t num_queues,uint32_t queue_depth,ublk_ctrl_cb ctrl_cb,void * cb_arg) ublk_start_disk() argument [all...] |
| H A D | ublk_internal.h | 60 uint32_t num_queues, uint32_t queue_depth,
|
| /spdk/lib/iscsi/ |
| H A D | tgt_node.h | 53 int queue_depth; member 92 int queue_depth,
|
| H A D | tgt_node.c | 1048 int queue_depth, in iscsi_tgt_node_construct() argument 1134 if (queue_depth > 0 && ((uint32_t)queue_depth <= g_iscsi.MaxQueueDepth)) { in iscsi_tgt_node_construct() 1135 target->queue_depth = queue_depth; in iscsi_tgt_node_construct() 1138 queue_depth, g_iscsi.MaxQueueDepth); in iscsi_tgt_node_construct() 1139 target->queue_depth = g_iscsi.MaxQueueDepth; in iscsi_tgt_node_construct() 1334 spdk_json_write_named_int32(w, "queue_depth", target->queue_depth); in iscsi_tgt_node_info_json()
|
| H A D | iscsi.h | 262 int queue_depth; member
|
| H A D | iscsi_rpc.c | 366 int32_t queue_depth; member 389 {"queue_depth", offsetof(struct rpc_target_node, queue_depth), spdk_json_decode_int32}, 438 req.queue_depth, in rpc_iscsi_create_target_node()
|
| /spdk/python/spdk/spdkcli/ |
| H A D | ui_node_iscsi.py | 81 pg_ig_mappings, queue_depth, g=None, d=None, r=None, argument 109 queue_depth = self.ui_eval_param(queue_depth, "number", None) 118 pg_ig_maps=pg_ig_maps, queue_depth=queue_depth,
|
| /spdk/scripts/ |
| H A D | fio-wrapper | 110 if [[ -e /sys/block/$dev/device/queue_depth ]]; then
|
| H A D | iostat.py | 63 self.queue_depth = value 259 "{:.2f}".format(_stat.queue_depth), 321 "{:.2f}".format(_stat.queue_depth),
|
| /spdk/test/ublk/ |
| H A D | ublk.sh | 76 [[ "$(jq -r ".[${i}].queue_depth" <<< "$ublk_dev")" = "$QUEUE_DEPTH" ]]
|
| /spdk/test/nvme/reset/ |
| H A D | reset.c | 221 submit_io(struct ns_worker_ctx *ns_ctx, int queue_depth) in submit_io() argument 223 while (queue_depth-- > 0) { in submit_io()
|
| /spdk/test/nvmf/ |
| H A D | nvmf.sh | |
| /spdk/lib/nvme/ |
| H A D | nvme_pcie_common.c | 247 pqpair->qpair.queue_depth = 0; in nvme_pcie_ctrlr_construct_admin_qpair() 671 pqpair->qpair.queue_depth); in nvme_pcie_qpair_complete_tracker() 716 (uint32_t)req->cmd.cid, (uint32_t)cpl->status_raw, pqpair->qpair.queue_depth); in nvme_pcie_qpair_complete_tracker() 740 pqpair->qpair.queue_depth--; in nvme_pcie_qpair_abort_trackers() 1694 pqpair->qpair.queue_depth++; in nvme_pcie_qpair_submit_request()
|
| H A D | nvme_tcp.c | 303 tqpair->qpair.queue_depth = 0; in nvme_tcp_alloc_reqs() 982 tqpair->qpair.queue_depth++; in nvme_tcp_qpair_capsule_cmd_send() 985 req->cmd.cdw10, req->cmd.cdw11, req->cmd.cdw12, tqpair->qpair.queue_depth); in nvme_tcp_qpair_capsule_cmd_send() 1032 tqpair->qpair.queue_depth--; in nvme_tcp_qpair_submit_request() 1034 (uint32_t)req->cmd.cid, (uint32_t)cpl.status_raw, tqpair->qpair.queue_depth); in nvme_tcp_qpair_submit_request()
|
| /spdk/test/app/fuzz/llvm_nvme_fuzz/ |
| H A D | llvm_nvme_fuzz.c | 749 run_cmds(uint32_t queue_depth) in run_cmds() argument 754 for (i = 0; i < queue_depth; i++) { in run_cmds()
|
| /spdk/lib/nvmf/ |
| H A D | rdma.c | 46 #define MAX_WR_PER_QP(queue_depth) (queue_depth * 3 + 2) argument 1054 rqpair->qpair.queue_depth = 0; in nvmf_rdma_qpair_initialize() 1995 rqpair->qpair.queue_depth--; in _nvmf_rdma_request_free() 2152 (uintptr_t)rdma_req, (uintptr_t)rqpair, rqpair->qpair.queue_depth); in nvmf_rdma_request_process() 2474 (uintptr_t)rdma_req, (uintptr_t)rqpair, rqpair->qpair.queue_depth); in nvmf_rdma_opts_init() 4772 rqpair->qpair.queue_depth++; in nvmf_rdma_poller_poll()
|
| H A D | tcp.c | 464 tqpair->qpair.queue_depth++; in nvmf_tcp_request_free() 487 tqpair->qpair.queue_depth--; 1450 tqpair->qpair.queue_depth = 0; in nvmf_tcp_discover() 3101 tqpair->qpair.queue_depth); in nvmf_tcp_req_process() 3336 tqpair->qpair.queue_depth); in nvmf_tcp_poll_group_remove()
|