Lines Matching defs:user_dev
885 struct spdk_vhost_user_dev *user_dev;
911 user_dev = to_user_dev(vdev);
912 pthread_mutex_lock(&user_dev->lock);
913 if (user_dev->registered == false) {
915 pthread_mutex_unlock(&user_dev->lock);
919 /* We expect sessions inside user_dev->vsessions to be sorted in ascending
924 if (user_dev->vsessions_num == UINT_MAX) {
925 pthread_mutex_unlock(&user_dev->lock);
931 user_dev->user_backend->session_ctx_size)) {
933 pthread_mutex_unlock(&user_dev->lock);
936 memset(vsession, 0, sizeof(*vsession) + user_dev->user_backend->session_ctx_size);
940 vsession->id = user_dev->vsessions_num++;
945 pthread_mutex_unlock(&user_dev->lock);
953 pthread_mutex_unlock(&user_dev->lock);
962 TAILQ_INSERT_TAIL(&user_dev->vsessions, vsession, tailq);
964 pthread_mutex_unlock(&user_dev->lock);
974 struct spdk_vhost_user_dev *user_dev = to_user_dev(vsession->vdev);
979 pthread_mutex_lock(&user_dev->lock);
981 backend = user_dev->user_backend;
986 pthread_mutex_unlock(&user_dev->lock);
1133 struct spdk_vhost_user_dev *user_dev;
1142 user_dev = to_user_dev(vdev);
1144 pthread_mutex_lock(&user_dev->lock);
1162 pthread_mutex_unlock(&user_dev->lock);
1170 struct spdk_vhost_user_dev *user_dev;
1177 user_dev = to_user_dev(vsession->vdev);
1179 pthread_mutex_lock(&user_dev->lock);
1181 pthread_mutex_unlock(&user_dev->lock);
1187 pthread_mutex_unlock(&user_dev->lock);
1194 struct spdk_vhost_user_dev *user_dev;
1201 user_dev = to_user_dev(vsession->vdev);
1203 pthread_mutex_lock(&user_dev->lock);
1206 pthread_mutex_unlock(&user_dev->lock);
1220 pthread_mutex_unlock(&user_dev->lock);
1249 struct spdk_vhost_user_dev *user_dev;
1254 user_dev = to_user_dev(vdev);
1256 pthread_mutex_lock(&user_dev->lock);
1257 TAILQ_FOREACH(vsession, &user_dev->vsessions, tailq) {
1259 pthread_mutex_unlock(&user_dev->lock);
1264 pthread_mutex_unlock(&user_dev->lock);
1303 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1306 if (pthread_mutex_trylock(&user_dev->lock) != 0) {
1312 user_dev->user_backend->stop_session(vdev, vsession, NULL);
1313 pthread_mutex_unlock(&user_dev->lock);
1322 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1329 pthread_mutex_unlock(&user_dev->lock);
1331 pthread_mutex_lock(&user_dev->lock);
1341 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1343 if (pthread_mutex_trylock(&user_dev->lock) != 0) {
1349 assert(user_dev->pending_async_op_num > 0);
1350 user_dev->pending_async_op_num--;
1355 pthread_mutex_unlock(&user_dev->lock);
1364 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1368 if (pthread_mutex_trylock(&user_dev->lock) != 0) {
1373 TAILQ_FOREACH(vsession, &user_dev->vsessions, tailq) {
1381 pthread_mutex_unlock(&user_dev->lock);
1392 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1406 pthread_mutex_lock(&user_dev->lock);
1407 assert(user_dev->pending_async_op_num < UINT32_MAX);
1408 user_dev->pending_async_op_num++;
1409 pthread_mutex_unlock(&user_dev->lock);
1447 struct spdk_vhost_user_dev *user_dev;
1455 user_dev = to_user_dev(vsession->vdev);
1459 pthread_mutex_lock(&user_dev->lock);
1461 pthread_mutex_unlock(&user_dev->lock);
1465 pthread_mutex_unlock(&user_dev->lock);
1468 pthread_mutex_lock(&user_dev->lock);
1471 pthread_mutex_unlock(&user_dev->lock);
1476 pthread_mutex_unlock(&user_dev->lock);
1481 pthread_mutex_lock(&user_dev->lock);
1489 pthread_mutex_unlock(&user_dev->lock);
1496 pthread_mutex_lock(&user_dev->lock);
1502 pthread_mutex_unlock(&user_dev->lock);
1518 struct spdk_vhost_user_dev *user_dev;
1528 user_dev = to_user_dev(vsession->vdev);
1555 pthread_mutex_lock(&user_dev->lock);
1557 pthread_mutex_unlock(&user_dev->lock);
1561 pthread_mutex_unlock(&user_dev->lock);
1565 pthread_mutex_lock(&user_dev->lock);
1567 pthread_mutex_unlock(&user_dev->lock);
1576 pthread_mutex_unlock(&user_dev->lock);
1679 vhost_user_dev_set_coalescing(struct spdk_vhost_user_dev *user_dev, uint32_t delay_base_us,
1694 user_dev->coalescing_delay_us = delay_base_us;
1695 user_dev->coalescing_iops_threshold = iops_threshold;
1730 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1733 *delay_base_us = user_dev->coalescing_delay_us;
1737 *iops_threshold = user_dev->coalescing_iops_threshold;
1778 struct spdk_vhost_user_dev *user_dev;
1791 user_dev = calloc(1, sizeof(*user_dev));
1792 if (user_dev == NULL) {
1796 vdev->ctxt = user_dev;
1800 free(user_dev);
1806 user_dev->user_backend = user_backend;
1807 user_dev->vdev = vdev;
1808 user_dev->registered = true;
1809 TAILQ_INIT(&user_dev->vsessions);
1810 pthread_mutex_init(&user_dev->lock, NULL);
1812 vhost_user_dev_set_coalescing(user_dev, SPDK_VHOST_COALESCING_DELAY_BASE_US,
1831 struct spdk_vhost_user_dev *user_dev;
1841 user_dev = to_user_dev(vdev);
1843 pthread_mutex_destroy(&user_dev->lock);
1844 free(user_dev);
1856 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1858 if (pthread_mutex_trylock(&user_dev->lock) != 0) {
1863 if (!TAILQ_EMPTY(&user_dev->vsessions) && g_vhost_user_started) {
1865 pthread_mutex_unlock(&user_dev->lock);
1868 pthread_mutex_unlock(&user_dev->lock);
1875 struct spdk_vhost_user_dev *user_dev = to_user_dev(vdev);
1878 if (pthread_mutex_trylock(&user_dev->lock) != 0) {
1882 if (user_dev->pending_async_op_num) {
1883 pthread_mutex_unlock(&user_dev->lock);
1888 if (!TAILQ_EMPTY(&user_dev->vsessions) && g_vhost_user_started) {
1890 pthread_mutex_unlock(&user_dev->lock);
1898 TAILQ_FOREACH_SAFE(vsession, &user_dev->vsessions, tailq, tmp_vsession) {
1900 TAILQ_REMOVE(&user_dev->vsessions, vsession, tailq);
1911 user_dev->registered = false;
1912 pthread_mutex_unlock(&user_dev->lock);
1920 pthread_mutex_destroy(&user_dev->lock);
1922 free(user_dev);
1971 struct spdk_vhost_user_dev *user_dev;
1976 user_dev = to_user_dev(vdev);
1978 pthread_mutex_lock(&user_dev->lock);
1979 TAILQ_FOREACH(vsession, &user_dev->vsessions, tailq) {
1984 pthread_mutex_unlock(&user_dev->lock);
2024 struct spdk_vhost_user_dev *user_dev;
2026 user_dev = to_user_dev(vdev);
2027 pthread_mutex_lock(&user_dev->lock);
2028 TAILQ_FOREACH(vsession, &user_dev->vsessions, tailq) {
2038 pthread_mutex_unlock(&user_dev->lock);